Skip to main content

Interact with AI without API key

Project description

python-tgpt

License PyPi Black Passing coverage Progress Downloads Downloads Latest release release date wakatime

AI for all

>>> import tgpt
>>> bot = tgpt.TGPT()
>>> bot.chat('Hello there')
"  Hello! It's nice to meet you. Is there something I can help you with or would you like to chat?"
>>> 

This project enables seamless interaction with LLaMA AI without requiring an API Key.

The name python-tgpt draws inspiration from its parent project tgpt, which operates on golang. Through this Python adaptation, users can effortlessly engage with LLaMA's capabilities, fostering a smoother AI interaction experience.

Features

  • 🗨️ Enhanced conversational chat experience
  • 💾 Capability to save prompts and responses (Conversation)
  • 🔄 Ability to load previous conversations
  • ⌨️ Command-line interface
  • 🐍 Python package
  • 🌊 Stream and non-stream response
  • 🚀 Ready to use (No API key required)

Prerequisites

Installation and Usage

Installation

Choose one of the following methods to get started.

  1. From PyPI:

    pip install --upgrade python-tgpt
    
  2. Directly from the source:

    pip install git+https://github.com/Simatwa/python-tgpt.git
    
  3. Clone and Install:

    git clone https://github.com/Simatwa/python-tgpt.git
    cd python-tgpt
    pip install .
    

Usage

This package offers a convenient command-line interface.

  • For a quick response:

    python -m tgpt generate "<Your prompt>"
    
  • For interactive mode:

    python -m tgpt interactive "<Kickoff prompt (though not mandatory)>"
    

You can also simply use tgpt instead of python -m tgpt.

Starting from version 0.0.6, generate is the default command. Therefore, something like this will still work: tgpt "<Your prompt>".

Developer Docs

  1. Generate a quick response
from tgpt import TGPT
bot = TGPT()
resp = bot.chat('<Your prompt>')
print(resp)
# Output : How may I help you.
  1. Get back whole response
from tgpt import TGPT
bot = TGPT()
resp = bot.ask('<Your Prompt')
print(resp)
# Output
"""
{'completion': "I'm so excited to share with you the incredible experiences...", 'stop_reason': None, 'truncated': False, 'stop': None, 'model': 'llama-2-13b-chat', 'log_id': 'cmpl-3NmRt5A5Djqo2jXtXLBwJ2', 'exception': None}
"""

Stream Response

Just add parameter stream with value true.

  1. Text Generated only
from tgpt import TGPT
bot = TGPT()
resp = bot.chat('<Your prompt>', stream=True)
for value in resp:
    print(value)
# output
"""
How may
How may I help 
How may I help you
How may I help you today?
"""
  1. Whole Response
from tgpt import TGPT
bot = TGPT()
resp = bot.ask('<Your Prompt>', stream=True)
for value in resp:
    print(value)
# Output
"""
{'completion': "I'm so", 'stop_reason': None, 'truncated': False, 'stop': None, 'model': 'llama-2-13b-chat', 'log_id': 'cmpl-3NmRt5A5Djqo2jXtXLBwxx', 'exception': None}

{'completion': "I'm so excited to share with.", 'stop_reason': None, 'truncated': False, 'stop': None, 'model': 'llama-2-13b-chat', 'log_id': 'cmpl-3NmRt5A5Djqo2jXtXLBwxx', 'exception': None}

{'completion': "I'm so excited to share with you the incredible ", 'stop_reason': None, 'truncated': False, 'stop': None, 'model': 'llama-2-13b-chat', 'log_id': 'cmpl-3NmRt5A5Djqo2jXtXLBwxx', 'exception': None}

{'completion': "I'm so excited to share with you the incredible experiences...", 'stop_reason': None, 'truncated': False, 'stop': None, 'model': 'llama-2-13b-chat', 'log_id': 'cmpl-3NmRt5A5Djqo2jXtXLBwxx', 'exception': None}
"""

To obtain more tailored responses, consider utilizing optimizers using the optimizer parameter. Its values can be set to either code or system_command.

from tgpt import TGPT
bot = TGPT()
resp = bot.ask('<Your Prompt>', optimizer='code')
print(resp)

Note: Commencing from v0.1.0, the default mode of interaction is conversational. This mode enhances the interactive experience, offering better control over the chat history. By associating previous prompts and responses, it tailors conversations for a more engaging experience.

You can still disable the mode:

bot = tgpt.TGPT(is_conversation=False)

Utilize the --disable-conversation flag in the console to achieve the same functionality.

Acknowledgements

  1. tgpt
  2. You

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

python-tgpt-0.1.0.tar.gz (17.6 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

python_tgpt-0.1.0-py3-none-any.whl (18.1 kB view details)

Uploaded Python 3

File details

Details for the file python-tgpt-0.1.0.tar.gz.

File metadata

  • Download URL: python-tgpt-0.1.0.tar.gz
  • Upload date:
  • Size: 17.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.9.18

File hashes

Hashes for python-tgpt-0.1.0.tar.gz
Algorithm Hash digest
SHA256 5adec8cb63aa49140f94257759e3394fd8de543826e11a16f4ba267e70dc47ec
MD5 bb3ba01c375c47a81f3e7b003443c0d8
BLAKE2b-256 1540d7a3a87bfa333da7ac341e536b8b8b3f07435c84b82696b3462828a32b79

See more details on using hashes here.

File details

Details for the file python_tgpt-0.1.0-py3-none-any.whl.

File metadata

  • Download URL: python_tgpt-0.1.0-py3-none-any.whl
  • Upload date:
  • Size: 18.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.9.18

File hashes

Hashes for python_tgpt-0.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 48b1ed5c7a0ffa59abd218400383dbab38b9cac8cf2d1751c194c7f0f2875664
MD5 8b94bc9bd84b371ff99478bc565a75bc
BLAKE2b-256 8ce86cc1c0862b5bcdd323564236d44dc49102075792649ff526a4e766f838c7

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page