Skip to main content

GPT's function calling feature wrapper

Project description

CallingGPT

PyPi

GPT's Function Calling Demo, a experiment of self-hosted ChatGPT-Plugins-like platform.

Recommend reading: function-calling

Abstract

OpenAI's GPT models provide a function calling feature, so we can easily create ChatGPT-Plugins-like tools. This repository is a proof-of-concept of the function calling feature.
In this experiment, we defined the Plugin as Namespace which contains a serial of functions. While user performing a conversation, the functions in Namespace will be called by the API and return the result to the user.

Usage

  1. Clone this repository and install the dependencies.

    git clone https://github.com/RockChinQ/CallingGPT
    cd CallingGPT
    pip install -r requirements.txt
    
  2. Run the main.py to generate config.yaml

    python main.py
    
  3. Edit the config.yaml to set your API key and other settings.

  4. Run the main.py and pass your modules.

    python main.py <module0> <module1> ...
    

Example

Use the example/greet.py, provides a greet function called when user ask GPT to greet someone.

python main.py example/greet.py

Then you can talk to the bot.

$ python main.py examples/greet.py 
Using module: examples.greet
>>> Hello and who are you?
<<< Hello! I am an AI assistant. How can I assist you today?
>>> say hello to Rock
call<examples-greet-greet>: {
  "user": "Rock"
}
<<< Hello, Rock! How can I assist you today?
>>> and to Alice
call<examples-greet-greet>: {
  "user": "Alice"
}
<<< Hello, Alice! How can I assist you today?
>>>

Type help to get help.
See wiki for the function format.

Other Examples

examples/draw_and_wrapper_md.py

Provides a dalle_draw function to use DALL·E model when user ask GPT to draw something.

python main.py examples/draw_and_wrapper_md.py 
$ python main.py examples/draw_and_wrapper_md.py 
Using module: examples.draw_and_wrapper_md
>>> hello!
<<< Hi there! How can I assist you today?
>>> draw a sunset for me please
call<examples-draw_and_wrapper_md-draw>: {
  "prompt": "sunset"
}
<<< Sure! Here's a beautiful sunset for you:

![Sunset](https://oaidalleapiprodscus.blob.core.windows.net/private/org-VS9HEpJba78GXVfOcmVo7qaM/user-OHa7Jo3kL4XJDg9lo7AzdWNT/img-QmDUiwp1IGFcu8pDGZh0i7r8.png)

I hope you like it! Let me know if there's anything else I can help you with.
>>> 

For Code

  1. Install the package

    pip install --upgrade CallingGPT
    
  2. Create your own functions in modules(these modules can also be used in the CLI mode)

    # your_module_a.py
    def func_a(prompt: str) -> str:  # Type hint of EACH argument and return value is REQUIRED.
        """
        The description of this func a, will be provided to the api.
    
        Args:
            prompt(str): The prompt of the function.
    
        Returns:
            The result of the function.
        """
        # Google style docstring is REQUIRED, it will be split into
        # `description` and `params`(required if there are args) and 
        # `returns`(optional), `\n\n` between each part.
        return "func_a: " + prompt
    
    # your_module_b.py
    def adder(a: int, b: int) -> int:
        """
        Add two numbers.
    
        Args:
            a: The first number.
            b: The second number.
    
        Returns:
            The sum of a and b.
        """
        # Type hints of args in docstring is optional.
        return a + b
    
  3. Call the wrapper

    from CallingGPT.session.session import Session
    import your_module_a, your_module_b
    import openai
    
    openai.api_key = 'your_openai_api_key'
    
    session = Session([your_module_a, your_module_b])
    
    for reply in session.ask("your prompt"):
        # session.ask will yield each time the api returns a result,
        # before calling function, it will print the function name and args.
        # e.g. here's a function call:
        # {
        #   "role": "assistant",
        #   "content": null,
        #   "function_call": {
        #     "name": "examples-draw_and_wrapper_md-draw",
        #     "arguments": "{\n  \"prompt\": \"cat\"\n}"
        #   }
        # }
        # 
        # while here's a normal reply:
        # {
        #   "role": "assistant",
        #   "content": "Hello, I am an AI assistant. How can I assist you today?"
        # }
        print(reply)
    

    Session will automatically manage context for you.

See wiki for the function format.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

CallingGPT-0.0.1.0.tar.gz (8.4 kB view details)

Uploaded Source

Built Distribution

CallingGPT-0.0.1.0-py3-none-any.whl (7.2 kB view details)

Uploaded Python 3

File details

Details for the file CallingGPT-0.0.1.0.tar.gz.

File metadata

  • Download URL: CallingGPT-0.0.1.0.tar.gz
  • Upload date:
  • Size: 8.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.10.8

File hashes

Hashes for CallingGPT-0.0.1.0.tar.gz
Algorithm Hash digest
SHA256 7cd173d2bca7202b9fcb311342675dd2a5d409ccdb0f245dd1e6596a8bbabf2d
MD5 dfbeafa864f71207dfac2a433ef514f0
BLAKE2b-256 0622a64557f5fe6b3db94f2aa899c566377a07c527345b08624f79c4e795b17e

See more details on using hashes here.

File details

Details for the file CallingGPT-0.0.1.0-py3-none-any.whl.

File metadata

File hashes

Hashes for CallingGPT-0.0.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 0d275cd667d20c29296d03ce81a8b2c0328b96528bd6d6da97e6dffd23c31608
MD5 b4235bd928fb1a83547fb46a4d970f3b
BLAKE2b-256 47f432bb0676a425afd1961018ce3432602d6a91b1616eca57ce02b478e47647

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page