Skip to main content

Toolkit For Creating Gemini Based Agents

Project description

img.png

gemini-agents-toolkit

The project is an SDK for implementing Agent-Driven Development (ADD) applications. ADD is aimed to move routine code tasks to the LLM agents and concentrate on the algorithmic and fun part of the business logic implementation.

license discord youtube

⭐ Add a star for a duck!


How It Works  •  Requirements  •  Getting Started  •  Run Examples  •  How To Contribute


🚀How It Works

gemini-agents-toolkit is an SDK that creates LLM agents and enables their integration into pipelines for modifying generated responses.

See the picture describing the process:

img.png

The roles of every component are as follows:

  1. Application: Define custom functions to guide LLM executions, launch agents for task execution, and combine tasks into pipelines.
  2. gemini-agents-toolkit: A tool for creating agents, executing pipelines and interacting with Gemini.
  3. Vertex AI API: A Google API for interacting with Gemini models.
  4. Gemini: LLM models that generate text, code, or instructions to guide agent execution.

📝Requirements

❗️ The project uses Vertex AI API to access Gemini models, which is paid Google API ❗

You can find pricing here.

Agents request text generating only, which is paid for Text Input and Text Output. Running a pipeline will produce several requests depending on the pipeline's complexity.

========

Supported Python versions: Python 3.8 to 3.12 (gcloud requirement)


🤖Getting Started

Big Picture

This project uses Gemini via Google Cloud API. To use the project you must first set up the following:

  • Google Cloud account with Billing
  • Allow requesting LLM models via APIs (pricing)
  • Set up gcloud tool in your environment
  • Set up the project

Google Cloud account with Billing

  1. Go to Google AI Studio, select "Get API key" in the menu on the left. Click "Create API" key button and select Google Cloud project to associate API key with.
  2. Click suggestion to "Set up Billing" near your new API key and initialize your Google Billing account.
  3. [Optional] You can add a Budget for your account, which will email you when your spendings reach configured thresholds. Choose "Budgets & alerts" in the menu on the left in Billing console and follow the instruction.
  4. After this step, you should be able to test your Google Cloud project set up running curl command presented under the API keys table.

Allow requesting LLM models via APIs

  1. The project uses Vertex AI API to request Gemini models, Vertex AI API should be allowed in your Google cloud account. You need to allow it from the Google Cloud Vertex AI API product page.
  2. [Optional] With these instructions you can limit or increase your API requests quota.

Set up gcloud tool in your environment

Follow the Google instructions to install gcloud and to initialize it.

Set up the project

  1. Clone gemini-agents-toolkit repository.
  2. For the development process, you probably will need other repositories from the project. You can find other packages here. Currently, the project uses this list of other repositories:
Module Description Link
json-agent Agent to convert data formats to JSON link
  1. gemini-agents-toolkit uses several environment variables that you need to set up:
Env Variable Description Which Value To Set Command To Set Up
GOOGLE_PROJECT Project id from Google AI Studio Open Google AI studio and choose from the table with projects EXPORT GOOGLE_PROJECT=my-amazing-project
GOOGLE_API_KEY API key from Google AI Studio to request your project Find out the API key in the row with your project EXPORT GOOGLE_API_KEY=my-amazing-project
GOOGLE_REGION You can choose which region API to request You can ignore env for the default value us-west1, or find all available regions in Google Cloud docs EXPORT GOOGLE_REGION=my-amazing-project
  1. Now, you need to register custom Python modules to use them in examples:
export PYTHONPATH=/my/custom/module/path:$PYTHONPATH

Modules you need to register:

Module Description Link
config Using env variables and constants link
json-agent Agent to convert data formats to JSON link

❗ To have your variables set up permanently, register them in .bashrc (.zshrc etc, depending on your operational system and shell).


🎉Run Examples

Now it's time for fun!

Run examples/simple_example.py with

python examples/simple_example.py

In this example, you have created a custom function and gave Gemini ability to use your function:

See code
import vertexai
from config import (PROJECT_ID, REGION, SIMPLE_MODEL)
from gemini_agents_toolkit import agent


def say_to_duck(say: str):
    """say something to a duck"""
    return f"duck answer is: duck duck {say} duck duck duck"


vertexai.init(project=PROJECT_ID, location=REGION)

all_functions = [say_to_duck]
duck_comms_agent = agent.create_agent_from_functions_list(functions=all_functions,
                                                          model_name=SIMPLE_MODEL)

print(duck_comms_agent.send_message("say to the duck message: I am hungry"))

=>>>><<<<=

You can create several agents, which will delegate functions execution to each other:

python examples/multi_agent_example.py
See code
import datetime
import vertexai
from config import (PROJECT_ID, REGION, SIMPLE_MODEL)
from gemini_agents_toolkit import agent

vertexai.init(project=PROJECT_ID, location=REGION)


def generate_duck_comms_agent():
    """create an agent to say to a duck"""

    def say_to_duck(say: str):
        """say something to a duck"""
        return f"duck answer is: duck duck {say} duck duck duck"

    return agent.create_agent_from_functions_list(
        functions=[say_to_duck],
        delegation_function_prompt=("""Agent can communicat to ducks and can say something to them.
                                    And provides the answer from the duck."""),
        model_name=SIMPLE_MODEL)


def generate_time_checker_agent():
    """create an agent to get the time"""

    def get_local_time():
        """get the current local time"""
        return datetime.datetime.now().strftime("%Y-%m-%d %H:%M:%S")

    return agent.create_agent_from_functions_list(
        functions=[get_local_time],
        delegation_function_prompt="Agent can provide the current local time.",
        model_name=SIMPLE_MODEL)


duck_comms_agent = generate_duck_comms_agent()
time_checker_agent = generate_time_checker_agent()

main_agent = agent.create_agent_from_functions_list(
    delegates=[time_checker_agent, duck_comms_agent],
    model_name=SIMPLE_MODEL)

print(main_agent.send_message("say to the duck message: I am hungry"))
print(main_agent.send_message("can you tell me what time it is?"))

=>>>><<<<=

You can execute your code periodicaly:

python examples/simple_scheduler_example.py
See code
import time
import vertexai
from config import (PROJECT_ID, REGION, SIMPLE_MODEL)
from gemini_agents_toolkit import agent


def say_to_duck(say: str):
    """say something to a duck"""
    return f"duck answer is: duck duck {say} duck duck duck"


def print_msg_from_agent(msg: str):
    """print message in console"""
    print(msg)


vertexai.init(project=PROJECT_ID, location=REGION)

all_functions = [say_to_duck]
duck_comms_agent = agent.create_agent_from_functions_list(functions=all_functions,
                                                          model_name=SIMPLE_MODEL,
                                                          add_scheduling_functions=True,
                                                          on_message=print_msg_from_agent)

# no need to print result directly since we passed to agent on_message
duck_comms_agent.send_message("can you be saying, each minute, to the duck that I am hungry")

# wait 3 min to see results
time.sleep(180)

Find more advanced examples in the examples directory.


💡How To Contribute

If you want to be a code contributor

  • Just pick up a task from the Issues in this repository, assign it to you and raise a pull request with proposed changes.
  • If you need help, join the Discord and ask for help. The contributors team will be happy to see you!

If you want to contribute ideas or participate discussions

Feel free to join our Discord, where you can discuss the project with the contributors and probably impact on the way the project evolves.


Back to top

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

gemini_agents_toolkit-3.1.1.tar.gz (17.8 kB view details)

Uploaded Source

Built Distribution

gemini_agents_toolkit-3.1.1-py3-none-any.whl (33.7 kB view details)

Uploaded Python 3

File details

Details for the file gemini_agents_toolkit-3.1.1.tar.gz.

File metadata

  • Download URL: gemini_agents_toolkit-3.1.1.tar.gz
  • Upload date:
  • Size: 17.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.12.5

File hashes

Hashes for gemini_agents_toolkit-3.1.1.tar.gz
Algorithm Hash digest
SHA256 3df503296a8b6d10ab22178a9c6cf97aed51749d5d504f42c57d2690666b7a5f
MD5 b506feae9594b6d5be9b9e4a38e6b897
BLAKE2b-256 c395e5123b0e6c42b0a413cca04973028be3f98bb1b2773b3801090d42317eda

See more details on using hashes here.

File details

Details for the file gemini_agents_toolkit-3.1.1-py3-none-any.whl.

File metadata

File hashes

Hashes for gemini_agents_toolkit-3.1.1-py3-none-any.whl
Algorithm Hash digest
SHA256 ab3faadea38949a6cd87dd516d7db9e8f2bdc79f0da3a6d09f73c68fd79f5369
MD5 dd85fcb4d8d933931db20276f90f2ba8
BLAKE2b-256 10e6d6150b07fa327c76052a895ca5bcec6dc8519dcb3dc29c5eed5986e7625e

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page