Skip to main content

Cutting-edge framework for orchestrating role-playing, autonomous AI agents. By fostering collaborative intelligence, CrewAI empowers agents to work together seamlessly, tackling complex tasks.

Project description

crewAI

Logo of crewAI, tow people rowing on a boat

🤖 Cutting-edge framework for orchestrating role-playing, autonomous AI agents. By fostering collaborative intelligence, CrewAI empowers agents to work together seamlessly, tackling complex tasks.

Why CrewAI?

The power of AI collaboration has too much to offer. CrewAI is designed to enable AI agents to assume roles, share goals, and operate in a cohesive unit - much like a well-oiled crew. Whether you're building a smart assistant platform, an automated customer service ensemble, or a multi-agent research team, CrewAI provides the backbone for sophisticated multi-agent interactions.

Getting Started

To get started with CrewAI, follow these simple steps:

  1. Installation:
pip install crewai
  1. Setting Up Your Crew:
import os
from crewai import Agent, Task, Crew, Process

os.environ["OPENAI_API_KEY"] = "Your Key"

# Define your agents with roles and goals
researcher = Agent(
  role='Researcher',
  goal='Discover new insights',
  backstory="You're a world class researcher working on a major data science company",
  verbose=True,
  allow_delegation=False
  # llm=OpenAI(temperature=0.7, model_name="gpt-4"). It uses langchain.chat_models, default is GPT4
)
writer = Agent(
  role='Writer',
  goal='Create engaging content',
  backstory="You're a famous technical writer, specialized on writing data related content",
  verbose=True,
  allow_delegation=False
)

# Create tasks for your agents
task1 = Task(description='Search and investigate the latest AI trends', agent=researcher)
task2 = Task(description='Write a blog post on AI advancements', agent=writer)

# Instantiate your crew with a sequential process
crew = Crew(
  agents=[researcher, writer],
  tasks=[task1, task2],
  verbose=2, # Crew verbose more will let you know what tasks are being worked on, you can set it to 1 or 2 to different logging levels
  process=Process.sequential # Sequential process will have tasks executed one after the other and the outcome of the previous one is passed as extra content into this next.
)

# Get your crew to work!
result = crew.kickoff()

Currently the only supported process is Process.sequential, where one task is executed after the other and the outcome of one is passed as extra content into this next.

Key Features

  • Role-Based Agent Design: Customize agents with specific roles, goals, and tools.
  • Autonomous Inter-Agent Delegation: Agents can autonomously delegate tasks and inquire amongst themselves, enhancing problem-solving efficiency.
  • Flexible Task Management: Define tasks with customizable tools and assign them to agents dynamically.
  • Processes Driven: Currently only supports sequential task execution but more complex processes like consensual and hierarchical being worked on.

CrewAI Mind Map

Examples

You can test different real life examples of AI crews in the examples repo

Code

Video

Trip Planner

Trip Planner

Stock Analysis

Stock Analysis

Local Open Source Models

crewAI supports integration with local models, thorugh tools such as Ollama, for enhanced flexibility and customization. This allows you to utilize your own models, which can be particularly useful for specialized tasks or data privacy concerns.

Setting Up Ollama

  • Install Ollama: Ensure that Ollama is properly installed in your environment. Follow the installation guide provided by Ollama for detailed instructions.
  • Configure Ollama: Set up Ollama to work with your local model. You will probably need to tweak the model using a Modelfile. I'd recommend adding Observation as a stop word and playing with top_p and temperature.

Integrating Ollama with CrewAI

  • Instantiate Ollama Model: Create an instance of the Ollama model. You can specify the model and the base URL during instantiation. For example:
from langchain.llms import Ollama
ollama_openhermes = Ollama(model="openhermes")
# Pass Ollama Model to Agents: When creating your agents within the CrewAI framework, you can pass the Ollama model as an argument to the Agent constructor. For instance:

local_expert = Agent(
  role='Local Expert at this city',
  goal='Provide the BEST insights about the selected city',
  backstory="""A knowledgeable local guide with extensive information
  about the city, it's attractions and customs""",
  tools=[
    SearchTools.search_internet,
    BrowserTools.scrape_and_summarize_website,
  ],
  llm=ollama_openhermes, # Ollama model passed here
  verbose=True
)

How CrewAI Compares

  • Autogen: While Autogen excels in creating conversational agents capable of working together, it lacks an inherent concept of process. In Autogen, orchestrating agents' interactions requires additional programming, which can become complex and cumbersome as the scale of tasks grows.

  • ChatDev: ChatDev introduced the idea of processes into the realm of AI agents, but its implementation is quite rigid. Customizations in ChatDev are limited and not geared towards production environments, which can hinder scalability and flexibility in real-world applications.

CrewAI's Advantage: CrewAI is built with production in mind. It offers the flexibility of Autogen's conversational agents and the structured process approach of ChatDev, but without the rigidity. CrewAI's processes are designed to be dynamic and adaptable, fitting seamlessly into both development and production workflows.

Contribution

CrewAI is open-source and we welcome contributions. If you're looking to contribute, please:

  • Fork the repository.
  • Create a new branch for your feature.
  • Add your feature or improvement.
  • Send a pull request.
  • We appreciate your input!

Installing Dependencies

poetry lock
poetry install

Virtual Env

poetry shell

Pre-commit hooks

pre-commit install

Running Tests

poetry run pytest

Packaging

poetry build

Installing Locally

pip install dist/*.tar.gz

License

CrewAI is released under the MIT License

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

crewai-0.1.15.tar.gz (14.7 kB view details)

Uploaded Source

Built Distribution

crewai-0.1.15-py3-none-any.whl (16.1 kB view details)

Uploaded Python 3

File details

Details for the file crewai-0.1.15.tar.gz.

File metadata

  • Download URL: crewai-0.1.15.tar.gz
  • Upload date:
  • Size: 14.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.11.7

File hashes

Hashes for crewai-0.1.15.tar.gz
Algorithm Hash digest
SHA256 59f69b4b3803d9a6643ba120ce1a814910238f754f4d121a05d4b4773779e395
MD5 6113bbec150720fa5e8da44d3337a83d
BLAKE2b-256 a796f3a71fbb36d371b8313d606d88a411572ef56aeb417c44c1297977c8b952

See more details on using hashes here.

File details

Details for the file crewai-0.1.15-py3-none-any.whl.

File metadata

  • Download URL: crewai-0.1.15-py3-none-any.whl
  • Upload date:
  • Size: 16.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.11.7

File hashes

Hashes for crewai-0.1.15-py3-none-any.whl
Algorithm Hash digest
SHA256 e07e9b3e11cacf754d608d31e91d8a392526439904b8f3ca3487387c78218906
MD5 d58119c8c0e2ac28b20950323abb34f0
BLAKE2b-256 4990f9a2a9a2b34db6a820e8e90824af071cb062f61f8722c6ab6985c358bcd2

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page