Skip to main content

A Python library for performing deep research using AI agents and Firecrawl, with support for custom models and report generation.

Project description

tinyagent_deepsearch

tinyagent_deepsearch Logo

tinyagent_deepsearch is a Python library from Alchemist Studios AI, developed by tunahorse21 (larock22), designed to facilitate deep research on various topics using AI agents, powered by OpenAI and Firecrawl for web scraping and content analysis. It leverages the tiny_agent_os framework for structuring AI agent interactions.

Note: Currently, tinyagent_deepsearch uses Firecrawl for web research and content extraction. In the future, the plan is to combine various agents and tools—including tiny_agent_os, browser-based agents (such as browser-use), and other agentic utilities—to enable even more powerful, multi-modal research workflows. Stay tuned for updates as the project evolves!

Features

  • Perform recursive, multi-step research on a given topic.
  • Generate focused search queries based on evolving learnings.
  • Utilize Firecrawl to scrape web content.
  • Employ OpenAI's language models to digest information and identify follow-up questions.
  • Configurable research depth and breadth.

Installation

You can install tinyagent_deepsearch using pip:

pip install tinyagent_deepsearch

(Note: This command assumes the package will be published to PyPI. For local installation from source, navigate to the project root directory where pyproject.toml is located and run pip install .)

Prerequisites

Before using the library, ensure you have the following API keys set as environment variables:

  • OPENAI_KEY: Your API key for OpenAI.
  • FIRECRAWL_KEY: Your API key for Firecrawl.

You can set them in your shell environment or by using a .env file in your project root (requires python-dotenv to be installed in your project).

Example .env file:

OPENAI_KEY="your_openai_api_key_here"
FIRECRAWL_KEY="your_firecrawl_api_key_here"

tiny_agent_os Configuration (config.yml)

This library relies on the tiny_agent_os framework. tiny_agent_os typically requires a config.yml file in the root of your project for its own operational settings (like default LLM choices, API endpoints for various services, etc.).

While tinyAgent_deepsearch allows you to specify the llm_model directly for its core research function, the underlying tiny_agent_os may still need a config.yml to function correctly for its internal operations or if you use tiny_agent_os features directly elsewhere in your project.

For detailed information on how to set up the config.yml for tiny_agent_os, please refer to its official documentation: https://github.com/alchemiststudiosDOTai/tinyAgent

Ensure this file is present and correctly configured in your project's root directory if you encounter issues related to tiny_agent_os configuration.

Usage

Here's a basic example of how to use the deep_research function:

import asyncio
from tinyagent_deepsearch import deep_research
from dotenv import load_dotenv # Optional: if you use a .env file

async def main():
    # Optional: Load environment variables from .env file
    # load_dotenv()

    topic = "The future of renewable energy sources"
    breadth = 3  # Number of search queries per depth level
    depth = 2    # Number of recursive research levels

    try:
        print(f"Starting deep research on: {topic}")
        results = await deep_research(
            topic=topic,
            breadth=breadth,
            depth=depth,
            llm_model="gpt-4o-mini", # Optional: specify LLM model
            concurrency=2,           # Optional: specify concurrency
            save_report=True,        # Optional: save research report to file
            report_dir="research_reports", # Optional: custom report directory
            report_format="json"    # Optional: 'json' or 'txt'
        )
        
        # If report was saved, show the path
        if "report_path" in results:
            print(f"Report saved to: {results['report_path']}")
        print("\n=== Research Complete ===")
        print("\nLearnings:")
        for i, learning in enumerate(results.get("learnings", [])):
            print(f"{i+1}. {learning}")

        print("\nVisited URLs:")
        for i, url in enumerate(results.get("visited", [])):
            print(f"{i+1}. {url}")

    except Exception as e:
        print(f"An error occurred: {e}")

if __name__ == "__main__":
    asyncio.run(main())

Configuration

The deep_research function accepts the following parameters:

Core Parameters:

  • topic (str): The initial research topic.
  • breadth (int): The number of search queries to generate at each depth level.
  • depth (int): The number of recursive research levels.

Model & Performance Settings:

  • llm_model (str, optional): The OpenAI model to use. Defaults to "gpt-4o-mini".
  • concurrency (int, optional): The maximum number of concurrent search and digest operations. Defaults to 2.

Report Generation:

  • save_report (bool, optional): Whether to save the research results to a file. Defaults to False.
  • report_dir (str | Path, optional): Directory where the report will be saved. If not provided, creates a 'reports' directory in the current working directory.
  • report_name (str, optional): Name of the report file. If not provided, a name is generated based on the topic and timestamp.
  • report_format (str, optional): Format of the report file. Can be 'json' or 'txt'. Defaults to 'json'.

Optional State:

  • learnings (List[str], optional): Optional list of initial learnings to build upon.
  • visited (List[str], optional): Optional list of initially visited URLs to avoid duplicates.

Contributing

Contributions are welcome! Please feel free to submit a pull request or open an issue. (Further details to be added)

License

This project is licensed under the MIT License. See the LICENSE file for details.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

tinyagent_deepsearch-0.1.3.tar.gz (3.0 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

tinyagent_deepsearch-0.1.3-py3-none-any.whl (10.2 kB view details)

Uploaded Python 3

File details

Details for the file tinyagent_deepsearch-0.1.3.tar.gz.

File metadata

  • Download URL: tinyagent_deepsearch-0.1.3.tar.gz
  • Upload date:
  • Size: 3.0 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.12.3

File hashes

Hashes for tinyagent_deepsearch-0.1.3.tar.gz
Algorithm Hash digest
SHA256 2cf7aff2630c5a96c614e9471428863040c60c728c36ad4ffdec14b8b0b60ef6
MD5 de991972ed2c0b2c209570699e239cc4
BLAKE2b-256 87c60d7f9f04a1d32abce0bdc4c7e0c32080a06277d6bd3f0ff6b54958b5ff1b

See more details on using hashes here.

File details

Details for the file tinyagent_deepsearch-0.1.3-py3-none-any.whl.

File metadata

File hashes

Hashes for tinyagent_deepsearch-0.1.3-py3-none-any.whl
Algorithm Hash digest
SHA256 47daa78eb63b5e4de869ef5c067c6bb989058b5928e71105b69f4eeb46ea2342
MD5 76754999b8e20f72e748d9f1ecc52de7
BLAKE2b-256 b41c5bf59764796150da39eb5e0cd12c8a7efeacbad3f5f458869a1c1f1dca7d

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page