Skip to main content

AI-powered Python library that extends pandas to import and analyze complex, unstructured files

Project description

Fundas

Fundamental Data Source - An AI-powered Python library that extends pandas to import and analyze complex, unstructured files.

Tests PyPI version Python 3.8+ License: MIT

Overview

Fundas leverages the OpenRouter API and generative AI to intelligently extract features and structured data from various file types based on simple prompts. It seamlessly converts any file into a clean pandas DataFrame for immediate analysis.

Features

  • 📄 read_pdf() - Extract structured data from PDF documents
  • 🖼️ read_image() - Extract data and text from images
  • 🎵 read_audio() - Process audio files and extract information
  • 🌐 read_webpage() - Scrape and structure web content
  • 🎥 read_video() - Analyze video content from frames, audio, or both

All functions return pandas DataFrames, making the data ready for immediate analysis!

Installation

pip install fundas

Or install from source:

git clone https://github.com/AMSeify/fundas.git
cd fundas
pip install -e .

Quick Start

Setup

First, set your OpenRouter API key. You can either:

Option 1: Use environment variable

export OPENROUTER_API_KEY="your-api-key-here"

Option 2: Use .env file (recommended)

# Copy the example file
cp .env.example .env

# Edit .env and add your credentials:
# OPENROUTER_API_KEY=your-api-key-here
# OPENROUTER_MODEL=openai/gpt-3.5-turbo  # Optional: set default model

Option 3: Pass directly to functions

import fundas as fd

df = fd.read_pdf("document.pdf", api_key="your-api-key-here")

Basic Usage

Read PDF Files

import fundas as fd

# Extract invoice data
df = fd.read_pdf(
    "invoice.pdf",
    prompt="Extract invoice items with product name, quantity, and price"
)
print(df)

Read Images

# Extract data from a chart or screenshot
df = fd.read_image(
    "sales_chart.png",
    prompt="Extract the sales data points from this chart"
)
print(df)

# Process a receipt
df = fd.read_image(
    "receipt.jpg",
    prompt="Extract items and their prices",
    columns=["item", "price", "quantity"]
)

Read Webpages

# Scrape product information
df = fd.read_webpage(
    "https://example.com/products",
    prompt="Extract product names, descriptions, and prices"
)
print(df)

# Extract article data
df = fd.read_webpage(
    "https://news.example.com/article",
    columns=["title", "author", "date", "content"]
)

Read Audio Files

# Transcribe and extract meeting notes
df = fd.read_audio(
    "meeting.mp3",
    prompt="Extract speaker names and key discussion points"
)

Read Video Files

# Analyze video frames
df = fd.read_video(
    "presentation.mp4",
    prompt="Extract slide titles and key points from this presentation",
    from_="pics"  # Extract from video frames
)

# Process audio track
df = fd.read_video(
    "lecture.mp4",
    prompt="Transcribe the lecture and identify key topics",
    from_="audios"  # Extract from audio track
)

# Analyze both video and audio
df = fd.read_video(
    "interview.mp4",
    prompt="Extract interview questions and answers",
    from_="both"  # or from_=["pics", "audios"]
)

Advanced Usage

Specify Columns

You can specify which columns you want to extract:

df = fd.read_pdf(
    "report.pdf",
    prompt="Extract quarterly financial data",
    columns=["quarter", "revenue", "expenses", "profit"]
)

Custom AI Models

Use different AI models via OpenRouter:

# Option 1: Pass model parameter to each function
df = fd.read_image(
    "complex_diagram.png",
    prompt="Extract relationships between components",
    model="anthropic/claude-3-opus"
)

# Option 2: Set default model in .env file
# OPENROUTER_MODEL=anthropic/claude-3-sonnet
df = fd.read_image("diagram.png", prompt="Extract data")  # Uses model from .env

# Option 3: Set via environment variable
import os
os.environ["OPENROUTER_MODEL"] = "openai/gpt-4"
df = fd.read_pdf("document.pdf", prompt="Extract info")

DataFrame Operations

Since all functions return pandas DataFrames, you can immediately use pandas operations:

import fundas as fd

# Read and analyze in one workflow
df = fd.read_pdf("sales.pdf", prompt="Extract sales data")
print(df.head())
print(df.describe())
print(df.groupby('region')['sales'].sum())

Requirements

  • Python >= 3.8
  • pandas >= 1.3.0
  • requests >= 2.25.0
  • PyPDF2 >= 3.0.0
  • Pillow >= 10.3.0
  • beautifulsoup4 >= 4.9.0
  • opencv-python >= 4.8.1.78

Advanced Features

Caching

Fundas includes an intelligent caching system to reduce redundant API calls:

import fundas as fd

# Enable caching (enabled by default)
df = fd.read_pdf("document.pdf", prompt="Extract data")

# The same file with the same prompt will use cached results
df2 = fd.read_pdf("document.pdf", prompt="Extract data")  # No API call

# Disable caching if needed
from fundas import OpenRouterClient
client = OpenRouterClient(api_key="key", use_cache=False)

Exporting Data

Export your DataFrames with AI-powered summarization:

import fundas as fd
import pandas as pd

# Create a DataFrame
df = pd.DataFrame({"product": ["A", "B", "C"], "sales": [100, 200, 150]})

# Export to CSV
fd.to_summarized_csv(df, "output.csv")

# Export to Excel with AI summary
fd.to_summarized_excel(
    df,
    "summary.xlsx",
    prompt="Add a summary row with totals"
)

# Generate AI summary
summary = fd.summarize_dataframe(df, prompt="Summarize sales performance")
print(summary)

Error Handling

Fundas includes robust error handling with automatic retries:

import fundas as fd

try:
    df = fd.read_pdf("document.pdf", prompt="Extract data")
except FileNotFoundError:
    print("File not found")
except ValueError as e:
    print(f"Invalid parameters: {e}")
except RuntimeError as e:
    print(f"API error: {e}")

Cache Management

Control the cache behavior:

from fundas import get_cache

cache = get_cache()

# Clear all cache entries
cache.clear()

# Clear only expired entries
cache.clear_expired()

# Disable/enable cache
cache.disable()
cache.enable()

API Reference

Read Functions

All read functions share similar parameters:

Common Parameters:

  • filepath or url (str | Path): Source file or URL
  • prompt (str): Description of what data to extract
  • columns (List[str], optional): Column names to extract
  • api_key (str, optional): OpenRouter API key
  • model (str, optional): AI model to use (default: gpt-3.5-turbo)

Returns: pandas DataFrame

Export Functions

All export functions accept:

Parameters:

  • df (pd.DataFrame): DataFrame to export
  • filepath (str | Path): Output file path
  • prompt (str, optional): AI transformation prompt
  • api_key (str, optional): OpenRouter API key
  • model (str, optional): AI model to use

read_pdf(filepath, prompt, columns=None, api_key=None, model=None)

Extract structured data from PDF files.

Parameters:

  • filepath (str | Path): Path to the PDF file
  • prompt (str): Description of what data to extract
  • columns (List[str], optional): Column names to extract
  • api_key (str, optional): OpenRouter API key
  • model (str, optional): AI model to use

Returns: pandas DataFrame

read_image(filepath, prompt, columns=None, api_key=None, model=None)

Extract structured data from image files.

Parameters:

  • filepath (str | Path): Path to the image file
  • prompt (str): Description of what data to extract
  • columns (List[str], optional): Column names to extract
  • api_key (str, optional): OpenRouter API key
  • model (str, optional): AI model to use

Returns: pandas DataFrame

read_audio(filepath, prompt, columns=None, api_key=None, model=None)

Extract structured data from audio files.

Parameters:

  • filepath (str | Path): Path to the audio file
  • prompt (str): Description of what data to extract
  • columns (List[str], optional): Column names to extract
  • api_key (str, optional): OpenRouter API key
  • model (str, optional): AI model to use

Returns: pandas DataFrame

read_webpage(url, prompt, columns=None, api_key=None, model=None)

Extract structured data from web pages.

Parameters:

  • url (str): URL of the webpage
  • prompt (str): Description of what data to extract
  • columns (List[str], optional): Column names to extract
  • api_key (str, optional): OpenRouter API key
  • model (str, optional): AI model to use

Returns: pandas DataFrame

read_video(filepath, prompt, from_='both', columns=None, api_key=None, model=None, sample_rate=30)

Extract structured data from video files.

Parameters:

  • filepath (str | Path): Path to the video file
  • prompt (str): Description of what data to extract
  • from_ (str | List[str]): Source to extract from - 'pics', 'audios', or 'both'
  • columns (List[str], optional): Column names to extract
  • api_key (str, optional): OpenRouter API key
  • model (str, optional): AI model to use
  • sample_rate (int): Frame sampling rate (default: 30)

Returns: pandas DataFrame

to_summarized_csv(df, filepath, prompt=None, api_key=None, model=None, **kwargs)

Export DataFrame to CSV with optional AI-powered summarization.

Parameters:

  • df (pd.DataFrame): DataFrame to export
  • filepath (str | Path): Path to save the CSV file
  • prompt (str, optional): Prompt to transform/summarize data
  • api_key (str, optional): OpenRouter API key
  • model (str, optional): AI model to use
  • **kwargs: Additional arguments for pd.DataFrame.to_csv()

to_summarized_excel(df, filepath, prompt=None, sheet_name="Sheet1", api_key=None, model=None, **kwargs)

Export DataFrame to Excel with optional AI-powered summarization.

to_summarized_json(df, filepath, prompt=None, api_key=None, model=None, orient="records", **kwargs)

Export DataFrame to JSON with optional AI-powered summarization.

summarize_dataframe(df, prompt="Provide a summary of this data", api_key=None, model=None)

Generate an AI-powered summary of a DataFrame.

Returns: str (AI-generated summary)

Configuration

Environment Variables

  • OPENROUTER_API_KEY: Your OpenRouter API key

Cache Settings

The cache is stored in ~/.fundas/cache/ by default. You can configure:

  • Cache directory location
  • Time-to-live (TTL) for cache entries
  • Enable/disable caching

Performance Tips

  1. Use caching: Keep caching enabled (default) to avoid redundant API calls
  2. Specify columns: When you know what columns you need, specify them explicitly
  3. Choose the right model: Balance speed, cost, and accuracy by selecting appropriate models
  4. Batch operations: Process multiple files in sequence to leverage cache warming

License

MIT License - see LICENSE file for details

Contributing

Contributions are welcome! We appreciate bug fixes, new features, documentation improvements, and more.

Please see our Contributing Guide for details on:

  • Setting up your development environment
  • Coding standards and style guide
  • Testing requirements
  • Pull request process

Quick start:

  1. Fork the repository
  2. Create a feature branch
  3. Make your changes with tests
  4. Submit a pull request

Support

For issues and questions, please open an issue on GitHub.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

fundas-0.1.0.tar.gz (40.6 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

fundas-0.1.0-py3-none-any.whl (15.7 kB view details)

Uploaded Python 3

File details

Details for the file fundas-0.1.0.tar.gz.

File metadata

  • Download URL: fundas-0.1.0.tar.gz
  • Upload date:
  • Size: 40.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for fundas-0.1.0.tar.gz
Algorithm Hash digest
SHA256 1cdc3fdf9c1c98e8bc57708bfaa45a8781567626a1aa5f6e8d80d0c2fa715aca
MD5 8da6be0361829de4dadf4324f40763c6
BLAKE2b-256 5c788dd0311314b7e122cc3b61ea88912b7d31723fae5f41cfe7a21ccf09ab51

See more details on using hashes here.

Provenance

The following attestation bundles were made for fundas-0.1.0.tar.gz:

Publisher: publish.yml on AMSeify/fundas

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file fundas-0.1.0-py3-none-any.whl.

File metadata

  • Download URL: fundas-0.1.0-py3-none-any.whl
  • Upload date:
  • Size: 15.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for fundas-0.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 4b21898442cdc643d6d77075ed544fb9e6413bee11d76319d7f8c61182be98f3
MD5 6d0302dccba46b7a99bc8cf507e932d2
BLAKE2b-256 4a61a37bbfd0349f7cec47e0df11351ec2f224623751f6515620bfb34e1db1b7

See more details on using hashes here.

Provenance

The following attestation bundles were made for fundas-0.1.0-py3-none-any.whl:

Publisher: publish.yml on AMSeify/fundas

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page