Skip to main content

AI-native machine learning tool that bridges natural language questions and predictive models using declarative Arc-Graph schemas

Project description

                         ▓▓▓▓▓╗   ▓▓▓▓▓▓╗    ▓▓▓▓▓▓╗
                        ▓▓╔══▓▓╗  ▓▓╔══▓▓╗  ▓▓╔════╝
                        ▓▓▓▓▓▓▓║  ▓▓▓▓▓▓╔╝  ▓▓║
                        ▓▓╔══▓▓║  ▓▓╔══▓▓╗  ▓▓║
                        ▓▓║  ▓▓║  ▓▓║  ╚▓▓╗ ╚▓▓▓▓▓▓╗
                        ╚═╝  ╚═╝  ╚═╝   ╚═╝  ╚═════╝
                         From Question to Prediction

Arc is an AI-native machine learning tool to enable machine learning accessible to everyone, from data analysts to seasoned ML engineers. It bridges the gap between natural language questions and predictive models, transforming how you work with data.

For Business Users & Analysts: Have you ever wanted to predict customer churn or forecast sales without writing complex code? With Arc, you can. Use plain English to explore data, build models, and get predictions. Arc's AI handles the complexity for you.

For Machine Learning Engineers & Data Scientists: Arc streamlines your ML workflow. Instead of boilerplate PyTorch, TensorFlow, or JAX, you use a declarative, AI-native approach. Arc translates your intent into a portable and declarative ML schema, letting you focus on high-level architecture and rapid iteration.

💡 How It Works

Arc is built on three foundational pillars:

  • Arc-Graph - Declarative YAML schema for ML model architecture and training configuration
  • Arc-Pipeline - Declarative YAML schema for feature engineering and data processing pipelines
  • Arc-Knowledge - Curated best practices and patterns (extendable via ~/.arc/knowledge/)

When you give a command in natural language, Arc's AI consults the Arc-Knowledge to generate optimal specifications:

Your Question → Arc AI (+ Arc-Knowledge) → Arc-Graph + Arc-Pipeline → Training → Predictions

The Arc-Knowledge includes:

  • Data loading patterns (CSV, Parquet, JSON, S3, Snowflake)
  • Feature engineering techniques (normalization, encoding, splits)
  • Model architectures (DCN, MMOE, Transformers, etc.)
  • Best practices and proven patterns

Extensibility: Add your own patterns and project-specific knowledge to ~/.arc/knowledge/ to guide Arc's AI for your use case.

This approach provides the best of both worlds:

  • Simplicity: A conversational interface - just describe what you want
  • Power & Portability: Declarative, version-controlled YAML files that run anywhere PyTorch runs
  • Transparency: Human-readable specifications you can review, modify, and share
  • Customizable: Extend the Arc-Knowledge with your own patterns and practices

✨ Key Features

  • 🤖 Natural Language to Model - Go from a question in plain English to a trained predictive model without writing a single line of ML code.
  • 📜 Declarative Schemas (Arc-Graph & Arc-Pipeline) - Arc's AI generates complete specifications in human-readable YAML. Arc-Graph defines your model architecture, Arc-Pipeline defines your feature engineering workflows. You review and approve; the AI handles the implementation.
  • 🧠 Extensible Arc-Knowledge - Built-in curated knowledge of ML best practices, data patterns, and model architectures. Extend it with your own project-specific patterns in ~/.arc/knowledge/ to customize Arc's AI for your domain.
  • 🗄️ Unified Data & ML with SQL - Connect your data sources via standard SQL. Arc manages your ML assets (models, features, results) in a dedicated database that you can query using standard SQL.
  • End-to-End & Portable - Arc-Graph and Arc-Pipeline files contain your complete ML workflow, ensuring train/serve parity and making your models easy to version, share, and reproduce.
  • 🎯 Smart & Interactive - AI-powered guidance and a user-friendly interactive mode are enabled by default to help you get started quickly.

🚀 Quick Start

Installation

# Clone and install the project
git clone https://github.com/non-linear-ai/arc
cd arc

Your First Model

Let's build a diabetes prediction model in 3 simple steps:

1. Configure Your API Key (One-Time Setup)

Start Arc and configure your API (saved to ~/.arc/, only needed once):

uv run arc chat
> /config

Example configuration:

◇ Configuration
  API Key            ********
  Base URL           https://api.deepseek.com/v1
  Model              deepseek-chat

Note: Arc works with agentic and OpenAI API-compatible models, such as Gemini, OpenAI GPT models, or Anthropic Sonnet models.

2. Ask Arc to Build Your Model

Simply describe what you want:

Download the Pima Indians Diabetes dataset and build a model to predict diabetes from patient health metrics

Arc will:

  • ✅ Download the dataset automatically
  • ✅ Analyze the data and engineer features
  • ✅ Generate an Arc-Graph model specification
  • ✅ Train and evaluate the model
  • ✅ Launch TensorBoard locally to monitor training curves and metrics in real-time
  • ✅ Show you predictions and performance metrics

3. Explore Your Results

Your model is trained! Use /sql SHOW TABLES and other SQL commands to explore your data and predictions. Check the logs for the TensorBoard URL to view training curves and metrics.

What Just Happened?

Arc generated an Arc-Graph specification that looks like this:

# Arc-Graph: Model Architecture
inputs:
  patient_data:
    dtype: float32
    shape: [null, 8]
    columns: [pregnancies, glucose, blood_pressure, skin_thickness,
              insulin, bmi, diabetes_pedigree, age]

graph:
  - name: classifier
    type: torch.nn.Linear
    params:
      in_features: 8
      out_features: 1
    inputs:
      input: patient_data

  - name: sigmoid
    type: torch.nn.Sigmoid
    inputs:
      input: classifier.output

outputs:
  prediction: sigmoid.output

This Arc-Graph specification is:

  • Human-readable - You can understand and modify it
  • Portable - Runs anywhere PyTorch runs
  • Versionable - Track changes in Git
  • Reproducible - Guarantees train/serve parity

For more details, see the Arc-Graph Specification Guide.

📚 Documentation

📖 Complete Documentation - Start here for comprehensive guides, examples, and API reference.

Quick Links

Getting Started:

Core Concepts:

User Guides:

Integrations:

  • AWS S3 - Load data from S3 buckets
  • Snowflake - Query Snowflake warehouses

For Contributors:

Development

Want to contribute? See CONTRIBUTING.md for guidelines.

Quick Dev Setup

# Install uv
curl -LsSf https://astral.sh/uv/install.sh | sh

# Clone and install
git clone https://github.com/non-linear-ai/arc
cd arc
uv sync --dev

# Run tests
uv run pytest

# Format and lint
uv run ruff format .
uv run ruff check . --fix

For detailed instructions, see Development Setup Guide.

Project Configuration

Project settings are in pyproject.toml. For API configuration, see Configuration Guide.

For S3 and Snowflake setup, see:

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

nl_arc-0.1.1.tar.gz (383.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

nl_arc-0.1.1-py3-none-any.whl (448.3 kB view details)

Uploaded Python 3

File details

Details for the file nl_arc-0.1.1.tar.gz.

File metadata

  • Download URL: nl_arc-0.1.1.tar.gz
  • Upload date:
  • Size: 383.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.8.22

File hashes

Hashes for nl_arc-0.1.1.tar.gz
Algorithm Hash digest
SHA256 ba1b2c998e3e4d771e1417eec18f46faedc4ce52a23d8397ec10f7bb5bef1a32
MD5 760b61ae6cb2b0d4de9cb201de0aa8a2
BLAKE2b-256 962bde04ee62b3a03b732564e3f16a312ed628355d50d4cf8044c724b6730fc5

See more details on using hashes here.

File details

Details for the file nl_arc-0.1.1-py3-none-any.whl.

File metadata

  • Download URL: nl_arc-0.1.1-py3-none-any.whl
  • Upload date:
  • Size: 448.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.8.22

File hashes

Hashes for nl_arc-0.1.1-py3-none-any.whl
Algorithm Hash digest
SHA256 da1d87ff5860dc16fcf2dee31683709cd3e72f67167362c3dfd250256b835750
MD5 e8cf987158dd5c44eed1f5991e23a46e
BLAKE2b-256 de5b4d1e5a0462a597d0223917cd671adaafc2478be70f945f22980115b01723

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page