Local agentic AI CLI powered by llama.cpp
Project description
llama-agentic
A local agentic AI CLI powered by llama.cpp. It runs on your machine, uses GGUF models, and exposes tools for files, shell commands, git, web access, memory, plugins, and MCP servers.
Think of it as a terminal coding agent driven by your local model instead of a hosted API.
You: Refactor the auth module to use JWT, run the tests, and commit
Agent: ⚙ view_file ✓ Read agent/auth.py
⚙ edit_file ✓ Applied JWT changes
⚙ run_shell ✓ pytest tests/test_auth.py
⚙ git_commit ✓ refactor: replace session auth with JWT
Done — all tests pass and changes are committed.
What it does
- Runs a ReAct loop: reason, choose a tool, observe results, repeat
- Works with local llama.cpp models through the OpenAI-compatible server API
- Includes built-in tools for files, editing, shell, Python, git, web, and memory
- Supports MCP servers so you can add GitHub, databases, browsers, Slack, and more
- Loads LLAMA.md automatically for project-specific context
- Supports persistent memory, session save/load, watch mode, and plugin loading
Documentation
The repo docs are the primary guide set:
| Guide | Description |
|---|---|
| Documentation Index | Overview of all guides |
| Getting Started | Install, setup, download a model, first run |
| User Guide | REPL usage, sessions, memory, watch mode |
| Tools Reference | Built-in tools and examples |
| Configuration | Environment variables and config hierarchy |
| Plugin Development | Add custom tools |
| MCP Integration | Configure and use MCP servers |
Requirements
| Requirement | Version | Install |
|---|---|---|
| macOS or Linux | macOS 12+ / Ubuntu 22+ | — |
| Python | 3.11+ | python.org |
| llama.cpp | latest | brew install llama.cpp |
| uv | latest | curl -LsSf https://astral.sh/uv/install.sh | sh |
Apple Silicon works well with Metal GPU offload by default.
Installation
From PyPI
pip install llama-agentic
# or
uv tool install llama-agentic
From source
git clone https://github.com/minrahim1999/llama-agentic.git
cd llama-agentic
uv tool install --editable .
Verify
llama-agent --help
Quick Start
1. Run first-time setup
llama-agent
This creates ~/.config/llama-agentic/config.env and can help detect llama-server, choose settings, and offer a starter model download.
2. Download a model
llama-agent download
llama-agent download qwen2.5-coder-7b
Recommended model:
llama-agent download qwen2.5-coder-7b
3. Start or verify the server
llama-agent doctor
llama-agent autostart enable
llama-agent autostart start
Or start it manually:
./scripts/start_server.sh /path/to/model.gguf
4. Generate project context
cd your-project
llama-agent --init
5. Start a session
llama-agent
llama-agent --task "Find and fix the failing tests"
Common Commands
llama-agent # interactive REPL
llama-agent --task "review the latest changes" # one-shot task
llama-agent --resume sessions/chat_2026-01-15.json
llama-agent doctor # environment checks
llama-agent download qwen2.5-coder-7b # download a model
llama-agent mcp list # list configured MCP servers
Common REPL commands:
/help/init/refresh/add <glob>/tools/memory/sessions/cost/exit
See docs/user-guide.md for the full command list.
Key Features
- Diff-aware editing:
edit_filepreviews changes before writing and keeps.bakbackups - Confirmation-gated actions: destructive tools require approval unless
UNSAFE_MODE=true - Persistent memory: store facts across sessions
- Session management: save, load, resume, and inspect history
- MCP integration: dynamically register tools from external MCP servers
- Plugin system: drop
.pyfiles intoplugins/to extend the agent .llamaignoresupport: block reads and writes to protected paths
Development
uv sync --dev
uv run pytest tests/ -v --tb=short
uv run ruff check agent/ tests/
uv build
Project layout and development conventions are documented in AGENTS.md.
License
MIT. See LICENSE.
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file llama_agentic-0.2.1.tar.gz.
File metadata
- Download URL: llama_agentic-0.2.1.tar.gz
- Upload date:
- Size: 170.7 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
ca37ee6d74f3055d20b6a7c391ca92b6f30a93646897579c2979f074f182da51
|
|
| MD5 |
34a2ef46c751d6ba1c551fd30fb492d2
|
|
| BLAKE2b-256 |
e779c761394215d73308c357ed4eedc14a38c094e01eecc0ee3947d02a37e54e
|
Provenance
The following attestation bundles were made for llama_agentic-0.2.1.tar.gz:
Publisher:
publish.yml on minrahim1999/llama-agentic
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
llama_agentic-0.2.1.tar.gz -
Subject digest:
ca37ee6d74f3055d20b6a7c391ca92b6f30a93646897579c2979f074f182da51 - Sigstore transparency entry: 1181149118
- Sigstore integration time:
-
Permalink:
minrahim1999/llama-agentic@86e236fc804f17eb62617d6a52eda90820e8c1b1 -
Branch / Tag:
refs/tags/v0.2.1 - Owner: https://github.com/minrahim1999
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@86e236fc804f17eb62617d6a52eda90820e8c1b1 -
Trigger Event:
push
-
Statement type:
File details
Details for the file llama_agentic-0.2.1-py3-none-any.whl.
File metadata
- Download URL: llama_agentic-0.2.1-py3-none-any.whl
- Upload date:
- Size: 55.9 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
ef925585cabe66bc04d3a266a48cbdfa766867d19c5d0677e6e134bb98bf3ff1
|
|
| MD5 |
2a6bb3e524242b514371e3c5be099f4a
|
|
| BLAKE2b-256 |
b035262a9b0c0e69ac5efc83e50ac9cb3ef1143a47820cc7fe1d6f5a3f67deb0
|
Provenance
The following attestation bundles were made for llama_agentic-0.2.1-py3-none-any.whl:
Publisher:
publish.yml on minrahim1999/llama-agentic
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
llama_agentic-0.2.1-py3-none-any.whl -
Subject digest:
ef925585cabe66bc04d3a266a48cbdfa766867d19c5d0677e6e134bb98bf3ff1 - Sigstore transparency entry: 1181149135
- Sigstore integration time:
-
Permalink:
minrahim1999/llama-agentic@86e236fc804f17eb62617d6a52eda90820e8c1b1 -
Branch / Tag:
refs/tags/v0.2.1 - Owner: https://github.com/minrahim1999
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@86e236fc804f17eb62617d6a52eda90820e8c1b1 -
Trigger Event:
push
-
Statement type: