Offline local-LLM terminal app for Jetson and edge Linux: chat with on-device models, run agent tools, and manage context safely.
Project description
OpenJet
open-jet is an offline-first agent runtime for Jetson-class and other edge Linux systems with tight memory budgets.
It is built for cases where the hard part is not just running a local model, but keeping the agent useful under constrained RAM, limited context windows, interrupted sessions, and hardware-specific failure modes.
Quickstart
pip install open-jet
OpenJet does not bundle an inference runtime. By default it uses llama-server
from llama.cpp, though setup can
also configure SGLang or TensorRT-LLM instead.
For the default llama.cpp path, you need:
- the
open-jetPython package - a working
llama-serverbinary - a local model file such as a
.gguf
Then run:
open-jet --setup
open-jet
Installation
1. Install the Python package
pip install open-jet
2. Install llama-server from llama.cpp
llama.cpp setup is required for the default OpenJet runtime. If you choose
SGLang or TensorRT-LLM in open-jet --setup, follow the runtime-specific
model/runtime setup for those backends instead.
Jetson:
git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
mkdir build && cd build
cmake .. \
-DGGML_CUDA=ON \
-DCMAKE_CUDA_ARCHITECTURES=87 \
-DGGML_CUDA_FA_ALL_QUANTS=ON
cmake --build . --target llama-server -j$(nproc)
Linux x86 + NVIDIA:
git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
mkdir build && cd build
cmake .. -DGGML_CUDA=ON -DGGML_CUDA_FA_ALL_QUANTS=ON
cmake --build . --target llama-server -j$(nproc)
CPU-only:
git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
mkdir build && cd build
cmake ..
cmake --build . --target llama-server -j$(nproc)
OpenJet looks for llama-server on PATH first, then at
~/llama.cpp/build/bin/llama-server.
3. Have a local model ready
You need a model reference for the selected runtime:
llama.cpp: typically a local.gguffile, or an Ollama-backed model selected during setupSGLang: a local model directory or supported HF model idTensorRT-LLM: a local model directory or supported HF model id
4. Run setup
open-jet --setup
The setup flow guides you through:
- hardware detection/profile
- model source selection
- model path or download choice
- context window size
- GPU offload configuration
After setup:
open-jet
What it provides
- bounded-context local chat with your on-device model
- automatic context condensing under pressure
- session resume and harness state recovery
- OpenTelemetry instrumentation with collector export
- hardware-aware runtime setup for Jetson and edge Linux
- controlled tool use and slash commands
- Python SDK access to the same backend
Why OpenJet exists
Most local LLM tools stop at “chat with a model on your box.” That breaks down on edge hardware.
OpenJet is built around:
- limited prompt memory on-device
- interrupted work and session recovery
- deterministic tool and approval boundaries
- real traces for evaluation
- reliable operator workflows on constrained local models
Docs
- Quickstart
- Installation
- Runtime: llama.cpp
- Runtime: SGLang
- Runtime: TensorRT-LLM
- Usage: CLI
- Usage: Slash commands
- Usage: Workflow harness
- Usage: Session state and logging
- Telemetry
- Python SDK
- Deployment: Jetson
- Deployment: Linux x86 + NVIDIA
- Deployment: CPU-only
- Configuration
- Licensing
License
AGPL-3.0-only, with commercial licensing available under separate terms.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distributions
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file open_jet-0.1.12-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.whl.
File metadata
- Download URL: open_jet-0.1.12-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.whl
- Upload date:
- Size: 11.3 MB
- Tags: CPython 3.10, manylinux: glibc 2.17+ ARM64
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.10.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
6d0c2b470a1dc7b9513db765a06713875cd3da460e7e5fb814b101bcd25fb8f5
|
|
| MD5 |
84346cc7755f5e5d8ab9c0aa72a14087
|
|
| BLAKE2b-256 |
45e9718ac2afa82c49962de7bef66ad1b8f1bae384a37508ee7c7dd97f7c2008
|