LITS: LLM Inference via Tree Search — modular agentic reasoning framework.
Project description
LiTS — Language Inference via Tree Search
A modular Python framework for LLM reasoning and planning with tree search (e.g., MCTS, BFS and other custom search algorithms) and chain reasoning (e.g.,ReAct).
Why LiTS?
| Concern | Challenge | LiTS Solution |
|---|---|---|
| Reusability | Reimplementing search algorithms for each new task | Task-agnostic data structures (Action → Step → State → Node) that hide search procedures from task-specific logic |
| Extensibility | Adding new tasks requires modifying many files | Modular components (Policy, Transition, RewardModel) + decorator-based registry — add a task by registering prompts and a transition |
| Observability | Tree search is expensive and hard to debug | Built-in InferenceLogger tracks token usage at component, instance, and search-phase levels; incremental checkpointing for fault tolerance |
Installation
pip install lits-llm # from PyPI
# or
pip install -e . # editable install from source
Requires Python >= 3.11.
Quick Start — 5-Minute Demo
Three steps: install, configure an LLM, run a search and see the artifacts.
1. Install
pip install lits-llm
2. Configure an LLM provider
LiTS needs an LLM for policy (action generation) and reward (scoring). Pick one provider:
OpenAI — set OPENAI_API_KEY:
export OPENAI_API_KEY="sk-..."
MODEL="openai/gpt-4o-mini"
AWS Bedrock — configure AWS credentials (SSO or env vars):
aws sso login --profile default # or export AWS_ACCESS_KEY_ID / AWS_SECRET_ACCESS_KEY
MODEL="bedrock/anthropic.claude-3-5-sonnet-20240620-v1:0"
Groq (free tier available) — set GROQ_API_KEY:
export GROQ_API_KEY="gsk_..."
MODEL="groq/llama-3.1-8b-instant"
Local HuggingFace — no API key needed (for testing only; prefer GPU or Apple Silicon for larger LLMs):
MODEL="Qwen/Qwen2.5-0.5B-Instruct" # auto-downloads from HuggingFace
# pass device via --search-arg: device=cuda (default), device=mps (Apple Silicon), or device=cpu
Any OpenAI-compatible API (vLLM, Together AI, Fireworks, etc.):
export OPENAI_API_KEY="your-key" # use "EMPTY" for local vLLM
export OPENAI_API_BASE="http://localhost:8000/v1" # vLLM / provider's base URL
MODEL="openai/meta-llama/Llama-3-8B-Instruct"
3. Run MCTS on a math problem
Save this as my_benchmark.py (the .py extension matters) in your working directory:
from lits.registry import register_dataset
@register_dataset("my_math", task_type="language_grounded")
def load_my_math(**kwargs):
return [
{
"question": (
"The proper divisors of 12 are 1, 2, 3, 4 and 6. "
"A proper divisor of an integer $N$ is a positive divisor of $N$ "
"that is less than $N$. What is the sum of the proper divisors "
"of the sum of the proper divisors of 284?"
),
"answer": "284",
}
]
Then run MCTS:
lits-search --include my_benchmark \
--dataset my_math \
--policy-model "$MODEL" \
--search-arg roll_out_steps=2 n_iters=50 force_terminating_on_depth_limit=false n_actions=3 max_steps=10 \
-o demo_results --override
4. Evaluate
lits-eval --result_dir demo_results
What you should see
demo_results/
├── checkpoints/ # Intermediate tree states per iteration
├── terminal_nodes/ # All terminal nodes found
├── config.json # Full config (reproducible)
├── execution.log # Execution log
├── inferencelogger.log # Per-call token usage with component/phase tags
├── llm_calls.jsonl # Raw LLM call records
└── eval.log # Accuracy + inference usage report
Validate config without LLM calls (no API key needed)
lits-search --include my_benchmark \
--dataset my_math --dry-run
This prints the resolved components, dataset info, and first example — useful for checking your setup before a real run.
CLI Commands
lits-search # Run tree search (MCTS, BFS)
lits-eval # Evaluate tree search results
lits-chain # Run chain agents (ReAct, EnvChain)
lits-eval-chain # Evaluate chain results
🎬 2.5-Minute Demo Video
See the full walkthrough — MCTS on math, Crosswords (environment-grounded), and evaluation:
👉 https://youtu.be/nRGX43YrR3I
The commands demonstrated in the video are listed below for direct copy-paste.
More CLI Examples
All commands below assume cd demos and a configured $MODEL (see Quick Start above).
MCTS on MATH500
lits-search --include lits_benchmark.math_qa \
--dataset math500 \
--policy concat --transition concat --reward generative \
--search-arg n_iters=50 n_actions=3 max_steps=10 \
--var limit=5
Swap to RAP (different components, same algorithm)
lits-search --include lits_benchmark.math_qa lits_benchmark.formulations.rap \
--dataset math500 \
--policy rap --transition rap --reward rap \
--search-arg n_iters=10 n_confidence=3
Swap to BFS (different algorithm, same components)
lits-search --include lits_benchmark.math_qa \
--dataset math500 \
--cfg search_algorithm=bfs \
--policy concat --transition concat --reward generative \
--search-arg roll_out_steps=2 n_actions=3 max_steps=10
Environment-grounded task (BlocksWorld)
lits-search --include lits_benchmark.blocksworld \
--dataset blocksworld \
--transition blocksworld \
--search-arg max_steps=6 n_iters=50
Tool-use task (MapEval-SQL)
lits-search --include lits_benchmark.mapeval \
--dataset mapeval-sql
No component flags needed — the framework auto-selects tool-use components.
Quick Start — Python API
Tree search algorithms are class-based, inheriting from BaseTreeSearch:
from lits.agents.tree.mcts import MCTSSearch, MCTSConfig
from lits.lm import get_lm
# Load model
model = get_lm("bedrock/us.anthropic.claude-3-5-haiku-20241022-v1:0")
# Configure search
config = MCTSConfig(
max_steps=10,
n_actions=3,
n_iters=50,
)
# Create search instance with components
search = MCTSSearch(
config=config,
policy=policy, # generates candidate actions
world_model=transition, # executes actions, produces new states
reward_model=reward, # evaluates action quality
)
# Run search
result = search.run(query="What is 25 * 17?", query_idx=0)
# Extract answers from terminal nodes
from lits.agents.tree.common import extract_answers_from_terminal_nodes
vote_answers, answer_rewards, best_node, trace = extract_answers_from_terminal_nodes(
terminal_nodes_collected=result.terminal_nodes_collected,
retrieve_answer=retrieve_answer_fn,
question="What is 25 * 17?"
)
ReAct Agent (tool use)
from lits.agents import create_tool_use_agent
agent = create_tool_use_agent(tools=tool_list, max_iter=50)
state = agent.run(query="Find restaurants near Sydney Opera House")
Supported LLM Providers
from lits.lm import get_lm
model = get_lm("bedrock/us.anthropic.claude-3-5-haiku-20241022-v1:0") # AWS Bedrock
model = get_lm("openai/gpt-4") # OpenAI
model = get_lm("Qwen/Qwen2.5-0.5B-Instruct", device="cuda") # HuggingFace
model = get_lm("groq/llama-3.1-8b-instant") # Groq
model = get_lm("tgi:///meta-llama/Meta-Llama-3-8B") # TGI
Architecture
Three core component abstractions compose into agents:
Policy → generates candidate actions from states
Transition → executes actions, produces new states
RewardModel → evaluates action quality (optional)
Search frameworks bundle these with an algorithm:
| Framework | Algorithm | Components |
|---|---|---|
| ReST-MCTS* | MCTS | ConcatPolicy + ConcatTransition + GenerativePRM |
| RAP | MCTS | RAPPolicy + RAPTransition + RapPRM |
| ToT-BFS | BFS | ConcatPolicy + ConcatTransition + GenerativePRM |
Extending with Custom Components
Register components via decorators — no core code changes needed:
from lits.components.registry import register_transition, register_dataset
@register_transition("my_domain")
class MyTransition(Transition):
def step(self, example, state, action, **kwargs):
...
def is_terminal(self, state, example, **kwargs):
...
@register_dataset("my_dataset", task_type="env_grounded")
def load_my_dataset(**kwargs):
...
Then use via CLI:
lits-search --include my_package \
--dataset my_dataset --transition my_domain
Task Types
| Task Type | State Space | Examples |
|---|---|---|
language_grounded |
Text context | Math reasoning (GSM8K, MATH500) |
env_grounded |
Symbolic/physical state | BlocksWorld, Crosswords |
tool_use |
Context + tool state | SQL queries, web search, APIs |
Project Structure
lits/ # Core framework
├── agents/ # MCTS, BFS, ReAct, EnvChain
├── components/ # Policy, Transition, RewardModel
├── lm/ # Multi-provider LLM interface
├── structures/ # State, Action, Step, Node
├── cli/ # CLI entry points
├── eval/ # Evaluation utilities
└── tools/ # Tool implementations
Documentation
License
Apache License 2.0
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file lits_llm-0.3.2.tar.gz.
File metadata
- Download URL: lits_llm-0.3.2.tar.gz
- Upload date:
- Size: 250.2 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
3b30a741dd32c94592ce845cc08abfaa9cf661f1b67b5fbf3198cfebdf642078
|
|
| MD5 |
ddf6992d5a7b05681664f7448bcd9021
|
|
| BLAKE2b-256 |
30c36edb86d52246888351efe9cc2c20726eafcb3b17fecb648a8b567c1a2607
|
Provenance
The following attestation bundles were made for lits_llm-0.3.2.tar.gz:
Publisher:
publish.yml on xinzhel/lits-llm
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
lits_llm-0.3.2.tar.gz -
Subject digest:
3b30a741dd32c94592ce845cc08abfaa9cf661f1b67b5fbf3198cfebdf642078 - Sigstore transparency entry: 1004624915
- Sigstore integration time:
-
Permalink:
xinzhel/lits-llm@4f522bb7bf5d5bfd68c42649efe44c566dfa039c -
Branch / Tag:
refs/tags/v0.3.2 - Owner: https://github.com/xinzhel
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@4f522bb7bf5d5bfd68c42649efe44c566dfa039c -
Trigger Event:
push
-
Statement type:
File details
Details for the file lits_llm-0.3.2-py3-none-any.whl.
File metadata
- Download URL: lits_llm-0.3.2-py3-none-any.whl
- Upload date:
- Size: 298.0 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
7100da5a3dbe3806fb574d6f08db21e817b77e92fddddf3227511c08f4b0e553
|
|
| MD5 |
eb8e88944ea6ecba6d4e085c76eb1324
|
|
| BLAKE2b-256 |
a45c588f5373e8de16e5c20c5451603d0c1f7798c00e2956be85c95d6ad70001
|
Provenance
The following attestation bundles were made for lits_llm-0.3.2-py3-none-any.whl:
Publisher:
publish.yml on xinzhel/lits-llm
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
lits_llm-0.3.2-py3-none-any.whl -
Subject digest:
7100da5a3dbe3806fb574d6f08db21e817b77e92fddddf3227511c08f4b0e553 - Sigstore transparency entry: 1004624917
- Sigstore integration time:
-
Permalink:
xinzhel/lits-llm@4f522bb7bf5d5bfd68c42649efe44c566dfa039c -
Branch / Tag:
refs/tags/v0.3.2 - Owner: https://github.com/xinzhel
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@4f522bb7bf5d5bfd68c42649efe44c566dfa039c -
Trigger Event:
push
-
Statement type: