Skip to main content

AI coding assistant agent for 1B–7B local models (Ollama, LMStudio, llama.cpp). Terminal REPL with file editing, project map, agents, scripts, and parallel multi-model queries.

Project description

1bcoder

AI coding assistant agent for 1B–7B local models running locally via Ollama, LMStudio, or LiteLLM.


Core idea: 1B models hallucinate badly when asked to rewrite large blocks of code. 1bcoder works around this by keeping changes small and structured — the model outputs a single-line fix (LINE N: content) or a minimal SEARCH/REPLACE block, which the tool then applies with a diff preview before writing to disk.

Planning and navigation are externalized: plans live in .txt files, project structure is indexed into a searchable map — so the model never has to hold the whole codebase in its head.

Target: programmers running qwen2.5-coder:0.6b or llama3.2:1b on a 4 GB machine — offline, no cloud, no subscription. The tool does the heavy lifting so the model doesn't have to.


What 1B models are actually good at

1B models fail at open-ended tasks. They excel at bounded tasks — where the answer is almost fully determined by the input you give them. The key is to use 1bcoder's navigation tools to narrow context to exactly what the model needs, then ask a specific question.

Generate

Task How
Dockerfile / docker-compose Describe the stack, ask the model to write it
SQL script Load schema with /script apply SQLiteSchema.txt, ask for INSERT / SELECT / migration
Simple algorithm Bubble sort, binary search, Newton gradient — model knows these cold
Function stub → implementation Write the signature + docstring, ask the model to fill the body
Unit test for one function /read the function, ask for a test
Type annotations /read a Python or TypeScript file, ask to add types
Regex pattern Describe the pattern precisely, ask for the regex
Config file nginx server block, systemd service, GitHub Actions single-job workflow
Shell script / Makefile Repetitive structure the model handles reliably
argparse / click parser Very formulaic — model knows these patterns exactly
Port a function to another language Paste 10–20 lines of Python/Go/Java, ask for the equivalent

Explain

Task How
What does this function do? /read file.py 10-30 → ask; works even for unfamiliar languages
What does this error mean? /run python main.py → output injected automatically → ask
What is this project? /tree ctx → injects directory tree → ask for overview
What does this module do? /find ClassName -c ctx or /map find \ClassName -y → inject → ask
What changed after my edit? /map idiff → inject diff → ask what the structural change means
Explain this SQL / regex / config Paste the fragment, ask; no file context needed

The rule behind the list

These tasks share one property: the model never needs to search for information. You deliver exactly the right context using /read, /run, /find, /tree, or /map find — then the model executes a single well-defined subtask. Context under 2K tokens, one clear question, deterministic output format: this is where 1B models are reliable.

Tasks that require the model to decide what to look at — refactoring across files, debugging a multi-file interaction, writing a new feature from scratch — need 32B+ models with /agent advance.


Features

  • Plain terminal REPL — works in any shell, IDE terminal, or SSH session; status line before each prompt shows active model, disk size, quantization, native context limit, and context fill %
  • /read injects files without line numbers (clean text, ideal for notes.txt and structured data); /readln injects with line numbers (use before /fix or /patch when line references matter)
  • Command autocorrection — typos in command names, file paths, and keywords are detected and fixed automatically before execution, for both human input and agent actions
  • /tree [path] — display directory tree of the whole project or any subtree; ask to inject into context (or pass ctx to skip the prompt)
  • /find <pattern> — search filenames and file content with regex; supports -f/-c/-i/--ext flags; highlights matches, asks to inject results into context
  • AI proposes a one-line fix (/fix) or a SEARCH/REPLACE patch (/patch) — always shows a diff before applying
  • Apply AI code blocks directly with /edit <file> code (new/full file) or /patch <file> code (SEARCH/REPLACE from reply, no line numbers needed) — preferred for agent mode
  • <think> tag support — reasoning blocks shown in terminal by default; /think hide suppresses terminal display; /think include keeps reasoning in context for chained turns
  • Run shell commands and inject their output with /run
  • Save AI replies to files with /save (code-fence stripping, multiple files, append modes)
  • Session persistence/ctx save / /ctx load dump and restore full conversations; /ctx list browses the .1bcoder/ctx/ project context library; /ctx compact summarizes and compresses the context via AI; /ctx savepoint marks a position for rollback or selective compaction; /ctx clear N drops the last N messages
  • Scripts — reusable sequences of commands stored as .txt files, run step-by-step or fully automated
  • Script from history/script create ctx captures this session's commands into a reusable script automatically
  • Project map — scan any codebase into a searchable index (/map index), query it (/map find), trace call chains (/map trace), and diff changes (/map idiff) — now includes ORPHAN_DRIFT alert (dead code delta) and GHOST ALERT (deleted file that other files depended on)
  • Ask mode/ask <question> is an alias for /agent ask: a read-only research loop for 4B models that explores the project with tree/find/map tools, never edits files, auto-truncates large results to protect context
  • Agent mode/agent <task> runs an autonomous loop; stops when the model outputs plain text with no ACTION; after the loop a [s]ummary / [a]ll / [n]one prompt lets you pull agent results into main context
  • Named agents — define custom agents in .1bcoder/agents/<name>.txt (system prompt, tools, max_turns, aliases, on_done, params, before, gates); call with /agent <name> task or /<name> task directly; agent-scoped aliases active only during that run
  • /plan <goal> — planning agent: researches the project, writes a natural-language step-by-step plan to plan.txt; run /agent <task> file: plan.txt to execute it step by step
  • /fill — fill agent: reads NaN session variables, scans project for .var files and config files, sets each value automatically
  • Session variables{{name}} placeholders substituted in any command; save/load from .var files for offline reuse without loading files into context
  • Project config/config save persists session state (host, model, ctx, params, vars, procs) to .1bcoder/config.yml; /config save global saves to ~/.1bcoder/config.yml; on startup, the first config with auto: true (local → global) is applied automatically
  • Aliases — define command shortcuts with /alias /name = expansion (supports {{args}}); persisted in aliases.txt; loaded from global then project directory at startup and survive /clear
  • Backup/restore/bkup save rotates existing backups (file.bkupfile.bkup(1), file.bkup(2)…) so no snapshot is ever overwritten; /bkup restore always restores the latest
  • /tempctx — agent-internal context control: /tempctx N sets a private token budget, /tempctx cut removes oldest messages, /tempctx clear resets to system+task, /tempctx show prints size; only active inside an agent loop so agents can't touch the main context; also settable via params = agent_ctx = N in agent files
  • Agent proc hooksbefore = (runs before every LLM call, injects output as [context]) and gates = (runs after every reply, FAIL: retries the current plan step); enables supervised loops, convention enforcement, and hallucination checks without changing the model
  • /scan <file> <theme> — named agent that reads any file chunk by chunk, extracts info relevant to a theme, and builds a summary in .1bcoder/scan_result.txt; uses /tempctx cut between chunks so the agent context never overflows; result is injected into main context at the end
  • MCP support — connect external tool servers (filesystem, web, git, database, browser…) via the Model Context Protocol
  • Parallel queries — send prompts to multiple models simultaneously with /parallel; control context sent (--ctx/--last/--no-ctx) and route replies back into main context (ctx output) for sub-agent workflows
  • Command hooks/hook before|after <cmd> <script> runs a script before or after edit/patch/fix/insert; before hook cancels the command if the script is missing; {{file}} and {{range}} injected automatically
  • Switch model or host at runtime without restarting (/model gemma3:1b, /host openai://localhost:1234)
  • Model parameters/param temperature 0.2, /param enable_thinking false — sent with every request, auto-cast to correct type
  • Multi-provider — connect to Ollama, LMStudio, or LiteLLM using ollama:// / openai:// URL scheme; plain host defaults to Ollama

Quick install

Option 1 — PyPI (recommended)

pip install 1bcoder

On first launch, default agents, procs, scripts, profiles, and aliases are copied to ~/.1bcoder/ automatically. On upgrade (pip install --upgrade 1bcoder), new entries in aliases.txt and profiles.txt are merged in without overwriting your customisations.

Option 2 — Clone and install locally

git clone https://github.com/szholobetsky/1bcoder.git
cd 1bcoder
pip install -e .

When installed from source with pip install -e ., the default data files are not copied automatically. Run the included script once to populate ~/.1bcoder/:

deploy_bcoder_data.bat

This copies everything from _bcoder_data\ (agents, procs, aliases, profiles, scripts) to %USERPROFILE%\.1bcoder\. Re-run after pulling updates to sync new defaults.

Warning: This will overwrite any files you have customised in ~/.1bcoder\. Back up your changes before running.

Option 3 — Install directly from GitHub

pip install git+https://github.com/szholobetsky/1bcoder.git

Then run anywhere:

1bcoder

Requirements

Dependency Version
Python ≥ 3.10
Ollama any recent version
requests ≥ 2.28

Instead of Ollama, any OpenAI-compatible backend works: LMStudio, LiteLLM, or any /v1/chat/completions proxy.

Optional (for MCP servers):

  • Node.js + npx (for @modelcontextprotocol/* servers)
  • uv / uvx (for Python-based MCP servers)

Installation

1. Install Ollama and pull a model

# Install Ollama from https://ollama.com, then:
ollama pull llama3.2:1b       # fast, minimal RAM
ollama pull qwen2.5-coder:1b  # good for code

2. Clone and install 1bcoder

git clone <repo-url>
cd 1bcoder

# Install with pip (creates the `1bcoder` command)
pip install -e .

Running

# Using the installed command
1bcoder

# Or run directly
python chat.py

On startup, 1bcoder checks for a config with auto: true (local .1bcoder/config.yml first, then ~/.1bcoder/config.yml) and connects to the host and model stored there. If no config is found, it connects to local Ollama and prompts for a model. Use --model or --host to override.

CLI options

1bcoder [--host URL] [--model NAME] [--init] [--scriptapply SCRIPT] [--param KEY=VALUE]

--host URL              Host URL — supports ollama:// and openai:// schemes (default: from config or http://localhost:11434)
--model NAME            Model to use; overrides config (shows list if not available on host)
--init                  Create .1bcoder/ scaffold in the current directory
--scriptapply SCRIPT        Run a script file non-interactively, then exit
--param KEY=VALUE       Plan parameter substitution (repeatable)

Examples:

1bcoder --host http://192.168.1.50:11434
1bcoder --model qwen2.5-coder:1b
1bcoder --scriptapply my-fixes.txt --param file=calc.py --param range=1-4

Quick start

 ██╗██████╗        ██████╗ ██████╗ ██████╗ ███████╗██████╗
███║██╔══██╗      ██╔════╝██╔═══██╗██╔══██╗██╔════╝██╔══██╗
╚██║██████╔╝█████╗██║     ██║   ██║██║  ██║█████╗  ██████╔╝
 ██║██╔══██╗╚════╝██║     ██║   ██║██║  ██║██╔══╝  ██╔══██╗
 ██║██████╔╝      ╚██████╗╚██████╔╝██████╔╝███████╗██║  ██║
 ╚═╝╚═════╝        ╚═════╝ ╚═════╝ ╚═════╝ ╚══════╝╚═╝  ╚═╝

  model    : gemma3:1b [815M Q4_K 32K]
  host     : http://localhost:11434
  provider : ollama
  dir      : /home/user/myproject

  /help for all commands   /init to create .1bcoder/ folder
  Ctrl+C interrupts stream   /exit to quit

> /init
> /map index .
> /read main.py 1-20
> what does the divide() function do?
> /fix main.py 5-5 wrong operator

Command Reference

File operations

Command Description
/read <file> [file2 ...] [start-end] Inject file(s) into AI context without line numbers (clean text)
/readln <file> [file2 ...] [start-end] Same as /read but with line numbers — use before /fix or /patch
/insert <file> <line> Insert last AI reply before line N (full text)
/insert <file> <line> code Insert extracted code block from last AI reply before line N
/insert <file> <line> <text> Insert literal text directly, preserving indentation (e.g. /insert main.py 14 x = 1)
/edit <file> <line> Manually replace a single line
/edit <file> code Apply last AI reply (code block) to whole file — creates file if missing, diff before applying
/edit <file> <line> code Apply code block starting at <line> — creates file if missing
/edit <file> <start>-<end> code Apply code block replacing exactly lines startend
/save <file> [mode] Save last AI reply to a file
/bkup save <file> Save a backup as <file>.bkup; rotates existing backup to <file>.bkup(N)
/bkup restore <file> Replace <file> with its .bkup copy (always the latest)
/diff <file_a> <file_b> [-y] Show colored unified diff between two files; -y auto-injects into context

/save modes: overwrite (default), append-above / -aa, append-below / -ab, add-suffix, code

/diff main.py main.py.bkup          # colored diff, asks to inject into context
/diff v1/calc.py v2/calc.py -y      # auto-inject without confirmation
/save out.txt
/save out.txt add-suffix        # → out_1.txt, out_2.txt, …
/save main.py code              # strips ```python…``` wrapper
/save index.html style.css code # block 1 → index.html, block 2 → style.css

AI edits

Command Description
/fix <file> [start-end] [hint] AI proposes one-line fix, shows diff, asks to apply
/patch <file> [start-end] [hint] AI proposes SEARCH/REPLACE block, shows unified diff
/patch <file> code Apply SEARCH/REPLACE block from last AI reply (no new LLM call)

/fix is designed for 1B models — output is strictly constrained to LINE N: content. /patch works better with larger models (7B+) and can replace multiple consecutive lines. /patch <file> code is the preferred agent mode edit — the agent writes the SEARCH/REPLACE block in its reply, then calls /patch <file> code to apply it without needing line numbers.

When /patch fails to find the SEARCH text it shows a diagnostic diff — the SEARCH lines vs the nearest matching lines in the file — so you can see immediately what doesn't match (e.g. trailing commas, wrong indentation). It also detects no-op patches where SEARCH and REPLACE are identical.

/fix main.py
/fix main.py 2-2 wrong operator
/patch main.py 10-40 fix the loop logic
/patch main.py code

Shell

/run <command>

Runs any shell command and injects stdout + stderr into the AI context.

/run python main.py
/run pytest tests/ -x

Codebase navigation

/tree — directory structure

/tree                     whole project from current directory (depth 4)
/tree src                 subtree rooted at src/
/tree src/java/com -d 6   deep subtree, 6 levels
/tree static ctx          show and auto-inject into context (no prompt)

Displays a Unicode box-drawing tree. Skips .git, node_modules, __pycache__, .venv, and other noise automatically. When depth is cut, shows … (N entries) so you know something's there. After displaying results asks Add tree to context? [Y/n] — pass ctx to skip.

/find — search filenames and content

/find <pattern> [-f] [-c] [-i] [--ext <ext>] [ctx]
Flag Effect
(none) search filenames and file content
-f filenames only
-c content only
-i case-insensitive
--ext py restrict to .py files
ctx auto-inject results into context (no Y/n prompt)

Pattern is a full regex. Matches highlighted in the terminal output.

/find MyClass                    filenames + content
/find user_id -c -i              content only, case-insensitive
/find config --ext py ctx        .py files, inject automatically
/find \.connect\(                regex: literal .connect(

After showing results asks Add results to context? [Y/n] (suppressed when there are no matches).


Project map

The map command scans your project with language-agnostic regex, extracts definitions (classes, functions, endpoints, tables…) and cross-references between files. No external dependencies — pure regex, works for Python, Java, SQL, HTML, Terraform, YAML, and anything else.

/map index [path] [depth]                      — scan project, save to .1bcoder/map.txt
/map find [query] [-d N] [-y]                  — search the map and inject results into context
/map trace <id> [-d N] [-y]                    — BFS backwards: who depends on this identifier?
/map trace deps <id> [-d N] [-leaf] [-y]       — forward: what does this identifier depend on?
/map trace <start> <end> [-y]                  — shortest dependency path between two points
/map idiff [path] [depth]                      — re-index then show diff vs previous snapshot
/map diff                                      — show diff without re-indexing (safe to repeat)
/map keyword index                             — build keyword vocabulary from map.txt
/map keyword extract <text> [-f] [-a] [-n] [-c] — extract real identifiers from keyword.txt matching text/file

Partial / incremental indexing — for large codebases where a full re-scan is slow:

/map index sonar_core/src/main/java/org/sonar/core/util

When path is a subfolder of the working directory, 1bcoder:

  1. Scans only that subtree (fast)
  2. Adjusts all paths to be relative to the project root
  3. Saves a named segment file: .1bcoder/map_sonar_core_src_main_java_org_sonar_core_util.txt
  4. Patches map.txt in-place — removes stale blocks for that subtree, appends the fresh ones
  5. Backs up the previous map.txt to map.prev.txt before patching

This lets you re-index a changed module in seconds instead of hours.

/map find search syntax:

Token Where Effect
term filename include if filename contains term
!term filename exclude if filename contains term
\term child lines include block if any child line contains term
\!term child lines exclude entire block if any child contains term
-term child lines show ONLY child lines containing term
-!term child lines hide child lines containing term
-d 1 filenames only
-d 2 filenames + defines/vars (no links)
-d 3 full blocks (default)
-y skip "add to context?" confirmation
/map find register                       — files named *register*
/map find \register                      — files that define/link "register"
/map find register \register             — both: in name AND in children
/map find \register !mock                — has "register" in children, skip mock files
/map find auth \UserService -!deprecated -y
/map find \py -!import                   — show only non-import lines in .py files
/map find password -d 1                  — just filenames, no details
/map find models -d 2                    — filenames + defines/vars only

/map trace — three modes:

1. Backwards BFS (who depends on this?):

/map trace register -d 2

auth/routes.py  [defines register(ln:45)]
  ← import:register  app/__init__.py
  ← call:register    tests/test_auth.py

2. Forward deps (what does this depend on?):

/map trace deps UserService -leaf

deps (leaves): UserService(ln:5)  [UserService.java]
  UserEntity.java   [firstName, lastName, accountNumber]
  UserRepo.java     [findById, findByEmail, save]

3. Path between two points (shortest dependency chain):

/map trace AccountNumber UserController

path 1: AccountNumber(ln:15) → UserController

  AccountEntity.java
    ↓ import:AccountNumber
  UserDAO.java
    ↓ import:UserDAO
  UserController.java

  [Y]es add + next / [s]kip next / [n]o stop:

After each path:

  • Y — add to context, find next alternative path
  • s — skip (don't add), find next alternative path
  • l N or l then enter N — auto-collect the next N paths without prompting (e.g. l 10)
  • n — stop and inject everything collected so far

/map keyword — build and query a keyword vocabulary from the map:

/map keyword index

Scans map.txt, extracts all real code identifiers, saves to .1bcoder/keyword.txt as CSV (word, count, line_numbers). Run once after /map index.

/map keyword extract <text or file> [-f] [-a] [-n] [-c]

Finds identifiers from keyword.txt that match words in the given text or file. Output is always real identifiers from keyword.txt — never synthetic splits.

Flag Effect
(none) Exact match: rule matches rule only, not RuleIndex
-f Fuzzy subword match: splits both query and keyword into parts
-a Alphabetical order (default: order of appearance in text)
-n Show codebase count: RuleIndex(25)
-c Comma-separated output

Fuzzy subword rules (-f):

  • rule → matches RuleIndex, RuleName, RuleUpdater (all contain subword rule)
  • RuleIndex → matches RuleIndex only (requires both rule AND index)
  • RuleIndex does not match Rule (missing index) or Index (missing rule)
  • Combined identifiers are more specific — no false positives from their parts

Handles all naming conventions at match time: camelCase, PascalCase, snake_case, UPPER_SNAKE_CASE, kebab-case.

/map keyword extract "fix rule search performance" -f -c
→ RuleIndex, RuleSearch, RuleUpdater, SearchService

/map keyword extract notes.txt -f -n
→ RuleIndex(47)
  RuleName(23)
  SearchService(18)

/map idiff re-indexes and shows what changed — use after every edit:

/map idiff

[map diff]  map.prev.txt → map.txt

  calc.py
  - defines: subtract(ln:12)
  ! WARNING: 1 identifier(s) removed

ORPHAN_DRIFT = +1  [DEGRADATION]
  before: 3 orphans
  after:  4 orphans
  + subtract    ← calc.py

! GHOST ALERT — 1 deleted file(s) had active callers:
  ! services/UserService.java
    called: findUser, createUser

ORPHAN_DRIFT catches dead code accumulation and deleted caller files. GHOST ALERT catches deleted callee files that other files still depend on — the blind spot that structural diff alone misses.

Standalone tools (usable without 1bcoder):

  • map_index.py — scanner only: python map_index.py [path] [depth]
  • map_query.py — query only: python map_query.py find \register / python map_query.py trace register

Ask mode — read-only research for 4B models

/ask is an alias for /agent ask — it loads its system prompt, tools, and aliases from .1bcoder/agents/ask.txt (global install) or your project's local override. Designed for 4B models (nemotron, qwen3:4b, ministral3:3b). The model navigates the project using tree, find, map find, and map keyword — it never edits files. All actions are auto-executed without confirmation.

/ask what does this project do
/ask -t 5 where is authentication handled
/ask how are books stored in the database

Each tool result is truncated at ~1K tokens to protect the model's context window — if a result is too large, the model receives a specific hint on how to narrow the query (e.g. use /tree <subfolder> or add more keywords to /map find).

The system prompt guides the model from broad to narrow: tree → map find → map keyword → find → read. On a second /ask in the same session the model skips /tree if the project structure is already in context.

When the loop finishes you are prompted: [s]ummary / [a]ll / [n]one — choose how much of the agent's conversation to pull into your main context.

What 4B models can do with /ask: 10-step investigation, understand project structure, predict where relevant code is located — tasks that normally require 30B+ models in open-ended agent mode.


Agent mode

/agent runs an autonomous loop: the model reads the task and decides which tool to use. The agent prompt instructs the model to emit one ACTION per turn; if it emits multiple, all are executed in order. Stops when the model outputs plain text with no ACTION.

/agent [-t N] [-y] <task> [plan step1, step2, ...]
  • -t N — override max_turns for this run only (e.g. -t 1 for a quick read+explain)
  • -y — skip per-action confirmation (execute all actions automatically)
  • Without -y: each proposed action pauses and asks [Y/n/e/f/q]:
Key Action
Y / Enter Execute the action
n Skip this action
e Edit the command before executing (copies to clipboard on Windows)
f Send feedback to the AI and skip the action (redirect the model mid-loop)
q Stop the agent
  • plan: step1, step2, ... — optional comma-separated list of items injected as hints one per turn
  • file: <steps.txt> — load steps from a .txt or .md file; numbered/bulleted list items become steps; ### Example / ### Summary sections are injected as context before step 1; max_turns is raised automatically if the file has more steps than the default limit; gate FAIL on a step retries it

When the loop finishes you are prompted: [s]ummary / [a]ll / [n]one — choose how much of the agent's conversation to pull into your main context.

/agent find and fix the divide by zero bug in calc.py
/agent -t 1 read models.py and explain the User class
/agent -y -t 5 refactor utils.py
/agent read files plan: models.py, views.py, urls.py
/agent implement the changes file: plan.txt    # load steps from plan.txt

Configure the default agent in .1bcoder/agent.txt:

max_turns = 10
auto_apply = true

tools =
    read
    insert
    save
    patch

Named agents

Custom agents are defined in .1bcoder/agents/<name>.txt (project-local) or <install>/.1bcoder/agents/<name>.txt (global). Local files override global ones. Call them with /agent <name> task or directly as /<name> task.

Agent file format:

# .1bcoder/agents/myagent.txt
description = What this agent does
max_turns = 10
auto_exec = false
auto_apply = false

system =
    You are a ... Complete the task using the available tools.

    To call a tool, write ACTION: followed by the command.
    ...

    Available tools:
    {tool_list}

tools =
    read
    find
    run

aliases =
    /search = /map find {{args}}
    /sql    = /run python db.py "{{args}}"

params =
    num_predict = 150
    agent_ctx = 4000
    temperature = 0.2

before =
    assist /short

gates =
    action-required
    pattern-gate "@Query" "use CriteriaBuilder only"
  • system = — inline multiline system prompt; indented lines continue the block; {tool_list} is substituted automatically from the tools = list
  • tools = — one tool name per indented line; controls what the agent knows about and what gets shown in its system prompt; empty tools = line means no tools (pure text agent)
  • aliases = — agent-scoped aliases; active only during this agent's run, restored to global state after; {{args}} is replaced by everything after the alias name
  • params = — model and agent parameters set for this run; agent_ctx sets the agent's private context limit (equivalent to /tempctx N); num_predict, temperature, etc. are forwarded to the model
  • before = — one proc per indented line; runs before every LLM call; all stdout injected as [context] into the agent's message list; useful for injecting a hint or parallel sub-query result each turn
  • gates = — one proc per indented line (with optional args); runs after each reply; if any proc prints FAIL:, the current plan step is retried and the FAIL reason is shown to the model as feedback
  • on_done = <command> — slash command executed once when the agent finishes naturally (no more ACTIONs); use to save the agent's final reply to a file (e.g. on_done = /ctx compact scan_result)
# Example: planning agent saves its output automatically
on_done = /save plan.txt -w

Gate procs — built-in procs designed for use in gates =:

Proc What it checks FAIL condition
action-required Agent reply contains ACTION: or a completion phrase Neither found — suggests the bare command if one appears anywhere in the reply
pattern-gate "regexp" "msg" Reply matches the given regular expression Match found — prints msg as the FAIL reason
grounding-check Identifiers in reply exist in map.txt Grounding score < 50% (prints warning; does not FAIL by default)

Before procs — built-in procs designed for use in before =:

Proc What it does
assist /short Reads last reply, asks the LLM for a one-sentence next-step hint, injects as [context]
assist /parallel profile <name> Same but uses a parallel profile to query the hint model

Built-in named agents (global install):

Agent Command Description
ask /ask <question> or /agent ask Read-only research — tree, find, map, never edits
advance /advance <task> or /agent advance Full toolset for 7B+ models
planning /plan <goal> Researches project, writes natural-language plan to plan.txt
fill /fill Reads NaN vars, finds .var files, sets missing values from project files
scan /scan <file> <theme> Reads large file chunk by chunk, extracts themed info, saves to .1bcoder/scan_result.txt

/agent advance — named agent from agents/advance.txt, full toolset for larger models (7B+), includes run, diff, map, bkup, and all edit tools. Shortcut: /advance:

/agent advance refactor the auth module
/advance read and summarise plan models.py, views.py

/concepts <topic> — alias for /agent concepts. Pure brainstorm agent: no tools, no files. Iterates through a list of concept seeds injected via plan:, produces a synthesis paragraph per turn. Useful for exploring design ideas, academic framing, or philosophical grounding without polluting the tool context.

/concepts symbol grounding in software engineering

Agent procs: before and gates

before = and gates = turn an agent into a supervised loop — useful when the model is small or the task requires strict compliance.

before = fires before every LLM call. All stdout is injected as a [context] message so the model sees it before generating its reply. Use cases:

  • assist /short — ask a second model for a one-sentence hint based on the last reply
  • assist /parallel profile ten_experts — poll multiple models for the next search direction

gates = fires after every reply. Each proc receives the reply on stdin. If any prints FAIL:, the agent:

  1. Re-injects the current plan step (so the model sees the hint again)
  2. Feeds the FAIL reason as user feedback
  3. Retries the turn (does not advance to the next plan step)

This is the key difference from a regular proc: a gate can hold the agent on a step until it produces a compliant reply.

Combined example — agent that must always emit an ACTION:

# agents/strict.txt
tools =
    read
    patch
    tempctx

gates =
    action-required

params =
    agent_ctx = 6000
    num_predict = 200
/agent strict fix the authentication bug file: plan.txt

If the model reasons without acting, action-required prints FAIL: and the step is retried with the failure reason visible.

Pattern gate — enforce coding conventions across all turns:

gates =
    pattern-gate "from dual" "no FROM DUAL allowed"
    pattern-gate "select \*" "use explicit columns"
    action-required

Multiple gates can be stacked. All run after each reply; a single FAIL from any of them retries the step.


Aliases

Define command shortcuts with /alias:

/alias /name = expansion          define an alias ({{args}} = everything after the name)
/alias                            list all active aliases
/alias clear /name                remove an alias (session only)
/alias save /name                 persist to .1bcoder/aliases.txt

Aliases are loaded at startup from the global aliases.txt then the project-local one and survive /clear. They are expanded before any command is dispatched, so aliases can expand to other aliases (up to 10 levels deep).

/alias /grep = /find {{args}} -c
/alias /kw   = /map keyword extract {{args}} -f -c
/alias save /grep

Agent-scoped aliases — an agent's aliases = section is merged into the active alias table before the loop starts and fully restored after. The agent can use its own shorthand commands in ACTION: lines; they disappear when the agent finishes.


Scripts

Scripts are .txt files containing one command per line, stored in .1bcoder/scripts/. Lines starting with [v] are already done and skipped. Lines starting with # are comments and skipped.

Command Description
/script list List all script files (* = current). Shows global scripts (g:) and project plans
/script open Select and load a script (type number). Includes global and project scripts
/script open <N> Load script by number directly — shows the list but skips the prompt
/script create [path] Create a new empty script
/script create ctx [path] Create script from this session's command history
/script show Display steps of the current script
/script add <command> Append a step to the current script
/script clear Wipe current script completely
/script reset Unmark all done steps (also happens automatically when a script runs to completion)
/script reapply [key=value ...] Reset all done steps then apply automatically; prompts for any NaN {{variables}} before running
/script refresh Reload script from disk and show contents
/script apply [file] [key=value ...] Run steps one by one (Y/n/q per step)
/script apply -y [file] [key=value ...] Run all pending steps automatically

/script create ctx captures all work commands typed this session (/read, /edit, /fix, /patch, /run, /save, /bkup, /map, /model, /host) into a ready-to-run plan:

> /host http://192.168.1.50:11434
> /model gemma3:1b
> /read calc.py
> /fix calc.py 11-11 divide by zero
> /run python calc.py

> /script create ctx fix-calc.txt
[script] created 'fix-calc.txt' from session history (5 step(s))

Scripts support {{key}} placeholders:

/read {{file}} {{range}}
what is wrong in lines {{range}}?
/fix {{file}} {{range}} {{hint}}
/script apply fix-fn.txt file=calc.py range=1-4 hint="wrong operator"

Run a script non-interactively from the command line:

1bcoder --model llama3.2:1b --scriptapply my-fixes.txt --param file=calc.py

Session variables (/var)

Session variables store named values that are substituted as {{name}} in any command, script step, or agent plan. Useful for parameterizing scripts without hard-coding values.

/var set port=5432               set directly (shorthand)
/var set port =5432              set directly (original form)
/var set port = 5432             set with spaces (same result)
/var set name =MyService         literal value
/var def port db host            declare multiple NaN variables (skips if already set)
/var get                         list all variables (NaN = unset)
/var get port                    print value of a single variable (useful with ->)
/var del port                    remove a variable

Capture from proc output:

/proc run regexp-extract "\bclass (\w+)" -g 1
/var set classname first         # grab first= key from proc stdout

Save and load .var files — for reuse across sessions without loading files into context:

/var set description=DB connection params for INVOICES project
/var set port=5432
/var set db=invoices.db
/var save invoices               # creates invoices.var

The first line of a .var file is always # <description> — an agent can read just that line (/read invoices.var 1-1) to decide relevance before loading the full file.

/var load invoices.var           # restore vars from file

Extract placeholders from script or file:

/var extract                     # scan current open script for {{placeholders}}
/var extract deploy.txt          # scan any file

Any {{key}} found but not yet set is registered as NaN — /script reapply will prompt for it before running.

/fill agent — for weak models that can't hold large files in context, use /fill to populate NaN variables automatically. The agent:

  1. Runs /var get to see what's missing
  2. Searches for .var files (/find . -f --ext var), reads only the first line (description) of each to decide relevance
  3. Loads relevant .var files
  4. Searches project config files for any remaining NaN values
/var extract                     # register NaN vars
/fill                            # agent fills them from project files
/script reapply                  # runs with all values set

Output capture (->, $ and ~)

Any command — LLM reply, tool output, or proc result — can be captured into a session variable using the -> suffix. Two special tokens expand anywhere in a command or message:

  • $ — last captured output (last AI reply or tool result)
  • ~ — last user input (last message or command you typed)
/map keyword extract auth.py -> keywords      # capture tool output into variable
/ask find files related to {{keywords}}       # use it in next command

summarize this for me -> myplan              # capture LLM reply
/agent planning $                            # pass it as task to next command

/find . -f --ext py -> filelist              # capture file listing
/agent ask "which of these handles auth? $"  # inline expand into message

/proc run my-extractor -> result             # capture proc stdout
/var set port result                         # also works: grab key from proc output

~ — repeat or redirect the last question:

як працює цей метод?                         # ask main model
/small ~                                     # same question → small model
/ask ~                                       # same question → agent mode
/explain "$"                                 # ask small model to explain the reply
поясни: $                                    # ask main model to explain its own reply

-> stores the full text (including ANSI-stripped terminal output) and also updates $ for immediate reuse. Variables captured with -> appear in /var get like any other session variable.


Project config (/config)

Save and restore session state (host, model, ctx, params, vars, procs). Two config locations are supported:

  • Local.1bcoder/config.yml in the current working directory (project-specific)
  • Global~/.1bcoder/config.yml (user-wide default for all projects)

Startup priority: on launch without --host/--model, 1bcoder checks local config first, then global. The first one with auto: true wins. If neither has auto: true, connects to local Ollama and prompts for a model.

/config save                     # save all current state to local config
/config save global              # save all current state to global config
/config save host                # save only host to local config
/config save global host         # save only host to global config
/config save model               # save only model to local config
/config save global model        # save only model to global config
/config save vars                # save only vars to local config
/config load                     # restore from local config
/config load global              # restore from global config
/config show                     # print local config contents
/config show global              # print global config contents
/config auto on                  # enable auto-load in local config
/config auto on global           # enable auto-load in global config
/config auto off                 # disable auto-load in local config

Selective delete:

/config del model                # remove model from config
/config del var project          # remove one variable
/config del vars                 # remove entire vars section
/config del param num_ctx        # remove one param
/config del procs                # remove entire procs section
/config del proc collect-files   # remove one proc

Config file format (.1bcoder/config.yml or ~/.1bcoder/config.yml):

auto: true
host: ollama://localhost:11434
model: qwen3:1.7b
ctx: 4096
params:
  num_ctx: 4096
  temperature: 0.7
vars:
  project: bookcrossing
  db: invoices.db
procs:
  - collect-files output.txt

When auto: true, host and model are used at startup to connect; ctx, params, vars, and procs are also restored.


MCP (Model Context Protocol)

Connect external tool servers to give the AI access to filesystems, databases, web pages, and more.

/mcp connect <name> <command>
/mcp tools [name]
/mcp call <server/tool> [json_args]
/mcp disconnect <name>
/mcp connect fs npx -y @modelcontextprotocol/server-filesystem .
/mcp connect web uvx mcp-server-fetch
/mcp call web/fetch {"url": "https://docs.python.org/3/"}
/mcp tools
/mcp disconnect fs

See /doc MCP for a full list of ready-to-use servers.


Parallel queries

Send a prompt to multiple models at the same time. No quoting required.

/parallel [main question]  [list: a1, a2, a3]  profile: name
          [ctx: full|last|none]  [file: path [-n N]]
          [collect: compact [profile: name]]  [--seq]

Prompt modes

Mode Behaviour
plain text Same prompt sent to all workers
list: a1, a2, a3 Aspects distributed one per worker; combined with the main question as {main question}\n\nAspect: {aspect_i}
comma-separated prompts Matched 1:1 to workers (last reused for remaining)

Use list: when you want to explore a single question from multiple angles simultaneously — each model gets the full question plus one specific aspect to focus on:

/parallel Hegel philosophy  list: axiology, epistemology, ethics, logic  profile: four-models

Context modes (default: ctx: last)

Keyword Context sent to workers
ctx: full Full conversation context
ctx: last Last message only (default)
ctx: none No context — prompt only

$ and ~ expansion$ expands to the last AI reply, ~ to your last input:

/parallel $  profile: short          # ask short model to summarise last reply
/parallel ~  list: pros, cons  profile: two-models  # weigh your last question from two angles

File chunking — split a file across workers:

/parallel file: bigfile.txt -n auto  profile: cluster  # -n auto = one chunk per worker
/parallel file: notes.md             profile: small     # === separator used if present

Collect — compact all worker replies into the context after they finish:

/parallel list: q1, q2  profile: small1  collect: compact
/parallel list: q1, q2  profile: small1  collect: compact profile: short

collect: compact sets a savepoint automatically before workers run, then calls /ctx compact savepoint (optionally with an external model) once all workers finish. The result is available as $.

Sequential mode--seq runs workers one after another instead of in parallel.

Profiles — save a set of workers for reuse:

/parallel profile create <name> host|model|file ...   # inline — workers as space-separated specs
/parallel profile create <name>                       # interactive wizard
/parallel profile list                                # show all profiles (local + global)
/parallel profile show <name>                         # print raw profile string
/parallel profile add <name>                          # append current host+model to a profile

Profiles stored in ~/.1bcoder/profiles.txt (global) or .1bcoder/profiles.txt (project-local):

review: localhost:11434|ministral3:3b|ans/review.txt localhost:11435|cogito:3b|ans/tests.txt  # code review + unit tests
fast:   localhost:11434|qwen2.5-coder:0.6b|ans/q.txt                                          # quick sanity check

Sub-agent profiles — built-in profiles that return answers directly to the main context:

small:    localhost:11434|qwen3:0.6b|ctx
explain:  localhost:11434|gemma3:1b|ctx
thinking: localhost:11434|lfm2.5-thinking:1.2b|ctx
short:    localhost:11434|llama3.2:1b|ctx

These are aliased as /small, /explain, /thinking, /short:

/small what does this function return?     # ask tiny model, last message as context
/explain $                                 # ask gemma to explain last reply
/short ~  ctx: none                        # repeat last question with no context

Hooks (/hook)

Run a script automatically before or after a command. Useful for backups before edits, linting after patches, or any pre/post workflow step.

/hook before <cmd> <script>   # run script before every <cmd>
/hook after  <cmd> <script>   # run script after  every <cmd>
/hook list                    # show active hooks
/hook clear <cmd>             # remove hooks for <cmd>
/hook clear                   # remove all hooks

<cmd> is the command name without the slash: edit, patch, fix, insert, run.

Two script types:

  • .txt — 1bcoder script (sequence of commands). {{file}} and {{range}} are injected as session variables.
  • .py — Python guard subprocess. Receives trigger content on stdin, outputs BLOCK:/ALERT:/ACTION: lines.

Auto-injected for .txt scripts:

Variable Value
{{file}} file argument of the triggering command
{{range}} line range (if specified), e.g. 10-25

Examples:

/hook before edit /bkup {{file}}              # backup before every edit (.txt script)
/hook before run sql_readonly_guard.py        # block dangerous SQL (.py guard)

Missing .txt script cancels a before hook. .py guard cancels only if it prints BLOCK:. Step errors inside .txt scripts do not cancel the command.


Prompt templates

Save any useful message as a reusable template and load it later with {{param}} substitution.

/prompt save ConvertJavaToPy     # saves last user message as ConvertJavaToPy.txt
/prompt load                     # numbered list, select by number, fill {{params}} interactively

Templates stored in <install>/.1bcoder/prompts/. Use {{keyword}} placeholders — values are prompted on load.


Post-processors (/proc)

Run a Python script against the last LLM reply. Useful for extracting filenames, validating identifiers against map.txt, collecting data across turns, and more.

/proc list                         # list available processors
/proc run <name>                   # one-shot: run against last reply
/proc run <name> -f <file>         # run against an external file instead of last reply
/proc on grounding-check           # persistent: run after every reply automatically
/proc off                          # stop persistent processor
/proc before on assist /short      # before-proc: run BEFORE every LLM call, inject output as [context]
/proc before off                   # stop before processor
/proc gate on action-required      # gate: run after reply; FAIL retries current plan step
/proc gate on pattern-gate "regexp" "message"   # gate with regexp
/proc gate off                     # stop gate processor
/proc new my-proc                  # create a new processor from template

Processor protocol:

  • stdin = last LLM reply
  • stdout = result (injected into context)
  • key=value lines = extracted params
  • ACTION: /command = confirmed and executed (run mode only)
  • ALERT: message = warning printed, continues
  • BLOCK: reason = cancels the triggering command (hook mode only)
  • FAIL: reason = gate mode: retries plan step and feeds reason to model; proc mode: prints warning
  • exit 1 = show stderr as warning, skip ACTION
  • BCODER_WORKDIR env var is set to the current working directory in all proc subprocesses (use to find map.txt, project files, etc.)

Built-in processors in ~/.1bcoder/proc/:

Processor Purpose Best mode
extract-files Extract filenames, ACTION: /read if one found one-shot
extract-code Extract code blocks; ACTION: /save <file> if one block + filename detected one-shot
extract-list Convert first bullet/numbered list in reply to comma-separated line one-shot
grounding-check Score identifiers against map.txt, warn if <50% persistent / gate
collect-files Accumulate filenames to .1bcoder/collected-files.txt persistent
md Render last reply as formatted Markdown in terminal one-shot
mdx Render last reply as Markdown + LaTeX (KaTeX) + Mermaid diagrams in browser one-shot
ctx_cut Auto /ctx cut when context exceeds threshold (default 90%) persistent
rude_words Alert if reply contains profanity (ua arg adds Ukrainian list) persistent
secret_check Alert if reply contains sensitive names (google, anthropic…) persistent
sql_readonly_guard Alert (proc) or block (hook) on write SQL statements both
action-required FAIL if agent reply has no ACTION: and no completion phrase gate
pattern-gate FAIL if reply matches given regexp (argv[1] = pattern, argv[2] = message) gate
assist Before-proc: reads last reply, asks LLM for one-sentence next-step hint before

Guard usage examples:

/proc on ctx_cut 80                          # auto cut at 80%
/proc on rude_words ua                       # profanity check + Ukrainian
/proc on secret_check client=acme            # + custom keyword
/hook before run sql_readonly_guard.py       # block /run with DELETE/DROP/UPDATE

See /doc PROC for the full protocol, built-in processor reference, and guide to writing your own.


Team runs (/team)

Spawn multiple 1bcoder workers in parallel, each running a different plan against the same project. Each worker gets its own context (e.g. /tree, /find, /map) and saves results to .1bcoder/results/.

/team list                                           # list team definitions
/team show code-analysis                             # show workers in a team
/team new my-team                                    # create team yaml from template
/team run code-analysis --param keyword=auth --param task="404 on login"

Team definition (.1bcoder/teams/<name>.yaml):

workers:
  - name: structure
    host: localhost:11434
    model: qwen2.5-coder:1.5b
    script: team-tree-worker.txt
  - name: search
    host: openai://localhost:1234
    model: qwen2.5-coder:1.5b
    script: team-search-worker.txt
    depends_on: structure
  - name: summary
    host: 192.168.0.10:11434
    model: gemma3:4b
    script: team-map-worker.txt
    depends_on: structure, search

name — optional worker label (auto-assigned 1, 2, 3… if omitted). depends_on — comma-separated worker names; worker waits until all listed workers finish before starting. Workers without depends_on start in parallel immediately.

--param values are forwarded to every worker script as {{placeholders}}. Each worker log goes to .1bcoder/team-logs/. After all workers finish, aggregate with a summary script:

/script apply team-summarize.txt --param keyword=auth --param task="404 on login"

Built-in team scripts in <install>/.1bcoder/scripts/:

Script Worker role
team-tree-worker.txt /tree → where does the keyword live structurally?
team-search-worker.txt /find → which functions implement it?
team-map-worker.txt /map find → what depends on it?
team-summarize.txt reads all results, produces unified answer

Session controls

Command Description
/model [-sc] Switch AI model interactively
/model <name> [-sc] Switch directly by name (e.g. /model gemma3:1b)
/host <url> [-sc] Switch host and provider (see below); -sc keeps context
/ctx <n> Set context window size in tokens (default 8192)
/ctx Show current usage vs limit
/ctx clear Clear all conversation messages (keeps /param and num_ctx)
/ctx clear <n> Remove last N messages from context
/ctx cut Remove oldest messages until context fits
/ctx compact Ask AI to summarize the conversation, replace context with summary
/ctx save <file> Save full conversation to file
/ctx load <file> Restore a saved conversation (bare name resolved from .1bcoder/ctx/)
/ctx list List files in .1bcoder/ctx/ project context library
/ctx savepoint set Mark current position as a savepoint
/ctx savepoint rollback Remove all messages added since the savepoint
/ctx savepoint compact Summarize messages since savepoint, replace with summary
/ctx savepoint show Show savepoint info and messages added since
/tempctx <N> Set agent context limit to N tokens for this run (also settable via params = agent_ctx = N in agent file)
/tempctx show Show agent context size — only available inside an agent loop
/tempctx cut Remove oldest messages from agent context until it fits
/tempctx clear Reset agent context to system prompt + task only
/think exclude Strip <think> blocks from context (default)
/think include Keep <think> blocks in context (pass model reasoning to next turn)
/think show Show <think> blocks in terminal (default)
/think hide Hide <think> blocks in terminal
/param <key> <value> Set a model parameter for every request (e.g. temperature, enable_thinking)
/param timeout <seconds> Set HTTP read timeout — increase when models are slow on large contexts (default: 120s)
/param Show currently set params including timeout
/param clear Remove all params and reset timeout to 120s
/role <persona> Set a system persona for the AI (e.g. /role You are a Linux kernel expert)
/role show Show the active persona
/role clear Remove the active persona
/format <description> Inject a strict output format constraint and call the AI (e.g. JSON array, one word)
/format clear Remove the active format constraint from context
/clear Clear conversation context, reset params, and reload model metadata
/help Show full command reference
/help <command> Show help for one command (e.g. /help map)
/help <command> ctx Same, and inject into AI context
/init Create .1bcoder/ scaffold in current directory
/exit Quit

Providers

1bcoder supports Ollama (default) and any OpenAI-compatible endpoint (LMStudio, LiteLLM, etc.). The provider is encoded in the URL scheme — no separate flag needed.

URL Provider
localhost:11434 Ollama (default, no scheme needed)
ollama://localhost:11434 Ollama (explicit)
openai://localhost:1234 LMStudio
openai://localhost:4000 LiteLLM
openai://api.example.com Any OpenAI-compatible proxy
/host openai://localhost:1234          # switch to LMStudio, clear context
/host openai://localhost:4000 -sc      # switch to LiteLLM, keep context
/host localhost:11434                  # back to Ollama

/parallel with mixed providers — each worker carries its own scheme:

/parallel "review this" \
    ollama://localhost:11434|llama3.2:1b|ans/ollama.txt \
    openai://localhost:1234|qwen2.5:7b|ans/lmstudio.txt \
    openai://localhost:4000|gpt-4o-mini|ans/litellm.txt

On startup the active provider is shown:

  model    : llama3.2:1b [1.3G Q4_K 131K]
  host     : http://localhost:11434
  provider : ollama

The status line before each > prompt shows the same info in compact form:

 llama3.2:1b [1.3G Q4_K 131K]  │  ctx 245 / 8192 (3%)
 gpt-4o-mini [128K]  │  ctx 1536 / 128000 (1%)     ← OpenAI (no disk size)

Project layout

1bcoder/
├── chat.py           # entire application — REPL, all commands
├── map_index.py      # standalone project scanner (usable without 1bcoder)
├── map_query.py      # standalone map query tool (find + trace)
├── map_query_help.txt # full map_query usage reference
├── requirements.txt  # pip dependencies
├── pyproject.toml    # build metadata
├── run.bat           # Windows quick-launch
├── MCP.md            # MCP server quick-reference
└── _bcoder_data/          # wheel defaults (copied to ~/.1bcoder/ on first run)
    ├── agents/            # built-in named agent definitions (ask.txt, advance.txt, ...)
    ├── aliases.txt        # global aliases loaded at startup
    ├── scripts/           # built-in plan .txt files (team workers, examples)
    ├── teams/             # /team yaml definitions
    ├── prompts/           # /prompt saved templates
    ├── proc/              # /proc post-processor scripts
    └── doc/               # /doc articles (PROC.md, ...)

~/.1bcoder/                # user global dir (created on first run, edit freely)
    └── (same structure as above — overrides wheel defaults)

<project>/.1bcoder/        # project-local dir (created by /init)
    ├── agents/            # project-specific agent definitions (override global)
    ├── aliases.txt        # project-local aliases (merged over global at startup)
    ├── scripts/           # project-specific plan .txt files
    ├── teams/             # project-specific team yamls (override global)
    ├── agent.txt          # default agent config (max_turns, auto_apply, tools)
    ├── config.yml         # /config save — host, model, ctx, params, vars, procs
    ├── profiles.txt       # /parallel worker profiles (project-local)
    ├── map.txt            # generated by /map index
    ├── map.prev.txt       # previous snapshot (for /map diff)
    ├── results/           # worker output files (tree-analysis.txt, etc.)
    └── team-logs/         # per-worker logs from /team run

Keyboard shortcuts

Key Action
Enter Submit message
Shift+Enter Insert newline (requires Kitty keyboard protocol support)
Ctrl+N Insert newline (reliable fallback for all terminals)
ESC Interrupt AI response mid-stream
Ctrl+C Interrupt streaming or exit prompt

On Windows: hold Shift and drag with the left mouse button to select and copy text from the terminal.


Tips for 1B models

  • Start small. Use /read file.py 10-25 instead of loading the whole file. Short context = better focus.
  • Use /fix not /patch. The LINE N: content format is much more reliable at 1B scale than free-form generation.
  • Build a map first. Run /map index . at the start of a session, then use /map find to load only the relevant parts into context.
  • Use scripts. Scripts make multi-step work reproducible — the model only needs to handle one step at a time.
  • Capture workflows. After solving a task manually, run /script create ctx to save the exact steps as a reusable script.
  • Use .var files instead of context. Save project constants (port, db, host, main_file) to a .var file once with /var save. Future sessions reload them instantly with /var load — no tokens wasted reading config files.
  • Let /fill do the research. If you have NaN variables and a 4B+ model, just run /fill — the agent finds and loads .var files, reads config files, and sets everything without you having to specify where to look.
  • Use /plan before /agent. For complex tasks, run /plan <goal> first to get a structured step-by-step plan in plan.txt, then /agent implement plan plan.txt to execute it turn by turn.
  • Agent mode needs a bigger model. /agent advance works best with 32B+ models. For 1B–7B, use scripts instead.
  • Ctrl+C interrupts streaming if the model starts going off-track.
  • Use /readln before /patch. Reading with line numbers lets you reference exact locations; /patch then sends the file without line numbers so the model's SEARCH block matches the real content.
  • Timeout errors? If you see Read timed out with a large context, run /param timeout 300 to extend the limit to 5 minutes.
  • Model behaving oddly after a long session? Run /clear — it clears context, resets all params, and reloads model metadata, equivalent to a restart.

Command autocorrection

1bcoder automatically detects and fixes common typos in commands before executing them — for both human input and agent-generated ACTION: lines.

Error type Example Fixed to
Command name typo /insrt models.py 14 /insert models.py 14
File path typo /read models.p /read models.py
Keyword prefix /insert main.py 14 co /insert main.py 14 code
Subcommand prefix /map fnd auth /map find auth
Subcommand typo /bkup resore calc.py /bkup restore calc.py

For human input, the corrected command is shown with [fix?] and you are asked to confirm. For agent actions (auto mode), the fix is applied silently with a [fix] warning. Prefix matching is used for keywords to avoid false positives on hint words.


Tips for reasoning models (Qwen3, DeepSeek-R1, etc.)

  • Disable thinking for simple tasks. /param enable_thinking false speeds up responses when reasoning isn't needed.
  • Use /think include to chain reasoning. Pass one model's <think> output as context to another model or the next turn.
  • /patch <file> code over /edit. Reasoning models write precise SEARCH/REPLACE blocks — no line numbers needed, no full-file rewrites.
  • /ctx compact after long sessions. Reasoning models produce verbose output; compact regularly to stay within context limits.
  • Connect via LMStudio. /host openai://localhost:1234 — full parameter control including enable_thinking, temperature, seed.

Compatible local inference backends

Name OS Server Default port Connector Description
Ollama Win / Mac / Linux built-in 11434 ollama:// Default backend, easiest setup, pulls models automatically
LM Studio Win / Mac / Linux built-in 1234 openai:// Desktop GUI, model browser, OpenAI-compatible server
llama.cpp Win / Mac / Linux ./llama-server 8080 openai:// Minimal C++ server, CPU + GPU, no install needed
LocalAI Linux / Docker Docker container 8080 openai:// Runs many formats in a container, drop-in OpenAI replacement
Jan.ai Win / Mac / Linux built-in 1337 openai:// Desktop app, offline-first, OpenAI-compatible local server
llamafile Win / Mac / Linux self-contained exe 8080 openai:// Single executable, no install, built-in server (Mozilla)
GPT4All Win / Mac / Linux built-in 4891 openai:// Desktop app, CPU-friendly, targets non-technical users
Kobold.cpp Win / Mac / Linux built-in 5001 openai:// Popular for creative use, OpenAI-compatible API endpoint
text-generation-webui Linux / Win --api flag 5000 openai:// oobabooga UI, needs --api flag to expose OpenAI endpoint
TabbyAPI Linux / Win built-in 5000 openai:// Focused on exl2/GPTQ quantized models, low VRAM
vLLM Linux built-in 8000 openai:// Production server, high throughput, requires significant VRAM

(c) 2026 Stanislav Zholobetskyi Institute for Information Recording, National Academy of Sciences of Ukraine, Kyiv PhD research: «Intelligent Technology for Software Development and Maintenance Support»

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

1bcoder-0.1.4.tar.gz (189.0 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

1bcoder-0.1.4-py3-none-any.whl (163.6 kB view details)

Uploaded Python 3

File details

Details for the file 1bcoder-0.1.4.tar.gz.

File metadata

  • Download URL: 1bcoder-0.1.4.tar.gz
  • Upload date:
  • Size: 189.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.9

File hashes

Hashes for 1bcoder-0.1.4.tar.gz
Algorithm Hash digest
SHA256 9c4f06e5ed82d0811e11650ceb287c92d0496fd26ef683ccacb9c4528b57e067
MD5 2dba49e48221389be52db1ee660ef70a
BLAKE2b-256 5d2b2e07dda554e7afd00e60f83973de0afc4c7840bc71503946791e4f88c53f

See more details on using hashes here.

File details

Details for the file 1bcoder-0.1.4-py3-none-any.whl.

File metadata

  • Download URL: 1bcoder-0.1.4-py3-none-any.whl
  • Upload date:
  • Size: 163.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.9

File hashes

Hashes for 1bcoder-0.1.4-py3-none-any.whl
Algorithm Hash digest
SHA256 8d1759356c14684558b2748e2d0f416614d1b4e7ff8c0dfee7ba715831e60850
MD5 44b200eea5db092c5757df837acbe0c2
BLAKE2b-256 02148ae87b0f3a62c6fc906b54cc196378e99bab4da5558ec22d1ca3029a467c

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page