Skip to main content

Hardware LLM capability scanner — know what runs on your machine

Project description

tinillm

Know what LLMs your hardware can run — locally, instantly.

pipx install tinillm
tinillm

What it does

tinillm is an interactive tool. Launch it by typing tinillm in your terminal, and you'll land in a welcome screen. From there, every feature is a slash command: /scan inspects your hardware, /models browses real LLMs, /run launches one in Ollama, and so on.

╭─── tinillm v1.9.0 ───────────────────────────────────────────────────╮
│  ████████╗                                                           │
│     ██╔══╝    Welcome back, Harish!      Tips for getting started    │
│     ██║                                  ────────────────────────    │
│     ██║       v1.9.0 · tinillm             Run /scan    to detect    │
│     ╚═╝       ~/tinillm_CLI                Run /models  to browse    │
│               ollama ● running             Run /run     to launch    │
│                                            Run /doctor  health chk   │
│                                            Run /help    list all     │
│                                                                      │
│                                          Recent activity             │
│                                          ────────────────────────    │
│                                            /scan                     │
│                                            /run llama3.2:3b          │
╰──────────────────────────────────────────────────────────────────────╯
  Type /help for commands · /exit or Ctrl+D to quit

tinillm> /scan

  LLM Capability Matrix

  Model    Fit        Best Quant   Mem Needed   Tokens/sec
  ~1B      Perfect    Q8_0          1.9 GB        580 t/s
  ~3B      Perfect    Q8_0          3.8 GB        195 t/s
  ~7B      Perfect    Q6_K          6.2 GB         88 t/s
  ~13B     Perfect    Q5_K_M       10.1 GB         47 t/s
  ~34B     Good       Q4_K_M       21.8 GB         18 t/s

tinillm>

Works on Linux, macOS, and Windows. No GPU required — CPU-only machines are supported too.


Install

pipx install tinillm     # recommended: isolated per-tool environment
# or
pip install tinillm

Requires Python 3.11+. No other tools needed.


Usage

Launch the tool with a single command:

tinillm

Inside the REPL, every feature is a slash command:

Command What it does
/scan Scan hardware and show which LLM sizes fit
/scan --verbose Include model sizes that don't fit
/scan --json Machine-readable JSON output
/models Browse real LLM models and see which fit
/models --fits-only Hide models that don't fit
/models --ollama Show which models are installed in local Ollama
/run Pick a compatible model interactively and run it
/run llama3.2:3b Launch a specific model directly
/suggest --use-case coding Personalised model recommendation
/doctor System health check (hardware + Ollama)
/help List every command
/clear Clear the terminal
/exit Quit (or Ctrl+D)

Tab-completion works on slash commands, subcommands, and flags.


First launch

The first time you run tinillm, it automatically runs /scan for you so you see your hardware capabilities immediately. On subsequent launches, just the welcome panel appears.


GPU support

Vendor Detection method
NVIDIA nvidia-smi → sysfs fallback
AMD rocm-smi → sysfs fallback
Apple Silicon system_profiler (unified memory)
Intel Arc sysfs + lspci
Windows (all) PowerShell WMI
Any vulkaninfo last-resort fallback

Fit levels explained

Level Meaning
Perfect Fits comfortably at Q4_K_M or better with ≥20% VRAM headroom
Good Fits but tightly
Marginal Fits only at heavy compression / reduced context, or CPU-only
TooTight Won't fit under any quantisation

Versioning

Version Feature
1.9 Dropped RAG — focused on hardware scanning + model runner ← current
1.8 Interactive REPL (single entry-point, slash commands)
1.7 Added RAG (/index, /ask, /rag)
1.1 First feature — hardware scanner

Part of the tini* family

Tool What it does
tiniRAG Privacy-first RAG CLI
tinillm Interactive LLM + hardware tool

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

tinillm-2.1.1.tar.gz (110.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

tinillm-2.1.1-py3-none-any.whl (104.9 kB view details)

Uploaded Python 3

File details

Details for the file tinillm-2.1.1.tar.gz.

File metadata

  • Download URL: tinillm-2.1.1.tar.gz
  • Upload date:
  • Size: 110.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.3

File hashes

Hashes for tinillm-2.1.1.tar.gz
Algorithm Hash digest
SHA256 8984bd65ec8366bc2309df216e8fc25e219e55363da7cb674888d66ccc4c39df
MD5 ebfd5b152b7f06f4260ad7cb1dc40600
BLAKE2b-256 dc645fa507a1f9aea3123d580803dfe86ef599e2f3c68f581ddf79e1cb3b2631

See more details on using hashes here.

File details

Details for the file tinillm-2.1.1-py3-none-any.whl.

File metadata

  • Download URL: tinillm-2.1.1-py3-none-any.whl
  • Upload date:
  • Size: 104.9 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.3

File hashes

Hashes for tinillm-2.1.1-py3-none-any.whl
Algorithm Hash digest
SHA256 cf96c0ea8fe5cdf753638eeb55eb6be246ef0a0b3dfeaf86d1856fe88f7d4436
MD5 fbc20f86f5d501facf0fd31798200efb
BLAKE2b-256 fede56e5d2cd75aec90a3318dcc2358bf2f4d8e0c0167d3ca9db83fc94690060

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page