Skip to main content

Hardware LLM capability scanner — know what runs on your machine

Project description

tinillm

Know what LLMs your hardware can run — locally, instantly.

pipx install tinillm
tinillm

What it does

tinillm is an interactive tool. Launch it by typing tinillm in your terminal, and you'll land in a welcome screen. From there, every feature is a slash command: /scan inspects your hardware, /models browses real LLMs, /run launches one in Ollama, and so on.

╭─── tinillm v1.9.0 ───────────────────────────────────────────────────╮
│  ████████╗                                                           │
│     ██╔══╝    Welcome back, Harish!      Tips for getting started    │
│     ██║                                  ────────────────────────    │
│     ██║       v1.9.0 · tinillm             Run /scan    to detect    │
│     ╚═╝       ~/tinillm_CLI                Run /models  to browse    │
│               ollama ● running             Run /run     to launch    │
│                                            Run /doctor  health chk   │
│                                            Run /help    list all     │
│                                                                      │
│                                          Recent activity             │
│                                          ────────────────────────    │
│                                            /scan                     │
│                                            /run llama3.2:3b          │
╰──────────────────────────────────────────────────────────────────────╯
  Type /help for commands · /exit or Ctrl+D to quit

tinillm> /scan

  LLM Capability Matrix

  Model    Fit        Best Quant   Mem Needed   Tokens/sec
  ~1B      Perfect    Q8_0          1.9 GB        580 t/s
  ~3B      Perfect    Q8_0          3.8 GB        195 t/s
  ~7B      Perfect    Q6_K          6.2 GB         88 t/s
  ~13B     Perfect    Q5_K_M       10.1 GB         47 t/s
  ~34B     Good       Q4_K_M       21.8 GB         18 t/s

tinillm>

Works on Linux, macOS, and Windows. No GPU required — CPU-only machines are supported too.


Install

pipx install tinillm     # recommended: isolated per-tool environment
# or
pip install tinillm

Requires Python 3.11+. No other tools needed.


Usage

Launch the tool with a single command:

tinillm

Inside the REPL, every feature is a slash command:

Command What it does
/scan Scan hardware and show which LLM sizes fit
/scan --verbose Include model sizes that don't fit
/scan --json Machine-readable JSON output
/models Browse real LLM models and see which fit
/models --fits-only Hide models that don't fit
/models --ollama Show which models are installed in local Ollama
/run Pick a compatible model interactively and run it
/run llama3.2:3b Launch a specific model directly
/suggest --use-case coding Personalised model recommendation
/doctor System health check (hardware + Ollama)
/help List every command
/clear Clear the terminal
/exit Quit (or Ctrl+D)

Tab-completion works on slash commands, subcommands, and flags.


First launch

The first time you run tinillm, it automatically runs /scan for you so you see your hardware capabilities immediately. On subsequent launches, just the welcome panel appears.


GPU support

Vendor Detection method
NVIDIA nvidia-smi → sysfs fallback
AMD rocm-smi → sysfs fallback
Apple Silicon system_profiler (unified memory)
Intel Arc sysfs + lspci
Windows (all) PowerShell WMI
Any vulkaninfo last-resort fallback

Fit levels explained

Level Meaning
Perfect Fits comfortably at Q4_K_M or better with ≥20% VRAM headroom
Good Fits but tightly
Marginal Fits only at heavy compression / reduced context, or CPU-only
TooTight Won't fit under any quantisation

Versioning

Version Feature
1.9 Dropped RAG — focused on hardware scanning + model runner ← current
1.8 Interactive REPL (single entry-point, slash commands)
1.7 Added RAG (/index, /ask, /rag)
1.1 First feature — hardware scanner

Part of the tini* family

Tool What it does
tiniRAG Privacy-first RAG CLI
tinillm Interactive LLM + hardware tool

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

tinillm-2.3.0.tar.gz (121.0 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

tinillm-2.3.0-py3-none-any.whl (110.9 kB view details)

Uploaded Python 3

File details

Details for the file tinillm-2.3.0.tar.gz.

File metadata

  • Download URL: tinillm-2.3.0.tar.gz
  • Upload date:
  • Size: 121.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.3

File hashes

Hashes for tinillm-2.3.0.tar.gz
Algorithm Hash digest
SHA256 b4d706cec0658972fd960c672cfa6f3364125ad53aa72afb9372527f8d1fa169
MD5 6735b6068c7ea29652cbb6aa65304abf
BLAKE2b-256 c9e535a01527b38fc20fb261c344845d1d55fe83058669507f67e9db7c60fe66

See more details on using hashes here.

File details

Details for the file tinillm-2.3.0-py3-none-any.whl.

File metadata

  • Download URL: tinillm-2.3.0-py3-none-any.whl
  • Upload date:
  • Size: 110.9 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.3

File hashes

Hashes for tinillm-2.3.0-py3-none-any.whl
Algorithm Hash digest
SHA256 51dd704b4642b7d729625585404c9895fe04d93ee42584cf36dea4e713a0749d
MD5 db97bc46409435e56b205cdcad7702e8
BLAKE2b-256 ec2f7a47434f7da7bfd29312aaf6d3ea9a961dd0f782bd7eb5c8c3c244ab89d2

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page