Skip to main content

Hardware LLM capability scanner — know what runs on your machine

Project description

tinillm

Know what LLMs your hardware can run — locally, instantly.

pipx install tinillm
tinillm

What it does

tinillm is an interactive tool. Launch it by typing tinillm in your terminal, and you'll land in a welcome screen. From there, every feature is a slash command: /scan inspects your hardware, /models browses real LLMs, /run launches one in Ollama, and so on.

╭─── tinillm v1.9.0 ───────────────────────────────────────────────────╮
│  ████████╗                                                           │
│     ██╔══╝    Welcome back, Harish!      Tips for getting started    │
│     ██║                                  ────────────────────────    │
│     ██║       v1.9.0 · tinillm             Run /scan    to detect    │
│     ╚═╝       ~/tinillm_CLI                Run /models  to browse    │
│               ollama ● running             Run /run     to launch    │
│                                            Run /doctor  health chk   │
│                                            Run /help    list all     │
│                                                                      │
│                                          Recent activity             │
│                                          ────────────────────────    │
│                                            /scan                     │
│                                            /run llama3.2:3b          │
╰──────────────────────────────────────────────────────────────────────╯
  Type /help for commands · /exit or Ctrl+D to quit

tinillm> /scan

  LLM Capability Matrix

  Model    Fit        Best Quant   Mem Needed   Tokens/sec
  ~1B      Perfect    Q8_0          1.9 GB        580 t/s
  ~3B      Perfect    Q8_0          3.8 GB        195 t/s
  ~7B      Perfect    Q6_K          6.2 GB         88 t/s
  ~13B     Perfect    Q5_K_M       10.1 GB         47 t/s
  ~34B     Good       Q4_K_M       21.8 GB         18 t/s

tinillm>

Works on Linux, macOS, and Windows. No GPU required — CPU-only machines are supported too.


Install

pipx install tinillm     # recommended: isolated per-tool environment
# or
pip install tinillm

Requires Python 3.11+. No other tools needed.


Usage

Launch the tool with a single command:

tinillm

Inside the REPL, every feature is a slash command:

Command What it does
/scan Scan hardware and show which LLM sizes fit
/scan --verbose Include model sizes that don't fit
/scan --json Machine-readable JSON output
/models Browse real LLM models and see which fit
/models --fits-only Hide models that don't fit
/models --ollama Show which models are installed in local Ollama
/run Pick a compatible model interactively and run it
/run llama3.2:3b Launch a specific model directly
/suggest --use-case coding Personalised model recommendation
/doctor System health check (hardware + Ollama)
/help List every command
/clear Clear the terminal
/exit Quit (or Ctrl+D)

Tab-completion works on slash commands, subcommands, and flags.


First launch

The first time you run tinillm, it automatically runs /scan for you so you see your hardware capabilities immediately. On subsequent launches, just the welcome panel appears.


GPU support

Vendor Detection method
NVIDIA nvidia-smi → sysfs fallback
AMD rocm-smi → sysfs fallback
Apple Silicon system_profiler (unified memory)
Intel Arc sysfs + lspci
Windows (all) PowerShell WMI
Any vulkaninfo last-resort fallback

Fit levels explained

Level Meaning
Perfect Fits comfortably at Q4_K_M or better with ≥20% VRAM headroom
Good Fits but tightly
Marginal Fits only at heavy compression / reduced context, or CPU-only
TooTight Won't fit under any quantisation

Versioning

Version Feature
1.9 Dropped RAG — focused on hardware scanning + model runner ← current
1.8 Interactive REPL (single entry-point, slash commands)
1.7 Added RAG (/index, /ask, /rag)
1.1 First feature — hardware scanner

Part of the tini* family

Tool What it does
tiniRAG Privacy-first RAG CLI
tinillm Interactive LLM + hardware tool

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

tinillm-2.1.0.tar.gz (110.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

tinillm-2.1.0-py3-none-any.whl (104.6 kB view details)

Uploaded Python 3

File details

Details for the file tinillm-2.1.0.tar.gz.

File metadata

  • Download URL: tinillm-2.1.0.tar.gz
  • Upload date:
  • Size: 110.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.3

File hashes

Hashes for tinillm-2.1.0.tar.gz
Algorithm Hash digest
SHA256 2401401fa3f91c2cd6e1d34fb0bbc3491aedcdee52d3da8929e5bc6b29e55170
MD5 e8e9b85f0d15ed4cc00c2861deb30d0a
BLAKE2b-256 0595cb93c8ec208aa2744f4904feffcf1172f26f85296655f8205196f80ffb39

See more details on using hashes here.

File details

Details for the file tinillm-2.1.0-py3-none-any.whl.

File metadata

  • Download URL: tinillm-2.1.0-py3-none-any.whl
  • Upload date:
  • Size: 104.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.3

File hashes

Hashes for tinillm-2.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 39c0765b9c30c0c4d4036bb33511eb2cb91f96610bfdb65497d54d7e71261fd2
MD5 e9f13e3cfcf782a99bdc2a51275407f6
BLAKE2b-256 77b549f6673a2d7da1f62e1987984e99b003849bcef7c2c3234db1c74450c16c

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page