Skip to main content

CLI to manage local language model inference endpoints

Project description

LMRunner

LMRunner is a lightweight CLI for starting and managing local LLM inference endpoints backed by llama.cpp.

Installation

From source

git clone https://github.com/jschw/LMRunner.git
cd LMRunner
python -m pip install -e .

With pip

If you want to compile llama-server by yourself or just download the binaries, just do:

python -m pip install --upgrade lmrunner

You will have to set the path according in llm_server_config.json . Standard path of llama.cpp is /lmrunner/Llamacpp/llama.cpp/build/bin/llama-server .

If you want to install the python-bindings for llama.cpp (easier but may not up-to-date):

python -m pip install --upgrade lmrunner[llamacppbindings]

Run

./lmrunner

Commands

Commands are entered inside the interactive prompt and always start with /.

Command Function
/getconfigpaths Output the paths of LLM config files.
/editlmconf Open llm_config.json in the default text editor.
/editserverconf Open llm_server_config.json in the default text editor.
/refreshconf Reload llm_config.json and llm_server_config.json.
/updatemodels Update the LLM model catalog from GitHub and print the available models.
/listendpoints List all available LLM endpoint configs.
/startendpoint <name> Start a specific LLM endpoint by config name.
/restartendpoint <name> Restart a specific LLM endpoint by config name.
/stopendpoint <name> Stop a specific LLM endpoint by config name.
/stopallendpnts Stop all LLM inference endpoints.
/llmstatus Show the status of local LLM inference endpoints.
/setautostartendpoint <name> Set a specific LLM endpoint for autostart on next startup.
/help Show the help message.
/exit Exit the CLI and stop running endpoints.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

lmrunner-0.1.1.tar.gz (9.3 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

lmrunner-0.1.1-py3-none-any.whl (9.4 kB view details)

Uploaded Python 3

File details

Details for the file lmrunner-0.1.1.tar.gz.

File metadata

  • Download URL: lmrunner-0.1.1.tar.gz
  • Upload date:
  • Size: 9.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.13

File hashes

Hashes for lmrunner-0.1.1.tar.gz
Algorithm Hash digest
SHA256 38025b5680877c292219161769f32e1f762cc54ade5d24acca12005be5ab6c10
MD5 cae933818cf034df8117c8665d57b8f1
BLAKE2b-256 8aae7a56c4c0fd56660052018758a1b16e0e5351c66460731fae07f41f50d64a

See more details on using hashes here.

File details

Details for the file lmrunner-0.1.1-py3-none-any.whl.

File metadata

  • Download URL: lmrunner-0.1.1-py3-none-any.whl
  • Upload date:
  • Size: 9.4 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.13

File hashes

Hashes for lmrunner-0.1.1-py3-none-any.whl
Algorithm Hash digest
SHA256 a6b2aa068ef109c513b222d89b6cc6ce6828ee53cf32f8b28bbbeb4312e5e1e4
MD5 ec71fe82adc8a4e628403622354f764f
BLAKE2b-256 9d9614748ede11544309dc4ab56fddd7a9482577e4c8c09bb220a81d44ac944d

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page