Skip to main content

CLI to manage local LLM inference endpoints

Project description

LMRunner

LMRunner is a lightweight CLI for starting and managing local LLM inference endpoints backed by llama.cpp.

Installation

From source

git clone https://github.com/jschw/LMRunner.git
cd LMRunner
python -m pip install -e .

With pip

If you want to compile llama-server by yourself or just download the binaries, just do:

python -m pip install --upgrade lmrunner

You will have to set the path according in llm_server_config.json . Standard path of llama.cpp is /lmrunner/Llamacpp/llama.cpp/build/bin/llama-server .

If you want to install the python-bindings for llama.cpp (easier but may not up-to-date):

python -m pip install --upgrade lmrunner[llamacppbindings]

Run

./lmrunner

Commands

Commands are entered inside the interactive prompt and always start with /.

Command Function
/getconfigpaths Output the paths of LLM config files.
/editlmconf Open llm_config.json in the default text editor.
/editserverconf Open llm_server_config.json in the default text editor.
/refreshconf Reload llm_config.json and llm_server_config.json.
/updatemodels Update the LLM model catalog from GitHub and print the available models.
/listendpoints List all available LLM endpoint configs.
/startendpoint <name> Start a specific LLM endpoint by config name.
/restartendpoint <name> Restart a specific LLM endpoint by config name.
/stopendpoint <name> Stop a specific LLM endpoint by config name.
/stopallendpnts Stop all LLM inference endpoints.
/llmstatus Show the status of local LLM inference endpoints.
/setautostartendpoint <name> Set a specific LLM endpoint for autostart on next startup.
/help Show the help message.
/exit Exit the CLI and stop running endpoints.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

lmrunner-0.1.0.tar.gz (9.3 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

lmrunner-0.1.0-py3-none-any.whl (9.4 kB view details)

Uploaded Python 3

File details

Details for the file lmrunner-0.1.0.tar.gz.

File metadata

  • Download URL: lmrunner-0.1.0.tar.gz
  • Upload date:
  • Size: 9.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.13

File hashes

Hashes for lmrunner-0.1.0.tar.gz
Algorithm Hash digest
SHA256 96e77b5e3495cfa4ab3ad47acb12f78124ecc1de44df4cca73c1b86171cd1918
MD5 6e9e5ee2c580fc9594ce7ccb353297ae
BLAKE2b-256 bd891749dda2dd56edb429697abc6f6a2e0f31cc484cfc1a3b8a3b6a53bd625d

See more details on using hashes here.

File details

Details for the file lmrunner-0.1.0-py3-none-any.whl.

File metadata

  • Download URL: lmrunner-0.1.0-py3-none-any.whl
  • Upload date:
  • Size: 9.4 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.13

File hashes

Hashes for lmrunner-0.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 7bd24a5f64e127615d1cc94f3b15673bf5a2ad65d8ddc4a4d72d2907ebf4cf67
MD5 6f046c5a5a39b0f29da17d9153b5409b
BLAKE2b-256 980995b279f057e33d3772a3d1e6bf66e7d2de7e5603c33c88867fb835214557

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page