Skip to main content

CLI application to manage local language model inference endpoints

Project description

LMRunner

LMRunner is a lightweight CLI for starting and managing local LLM inference endpoints backed by llama.cpp.

Installation

From source

git clone https://github.com/jschw/LMRunner.git
cd LMRunner
python -m pip install -e .

With pip

If you want to compile llama-server by yourself or just download the binaries, just do:

python -m pip install --upgrade lmrunner

You will have to set the path according in llm_server_config.json . Standard path of llama.cpp is /lmrunner/Llamacpp/llama.cpp/build/bin/llama-server .

If you want to install the python-bindings for llama.cpp (easier but may not up-to-date):

python -m pip install --upgrade lmrunner[llamacppbindings]

Run

./lmrunner

Commands

Commands are entered inside the interactive prompt and always start with /.

Command Function
/getconfigpaths Output the paths of LLM config files.
/editlmconf Open llm_config.json in the default text editor.
/editserverconf Open llm_server_config.json in the default text editor.
/refreshconf Reload llm_config.json and llm_server_config.json.
/updatemodels Update the LLM model catalog from GitHub and print the available models.
/listendpoints List all available LLM endpoint configs.
/startendpoint <name> Start a specific LLM endpoint by config name.
/restartendpoint <name> Restart a specific LLM endpoint by config name.
/stopendpoint <name> Stop a specific LLM endpoint by config name.
/stopallendpnts Stop all LLM inference endpoints.
/llmstatus Show the status of local LLM inference endpoints.
/setautostartendpoint <name> Set a specific LLM endpoint for autostart on next startup.
/help Show the help message.
/exit Exit the CLI and stop running endpoints.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

lmrunner-0.2.0.tar.gz (9.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

lmrunner-0.2.0-py3-none-any.whl (9.6 kB view details)

Uploaded Python 3

File details

Details for the file lmrunner-0.2.0.tar.gz.

File metadata

  • Download URL: lmrunner-0.2.0.tar.gz
  • Upload date:
  • Size: 9.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.13

File hashes

Hashes for lmrunner-0.2.0.tar.gz
Algorithm Hash digest
SHA256 f813541802ffc1e034e10c2f5bb36b4df975dbd74bb715c912e5967c758930eb
MD5 3efe35313711f5e0853ab35899e65b84
BLAKE2b-256 573aac1542f540287282bb02ea1faffbf87f42d6281bf3246c5dded869282c62

See more details on using hashes here.

File details

Details for the file lmrunner-0.2.0-py3-none-any.whl.

File metadata

  • Download URL: lmrunner-0.2.0-py3-none-any.whl
  • Upload date:
  • Size: 9.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.13

File hashes

Hashes for lmrunner-0.2.0-py3-none-any.whl
Algorithm Hash digest
SHA256 8c7006dd36c6e2576151b5588e0bbc007405ce880b66c7c24960377c8839c817
MD5 46a0c38c2618b5f3b67909114a27e371
BLAKE2b-256 a44d58087b37fe7f9964bdd6db08e3717b93fd8624ec4063038bb9257461e95d

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page