Skip to main content

ollama proxy to other cloud model service

Project description

ollama-proxy

ollama-proxy is a service that adheres to the Ollama external interface protocol. It enables connections to various large model services, such as glm-4-plus and deepseek-coder-v2, through the Ollama interface protocol. The primary goal of this project is to facilitate integration with zed.dev, allowing it to utilize additional large model services beyond the official configuration.

Installation

To install ollama-proxy, a command-line tool, you can use the following command:

pip install ollama-proxy

Command Line Usage

The ollama-proxy command-line interface (CLI) allows you to run the proxy server with specific configurations. Below are the details on how to use the CLI, including available options and their descriptions.

Command Structure

To run the ollama-proxy, use the following command structure:

ollama-proxy <model_name> --config <config_path> --host <host> --port <port>

Arguments

  • model_name:

    • Type: String
    • Description: A required argument that specifies the name of the model you want to run.
  • config:

    • Type: String
    • Default: keys.toml
    • Description: The path to the TOML configuration file.

Options

  • --host:

    • Type: String
    • Default: 127.0.0.1
    • Description: The host address for the server. For zed.dev, only localhost is allowed.
  • --port:

    • Type: Integer
    • Default: 11434
    • Description: For zed.dev, the port number must be 11434.

Example Usage

To start the ollama-proxy with a specific model and configuration, you can run:

ollama-proxy glm-4-plus --config path/to/config.toml --host localhost --port 11434 --reload

In this example:

  • glm-4-plus is the name of the model you want to run.
  • The configuration file is located at path/to/config.toml.
  • The server will listen on localhost at port 11434.
  • Hot reloading is enabled.

Notes

  • Ensure that the specified configuration file exists and is correctly formatted in TOML.
  • The server will print logs to the console, indicating its status and any errors that may occur.

Configuration TOML File

To configure ollama-proxy, create a configuration file in TOML format. This file allows you to specify various settings for the proxy.

Example Configuration (config.toml)

[glm-4-plus]
provider = "zhipu"
url = "https://open.bigmodel.cn/api/paas/v4/chat/completions"
api_key = "your_api_key_here"

[deepseek-coder-v2]
provider = "deepseek"
url = "https://api.deepseek.com/chat/completions"
api_key = "your_api_key_here"
  • The section name corresponds to the model name, such as glm-4-plus or deepseek-coder-v2.
  • The provider specifies the name of the large model service provider.
  • The url is the endpoint of the large model service provider.
  • The api_key is the API key for the large model service provider.

When you run ollama-proxy, it will load the configuration file. For example, running ollama-proxy glm-4-plus will use glm-4-plus as the running model name. If glm-4-plus is not specified in the configuration file, an error will be raised.

TODO

  • Support more large model providers
  • Add xinference as a local model provider
  • Support cursor

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

ollama_proxy-0.2.1.tar.gz (11.9 kB view details)

Uploaded Source

Built Distribution

ollama_proxy-0.2.1-py3-none-any.whl (13.7 kB view details)

Uploaded Python 3

File details

Details for the file ollama_proxy-0.2.1.tar.gz.

File metadata

  • Download URL: ollama_proxy-0.2.1.tar.gz
  • Upload date:
  • Size: 11.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.12.6

File hashes

Hashes for ollama_proxy-0.2.1.tar.gz
Algorithm Hash digest
SHA256 899d68fdb95b3b710b0ff548c5eeb32d59e2bff1be63b6154d366c993fc680c6
MD5 201c0d95bfce95720721eacf29111bc0
BLAKE2b-256 57b0351cfeed130948915cec3097859044803b9d330dc64c22e73e3cd5fd9463

See more details on using hashes here.

File details

Details for the file ollama_proxy-0.2.1-py3-none-any.whl.

File metadata

File hashes

Hashes for ollama_proxy-0.2.1-py3-none-any.whl
Algorithm Hash digest
SHA256 3ce4a5dbf8c1e660849cb6d52402f6b4637cd03a8b766b5c91ee3b9afa849eb3
MD5 d38ddbca1e923ea6d1a065e6258306d9
BLAKE2b-256 cb1c9ba3cfb2db220c4d55ad3c92248a54df2b57b195409c577cd76e471cd75b

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page