Skip to main content

Add your description here

Project description

llama-cpp-server-py-core

Describe your project here.

Some tools

Convert huggingface model to gguf model

rye run hf2gguf /opt/models/llm/qwen/Qwen2.5-Coder-14B-Instruct --outfile /opt/models/llm/qwen/Qwen2.5-Coder-14B-Instruct-f16.gguf

Quantize gguf model

rye run quantize /opt/models/llm/qwen/Qwen2.5-Coder-14B-Instruct-f16.gguf /opt/models/llm/qwen/Qwen2.5-Coder-14B-Instruct-Q4_k_m.gguf Q4_k_m

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

llama_cpp_server_py_core-0.1.2.tar.gz (33.9 MB view details)

Uploaded Source

File details

Details for the file llama_cpp_server_py_core-0.1.2.tar.gz.

File metadata

  • Download URL: llama_cpp_server_py_core-0.1.2.tar.gz
  • Upload date:
  • Size: 33.9 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.0.1 CPython/3.10.15

File hashes

Hashes for llama_cpp_server_py_core-0.1.2.tar.gz
Algorithm Hash digest
SHA256 5d1ca66d351286c7995ca47c4aaa7cb8d117120a35cfb52d97b5b45f34c7a9dc
MD5 2c19420c489b6fad0a1694ae3d626e91
BLAKE2b-256 e9af068376aa50a4312f67121b06f3ff360c98a9692b5b7e99f3ea7e41989512

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page