Skip to main content

gguf connector core built on llama.cpp

Project description

llama-core

Static Badge

This is a solo llama connector also; being able to work independently.

install via (pip/pip3):

pip install llama-core

run it by (python/python3):

python -m llama_core

Prompt to user interface selection menu above; while chosen, GGUF file(s) in the current directory will be searched and detected (if any) as below.

include interface selector to your code by adding:

from llama_core import menu

include gguf reader to your code by adding:

from llama_core import reader

include gguf writer to your code by adding:

from llama_core import writer

remark(s)

Other functions are same as llama-cpp-python; for CUDA(GPU, Nvida) and Metal(M1/M2/M3, Apple) supported settings, please specify CMAKE_ARGS following Abetlen's repo below; if you want to install it by source file (under releases), you should opt to do it by .tar.gz file (then build your machine-customized installable package) rather than .whl (wheel; a pre-built binary package) with an appropriate cmake tag(s).

references

repo llama-cpp-python llama.cpp page gguf.us

build from llama_core-(version).tar.gz (examples for CPU setup below)

According to the latest note inside vs code, msys64 was recommended by Microsoft; or you could opt w64devkit or etc. as source/location of your gcc and g++ compilers.

for windows user(s):

$env:CMAKE_GENERATOR = "MinGW Makefiles"
$env:CMAKE_ARGS = "-DCMAKE_C_COMPILER=C:/msys64/mingw64/bin/gcc.exe -DCMAKE_CXX_COMPILER=C:/msys64/mingw64/bin/g++.exe"
pip install llama_core-(version).tar.gz

In mac, xcode command line tools were recommended by Apple for dealing all coding related issue(s); or you could bypass it for your own good/preference.

for mac user(s):

pip3 install llama_core-(version).tar.gz

example setup for metal - faster performance

Metal (M1/M2/M3 - Apple)

CMAKE_ARGS="-DGGML_METAL=on" pip3 install llama_core-(version).tar.gz

example setup for cuda - faster x2; depends on your model (how rich you are)

Cuda (GPU - Nvida)

CMAKE_ARGS="-DGGML_CUDA=on" pip install llama_core-(version).tar.gz

Make sure your gcc and g++ are >=11; you can check it by: gcc --version and g++ --version; other setting(s) include: cmake>=3.21, etc.; however, if you opt to install it by the pre-built wheel (.whl) file then you don't need to worry about that.

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

llama_core-0.3.7.tar.gz (64.0 MB view details)

Uploaded Source

Built Distributions

llama_core-0.3.7-cp312-cp312-macosx_14_0_arm64.whl (3.5 MB view details)

Uploaded CPython 3.12 macOS 14.0+ ARM64

llama_core-0.3.7-cp312-cp312-macosx_11_0_x86_64.whl (3.9 MB view details)

Uploaded CPython 3.12 macOS 11.0+ x86-64

llama_core-0.3.7-cp311-cp311-win_amd64.whl (3.8 MB view details)

Uploaded CPython 3.11 Windows x86-64

File details

Details for the file llama_core-0.3.7.tar.gz.

File metadata

  • Download URL: llama_core-0.3.7.tar.gz
  • Upload date:
  • Size: 64.0 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.11.4

File hashes

Hashes for llama_core-0.3.7.tar.gz
Algorithm Hash digest
SHA256 5780eae31ca0dfbbb0137daf6f1c50c6ee678427fbf0ad50b5019529d0cc5d27
MD5 d5105d9fc08d30808c468036284d2157
BLAKE2b-256 cdcf8f857db5e6380de3f2ce77e0c13a14ab0e2403335e8bd919d9d786789ec2

See more details on using hashes here.

File details

Details for the file llama_core-0.3.7-cp312-cp312-macosx_14_0_arm64.whl.

File metadata

File hashes

Hashes for llama_core-0.3.7-cp312-cp312-macosx_14_0_arm64.whl
Algorithm Hash digest
SHA256 55269707cf4b10fa2467879db31034bec0d334af6dbe03f061e088ee89c71a30
MD5 cadbe06634209440c51c9ac6f3395623
BLAKE2b-256 68ef7911db6f878473d50fa505ae433af19353fadf6ab0e546df3d7c5d95b7cf

See more details on using hashes here.

File details

Details for the file llama_core-0.3.7-cp312-cp312-macosx_11_0_x86_64.whl.

File metadata

File hashes

Hashes for llama_core-0.3.7-cp312-cp312-macosx_11_0_x86_64.whl
Algorithm Hash digest
SHA256 71e9b567ba50d4d62a9afb5a471fa233ae52c7ad2d0b2de86432a30507c04591
MD5 310b1163532dce78e5aeed88df9a9451
BLAKE2b-256 ba081925fc4a6841f5dced68b64398a20d76a57b323d07c12f1ebea0d9b12746

See more details on using hashes here.

File details

Details for the file llama_core-0.3.7-cp311-cp311-win_amd64.whl.

File metadata

File hashes

Hashes for llama_core-0.3.7-cp311-cp311-win_amd64.whl
Algorithm Hash digest
SHA256 40d18c4bc46f40e2d41aee5a4a2c0f4f59b66459e19d4f588717982a5371c6a4
MD5 8be1a14b193b5cc30d76b30ec06ae706
BLAKE2b-256 38c38aa942c24ed3646ffaa23f61772f4f53ca94893c779fc0dd8c92b4f312df

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page