Skip to main content

gguf connector core built on llama.cpp

Project description

llama-core

Static Badge

This is a solo llama connector also; being able to work independently.

install via (pip/pip3):

pip install llama-core

run it by (python/python3):

python -m llama_core

Prompt to user interface selection menu above; while chosen, GGUF file(s) in the current directory will be searched and detected (if any) as below.

include interface selector to your code by adding:

from llama_core import menu

include gguf reader to your code by adding:

from llama_core import reader

include gguf writer to your code by adding:

from llama_core import writer

remark(s)

Other functions are same as llama-cpp-python; for CUDA(GPU, Nvida) and Metal(M1/M2/M3, Apple) supported settings, please specify CMAKE_ARGS following Abetlen's repo below; if you want to install it by source file (under releases), you should opt to do it by .tar.gz file (then build your machine-customized installable package) rather than .whl (wheel; a pre-built binary package) with an appropriate cmake tag(s).

references

repo llama-cpp-python llama.cpp page gguf.us

build from llama_core-(version).tar.gz (examples for CPU setup below)

According to the latest note inside vs code, msys64 was recommended by Microsoft; or you could opt w64devkit or etc. as source/location of your gcc and g++ compilers.

for windows user(s):

$env:CMAKE_GENERATOR = "MinGW Makefiles"
$env:CMAKE_ARGS = "-DCMAKE_C_COMPILER=C:/msys64/mingw64/bin/gcc.exe -DCMAKE_CXX_COMPILER=C:/msys64/mingw64/bin/g++.exe"
pip install llama_core-(version).tar.gz

In mac, xcode command line tools were recommended by Apple for dealing all coding related issue(s); or you could bypass it for your own good/preference.

for mac user(s):

pip3 install llama_core-(version).tar.gz

for high (just a little bit better) performance seeker(s):

example setup for metal (M1/M2/M3 - Apple) - faster

CMAKE_ARGS="-DGGML_METAL=on" pip3 install llama_core-(version).tar.gz

example setup for cuda (GPU - Nvida) - faster x2; depends on your model (how rich you are)

CMAKE_ARGS="-DGGML_CUDA=on" pip install llama_core-(version).tar.gz

make sure your gcc and g++ are >=11; you can check it by: gcc --version and g++ --version; other setting(s) include: cmake>=3.21, etc.; however, if you opt to install it by the pre-built wheel (.whl) file then you don't need to worry about that.

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

llama_core-0.3.8.tar.gz (64.0 MB view details)

Uploaded Source

Built Distributions

llama_core-0.3.8-cp312-cp312-macosx_14_0_arm64.whl (3.5 MB view details)

Uploaded CPython 3.12 macOS 14.0+ ARM64

llama_core-0.3.8-cp312-cp312-macosx_11_0_x86_64.whl (3.3 MB view details)

Uploaded CPython 3.12 macOS 11.0+ x86-64

llama_core-0.3.8-cp311-cp311-win_amd64.whl (3.8 MB view details)

Uploaded CPython 3.11 Windows x86-64

File details

Details for the file llama_core-0.3.8.tar.gz.

File metadata

  • Download URL: llama_core-0.3.8.tar.gz
  • Upload date:
  • Size: 64.0 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.11.4

File hashes

Hashes for llama_core-0.3.8.tar.gz
Algorithm Hash digest
SHA256 ed0f2eae83bd3d1cd0904120f09d2557a28fc7d46fcbd2102afd89db667386ac
MD5 3847d81a264a64195be1966a176f8e86
BLAKE2b-256 81d0a33f84f3a8fd004c768402de67a69c7e51ba95576b6fce21804b5cc90ab5

See more details on using hashes here.

File details

Details for the file llama_core-0.3.8-cp312-cp312-macosx_14_0_arm64.whl.

File metadata

File hashes

Hashes for llama_core-0.3.8-cp312-cp312-macosx_14_0_arm64.whl
Algorithm Hash digest
SHA256 f3c856112fb5973931d28c80375a3a487e48a377bdb9e14a8c22bdeef5a009df
MD5 705103a37e6f5655d660428c72834434
BLAKE2b-256 33472f2044a13b349b29125682b5d4d5db41c4838106d0295f88efcc619ef060

See more details on using hashes here.

File details

Details for the file llama_core-0.3.8-cp312-cp312-macosx_11_0_x86_64.whl.

File metadata

File hashes

Hashes for llama_core-0.3.8-cp312-cp312-macosx_11_0_x86_64.whl
Algorithm Hash digest
SHA256 3774b8ac88ea5e61401d6fbfec79e9bf68195452c6a32bb5a4fcee3621a85f38
MD5 682585ce36f6e575a238d4e223957e0e
BLAKE2b-256 d84c8138b154c9664f159dd13497b0188b4031d0191ab2b70655e61c6d1ccb04

See more details on using hashes here.

File details

Details for the file llama_core-0.3.8-cp311-cp311-win_amd64.whl.

File metadata

File hashes

Hashes for llama_core-0.3.8-cp311-cp311-win_amd64.whl
Algorithm Hash digest
SHA256 3753dbfd46c0d65058398c93eb1813fe1d337bf6e3ca253e26e35c8b02e91489
MD5 bc4363dc495d9a95e478c3186f62f09c
BLAKE2b-256 7b50bb4981aef5e71eb46904760a64785af34c07b92d9dd61a7f9265e34be438

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page