Skip to main content

gguf connector core built on llama.cpp

Project description

llama-core

Static Badge

This is a solo llama connector also; being able to work independently.

install via (pip/pip3):

pip install llama-core

run it by (python/python3):

python -m llama_core

Prompt to user interface selection menu above; while chosen, GGUF file(s) in the current directory will be searched and detected (if any) as below.

include interface selector to your code by adding:

from llama_core import menu

include gguf reader to your code by adding:

from llama_core import reader

include gguf writer to your code by adding:

from llama_core import writer

remark(s)

Other functions are same as llama-cpp-python; for CUDA(GPU, Nvida) and Metal(M1/M2, Apple) supported settings, please specify CMAKE_ARGS following Abetlen's repo below; if you want to install it by source file (under releases), you should opt to do it by .tar.gz file (then build your machine-customized installable package) rather than .whl (wheel; a pre-built binary package) with an appropriate cmake tag(s).

references

repo llama-cpp-python llama.cpp page gguf.us

build from llama_core-(version).tar.gz (examples below are for CPU)

According to the latest note inside vs code, msys64 was recommended by Microsoft; or you could opt w64devkit or etc. as source/location of your gcc and g++ compilers.

for windows user(s):

$env:CMAKE_GENERATOR = "MinGW Makefiles"
$env:CMAKE_ARGS = "-DCMAKE_C_COMPILER=C:/msys64/mingw64/bin/gcc.exe -DCMAKE_CXX_COMPILER=C:/msys64/mingw64/bin/g++.exe"
pip install llama_core-(version).tar.gz

In mac, xcode command line tools were recommended by Apple for dealing all coding related issue(s); or you could bypass it for your own good/preference.

for mac user(s):

pip3 install llama_core-(version).tar.gz

Make sure your gcc and g++ are >=11; you can check it by: gcc --version and g++ --version; other setting(s) include: cmake>=3.21, etc.; however, if you opt to install it by the pre-built wheel (.whl) file then you don't need to worry about that.

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

llama_core-0.3.2.tar.gz (11.1 MB view details)

Uploaded Source

Built Distributions

llama_core-0.3.2-cp312-cp312-macosx_11_0_x86_64.whl (3.1 MB view details)

Uploaded CPython 3.12 macOS 11.0+ x86-64

llama_core-0.3.2-cp311-cp311-win_amd64.whl (3.7 MB view details)

Uploaded CPython 3.11 Windows x86-64

File details

Details for the file llama_core-0.3.2.tar.gz.

File metadata

  • Download URL: llama_core-0.3.2.tar.gz
  • Upload date:
  • Size: 11.1 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.11.4

File hashes

Hashes for llama_core-0.3.2.tar.gz
Algorithm Hash digest
SHA256 40b86bd45b9fcdcc1988c142390089e3cd0678ebc6aa9e005d30f3e8cf011948
MD5 f7f4604a2a508f7a6c98833a780d417d
BLAKE2b-256 2a0eaa2414538c3c0611eaf948df6402da957ee1ce144314d554ab1c45f7fbbb

See more details on using hashes here.

File details

Details for the file llama_core-0.3.2-cp312-cp312-macosx_11_0_x86_64.whl.

File metadata

File hashes

Hashes for llama_core-0.3.2-cp312-cp312-macosx_11_0_x86_64.whl
Algorithm Hash digest
SHA256 229c6f29e3dbe58471af93f8a682c68d7868c1086e3611de3697be0d6240a8dd
MD5 00837d1d602b1b5f1219891c1871acb4
BLAKE2b-256 cdfde9b18ab7387f5cabf3c663891a6a49402de637cb7359d5173933b3902bc1

See more details on using hashes here.

File details

Details for the file llama_core-0.3.2-cp311-cp311-win_amd64.whl.

File metadata

File hashes

Hashes for llama_core-0.3.2-cp311-cp311-win_amd64.whl
Algorithm Hash digest
SHA256 67ba3028157a472281f3fdfc7aa51f00e1af56f86474a65d5c2d95ad778c70ee
MD5 7be3413537b80a7d821557ba494fe21c
BLAKE2b-256 04d9bdd9e19b86f0474657f1d0dba7fa3df6bb6c05c50fd17821c1762fde14dd

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page