Skip to main content

Wrapper for simplified use of Llama2 GGUF quantized models.

Project description

gguf_llama

Provides a LlamaAI class with Python interface for generating text using Llama models.

Features

  • Load Llama models and tokenizers automatically from gguf file
  • Generate text completions for prompts
  • Automatically adjust model size to fit longer prompts up to a specific limit
  • Convenient methods for tokenizing and untokenizing text
  • Fix text formatting issues before generating

Usage

from llama_ai import LlamaAI

ai = LlamaAI("my_model.gguf", max_tokens=500, max_input_tokens=100)"

Generate text by calling infer():

text = ai.infer("Once upon a time")  
print(text)"

Installation

pip install gguf_llama

Documentation

See the API documentation for full details on classes and methods.

Contributing

Contributions are welcome! Open an issue or PR to improve gguf_llama.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

gguf_llama-0.0.18.tar.gz (4.6 kB view details)

Uploaded Source

Built Distribution

gguf_llama-0.0.18-py3-none-any.whl (4.7 kB view details)

Uploaded Python 3

File details

Details for the file gguf_llama-0.0.18.tar.gz.

File metadata

  • Download URL: gguf_llama-0.0.18.tar.gz
  • Upload date:
  • Size: 4.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.9.18

File hashes

Hashes for gguf_llama-0.0.18.tar.gz
Algorithm Hash digest
SHA256 485c467705699bcc7f4f77034167a33ae4380da73bb10ac5ca93817cbc87fdde
MD5 3202b9ceb859947633a4e197116e2ecd
BLAKE2b-256 c294548a808b940863e91e4a377d9765d4ca5966e02112602d36b275ad1c2682

See more details on using hashes here.

File details

Details for the file gguf_llama-0.0.18-py3-none-any.whl.

File metadata

  • Download URL: gguf_llama-0.0.18-py3-none-any.whl
  • Upload date:
  • Size: 4.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.9.18

File hashes

Hashes for gguf_llama-0.0.18-py3-none-any.whl
Algorithm Hash digest
SHA256 77f0ae941b1cf4766c25b5071ad718ceab8a31abb0dd28e852227c76785e8596
MD5 608e42ce2bfce8ad00b4e21967f59e2f
BLAKE2b-256 07bd9a12ace7bdfc65d458c0d9a19c6394172bd187e779dcb865bb70364120c1

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page