Skip to main content

Wrapper for simplified use of Llama2 GGUF quantized models.

Project description

gguf_llama

Provides a LlamaAI class with Python interface for generating text using Llama models.

Features

  • Load Llama models and tokenizers automatically from gguf file
  • Generate text completions for prompts
  • Automatically adjust model size to fit longer prompts up to a specific limit
  • Convenient methods for tokenizing and untokenizing text
  • Fix text formatting issues before generating

Usage

from llama_ai import LlamaAI

ai = LlamaAI("my_model.gguf", max_tokens=500, max_input_tokens=100)"

Generate text by calling infer():

text = ai.infer("Once upon a time")  
print(text)"

Installation

pip install gguf_llama

Documentation

See the API documentation for full details on classes and methods.

Contributing

Contributions are welcome! Open an issue or PR to improve gguf_llama.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

gguf_llama-0.0.18.tar.gz (4.6 kB view hashes)

Uploaded Source

Built Distribution

gguf_llama-0.0.18-py3-none-any.whl (4.7 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page