Core library for training Tlama models.
Project description
Tlama Core 🚀
Welcome to Tlama Core, the foundational repository for the Tlama models! This is the heart of our mission to create scalable, efficient, and cutting-edge AI models optimized for training and inference on GPUs. Whether you’re an AI researcher, developer, or GPU enthusiast, this repository is your playground to push the limits of performance, scalability, and innovation.
We believe in community-driven development—together, we can shape the future of AI. Join us to revolutionize machine learning with state-of-the-art optimizations for large-scale, high-performance models.
Why Tlama Core? 🤔
Tlama Core is the foundation for next-gen AI models, designed to enhance Tlama models with unmatched efficiency, power, and scalability. From high-performance computing to deep learning and robotics, we’re building the infrastructure for groundbreaking research and production-ready solutions.
Core Areas of Focus ⚙️
We’re targeting key optimizations to make Tlama models faster and more scalable. Explore our focus areas:
- Custom CUDA Kernels 🔥: Unlock hardware potential with tailored kernels for attention, matrix ops, and more.
- Mixed Precision Training 💎: Leverage Tensor Cores to train larger models faster.
- Distributed & Multi-GPU Support 🌐: Scale effortlessly with optimized multi-GPU training.
- Memory Optimizations 🧠: Use checkpointing and dynamic allocation for efficient large-scale training.
- Profiling Tools 🕵️♂️: Analyze and optimize performance with precision.
- Innovative Algorithms 💡: Push beyond PyTorch and cuBLAS with fresh approaches.
- Compression Techniques 📦: Lightweight models via quantization and pruning.
- Fine-tuning & Transfer Learning 🔄: Adapt models quickly to new tasks.
- Reinforcement Learning Support 🎮: Tools for RL experimentation and deployment.
- Research Support 🔬: Utilities for rapid experimentation and innovation.
Roadmap 📅
Join our journey to build the future of AI! Track our progress and contribute via our Roadmap Project Board and Issues.
| Phases | Focus | Key Goals | Status |
|---|---|---|---|
| 1 | Bases | Repo structure, docs, Migrate Tlama-124M | In Progress |
| 2 | Foundations | CUDA kernels, mixed precision, docs | Pending |
| 3 | Scalability | Multi-GPU, profiler, Tlama-500M | Pending |
| 4 | Innovation/Robotics | Optimized attention, RL, simulation | Pending |
| 5 | Consolidation | Tlama-1B, API, hackathon | Pending |
How to Set Up Tlama Core 💻
Get started easily on Windows, Linux, or macOS!
Prerequisites
- CUDA Toolkit (GPU support)
- Python 3.8+ (virtual env recommended)
- PyTorch (GPU version preferred)
- NVIDIA Driver (GPU-compatible)
- CMake (for CUDA kernels)
CUDA Toolkit Setup 🛠️
- Windows: Download from NVIDIA and install.
- Linux: Use
sudo apt install nvidia-cuda-toolkitor download from NVIDIA. - macOS: CPU-only (no CUDA support).
Installation Steps 🚀
- Clone the Repository:
git clone https://github.com/your-username/tlama-core.git cd tlama-core
- Set Up Environment:
python3 -m venv tlama-env source tlama-env/bin/activate # Linux/macOS .\tlama-env\Scripts\activate # Windows
- Install Dependencies:
pip install -r requirements.txt
- Install CUDA Kernels:
python setup.py install
- Verify:
python -c "import torch; print(torch.cuda.is_available())" # Should output True
Run Tlama 124M Model 🏃♂️
Test it out:
from transformers import AutoModel, AutoTokenizer
model = AutoModel.from_pretrained("eigencore/tlama-124M", trust_remote_code=True)
tokenizer = AutoTokenizer.from_pretrained("eigencore/tlama-124M", trust_remote_code=True)
prompt = "Once upon a time in a distant kingdom..."
input_ids = tokenizer.encode(prompt, return_tensors="pt")
output = model.generate(input_ids, max_length=50)
print(tokenizer.decode(output[0], skip_special_tokens=True))
More at Hugging Face.
How to Contribute 🌟
We’re all about collaboration! Here’s how to jump in:
- Check out the Roadmap Project Board.
- Pick an Issue to tackle.
- Follow our Contribution Guidelines.
- Join the convo on Discord.
New to contributing? Look for good-first-issue tags!
Our Vision 🌍
Our mission: scalable, efficient, high-performance AI models. We’re empowering researchers and developers to train bigger, deploy faster, and innovate freely. Learn more at Eigen Core.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file tlama_core-0.0.3.tar.gz.
File metadata
- Download URL: tlama_core-0.0.3.tar.gz
- Upload date:
- Size: 34.1 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.10.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
6b55ce7daa1bf5f043e0fdaa71a31cf3a0d2c40f65d1cff1712ad97cdae560ba
|
|
| MD5 |
8632c2f92fcc45161f6b42e88d63c8cf
|
|
| BLAKE2b-256 |
b49fa20d0b468423de6cea5196bff18df306b6984cb161d35fdb487c440de51a
|
File details
Details for the file tlama_core-0.0.3-py3-none-any.whl.
File metadata
- Download URL: tlama_core-0.0.3-py3-none-any.whl
- Upload date:
- Size: 36.6 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.10.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
8509a6ca1d74303889cc42ffe5be0132ef3f26182a8685789fc8548178d9058c
|
|
| MD5 |
87e52e332ec47c624da32c8ad4325bea
|
|
| BLAKE2b-256 |
efee7ad0a8ec4ef94b3f43cf413dfd442ba2dd739fa7b53a00005057b5a6fa0e
|