Skip to main content

A lightweight LLM inference framework

Project description

hllm - 轻量级 LLM 推理框架

在 CPU 上运行的简化推理框架

快速开始

from hllm import HLLM

# 初始化模型
model = HLLM(model_path="microsoft/Phi-3-mini-4k-instruct", device="cpu")

# 生成文本
result = model.generate("Write a short story about a robot.")
print(result)

目录结构

hllm/
├── hllm/              # 核心模块
│   ├── __init__.py
│   ├── model.py       # 模型加载与推理
│   ├── tokenizer.py   # 分词器封装
│   └── generate.py    # 生成逻辑
├── tests/             # 测试
└── examples/          # 示例

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distributions

No source distribution files available for this release.See tutorial on generating distribution archives.

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

light_llm_hp-0.1.0-py3-none-any.whl (5.9 kB view details)

Uploaded Python 3

File details

Details for the file light_llm_hp-0.1.0-py3-none-any.whl.

File metadata

  • Download URL: light_llm_hp-0.1.0-py3-none-any.whl
  • Upload date:
  • Size: 5.9 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.5

File hashes

Hashes for light_llm_hp-0.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 539ff2de52a14138ebc97a0fc0c7e247666adf1ac21dd6b8096f0665814e5583
MD5 43e0c9b189f6cf5cf2a3271163089670
BLAKE2b-256 e3f119099bd9cc8c2847e3175b6e61956150ef7c8bdf8d3c32bfa474697e664e

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page