vLLM plugin for Qwerky AI MambaInLlama hybrid models
Project description
Qwerky vLLM Models
A vLLM plugin for serving Qwerky AI's MambaInLlama hybrid models without the --trust-remote-code flag.
Installation
pip install vllm qwerky-vllm-models
Usage
After installing, serve Qwerky models with vLLM:
vllm serve QwerkyAI/Qwerky-Llama3.2-Mamba-3B-Llama3.3-70B-base-distill --max-model-len 4096
The plugin automatically registers the model architecture with vLLM on import.
Supported Models
QwerkyAI/Qwerky-Llama3.2-Mamba-3B-Llama3.3-70B-base-distill
How It Works
This package uses vLLM's plugin system (vllm.general_plugins entry point) to register the MambaInLlama model architecture. This means:
- No fork of vLLM required
- No
--trust-remote-codeflag needed - Works with standard vLLM installation
- Uses vLLM's native Triton-accelerated Mamba kernels
Requirements
- Python >= 3.10
- vLLM >= 0.14.0
- PyTorch >= 2.0.0
Changelog
0.2.8
- Fixed dtype mismatch in SSM scan:
F.softplus/torch.expcompute in float32, now cast back to original dtype - This caused "expected BFloat16 but found Float" error in einsum
0.2.7
- Fixed tensor broadcasting bug in
_ssm_scan:A.unsqueeze(0).unsqueeze(-1)->A.unsqueeze(0).unsqueeze(2) - This caused shape mismatch (8192 vs 16) during SSM discretization
0.2.6
- Added
embed_input_idsmethod required by vLLM'sVllmModelForTextGenerationinterface - This was the root cause of "This model does not support
--runner generate" error
0.2.5
- Fixed vLLM runner detection: added
MambaInLlamaMambaForCausalLMalias for HF config compatibility - Added proper protocol inheritance (
HasInnerState,IsHybrid) fromvllm.model_executor.models.interfaces - Fixed class variable type hints (
ClassVar[Literal[True]]) for vLLM model inspection - Simplified model registration code
0.2.4
- Complete architecture rewrite with explicit state cache management
- Separate prefill and decode paths for Mamba layers
- Grouped-head Mamba support (
num_xb_head,num_C_head,repeat_group) - Pure PyTorch SSM implementation (preparing for vLLM Triton op integration)
0.2.3
- Fixed
d_xbdefault value computation in configuration - Removed unsupported
device/dtypekwargs from RMSNorm calls
0.2.2
- Fixed vLLM 0.14+ compatibility issues with Mamba ops API
0.2.1
- Updated README, removed SFT model reference
0.2.0
- Initial public release with vLLM plugin system integration
License
Apache 2.0
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file qwerky_vllm_models-0.2.8.tar.gz.
File metadata
- Download URL: qwerky_vllm_models-0.2.8.tar.gz
- Upload date:
- Size: 15.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.3
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
9755ed983bf84ef0f2ae2c50439ef6474645b6d6a56defb7217243a1ede9eb7b
|
|
| MD5 |
15ab3227925dfc858a9b8256ea91a7aa
|
|
| BLAKE2b-256 |
59ad280edc2bab02effdffc9cbc5d9913f803aa867130334ba6b9e82414d23b5
|
File details
Details for the file qwerky_vllm_models-0.2.8-py3-none-any.whl.
File metadata
- Download URL: qwerky_vllm_models-0.2.8-py3-none-any.whl
- Upload date:
- Size: 15.9 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.3
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
e202feac1360351d33de0d34e0ce6788331cf46c3fdfe470a0d68720ee6aaa7b
|
|
| MD5 |
d5e6405c7dd13f2204d296b9b47be2e0
|
|
| BLAKE2b-256 |
971a6e2581a2f708c81c8789f6704b1c1cb5f89d9fab162a062cbbfefb0b94fa
|