High-performance Markov Chains & HMMs using OpenCL
Project description
⚡ MarkovGPU
Massive Scale Markov Models on Consumer Hardware
Run million-state HMMs on your laptop GPU.
No CUDA required • Hybrid CPU/GPU Backend • Production Ready
🌟 The Engine for Stochastic Intelligence
MarkovGPU is a high-performance probabilistic modeling library built for speed. It breaks the "NVIDIA Monopoly" by using OpenCL to accelerate Hidden Markov Models (HMM) and Markov Chains on any GPU—including AMD Radeon, Intel Arc, and Apple Silicon.
It doesn't just run; it thinks. The Smart Hybrid Backend automatically routes small tasks to the CPU (NumPy) and massive workloads to the GPU, giving you optimal performance at every scale.
🚀 Core Superpowers
| Feature | Magic Behind It |
|---|---|
| ⚡ Hardware Agnostic | Built on OpenCL — runs on AMD, Intel, NVIDIA, and Apple M1/M2/M3 chips. |
| 🧠 Smart Hybrid Backend | Auto-detects problem size ($N$). Uses NumPy for speed on small data, GPU for massive throughput. |
| 📉 Log-Space Stability | Implements Log-Sum-Exp kernels to prevent underflow on long time-series (1M+ steps). |
| 🕵️ Viterbi Decoding | Finds the "Hidden Truth" in noisy data (e.g., market regimes, DNA sequences) in milliseconds. |
| 🎓 Unsupervised Learning | Baum-Welch (EM) algorithm trains models directly on the GPU, learning rules from raw data. |
| 📦 Zero-Config Install | pip install markovgpu-rane. No driver hell. No CUDA toolkit nightmares. |
🏗️ Architecture: The Hybrid Pipeline
graph LR
A[User Code] -->|Request Fit/Predict| B{Smart Dispatcher}
B -->|Small N < 64| C["CPU Engine
(NumPy AVX2)"]
B -->|Large N >= 64| D["GPU Engine
(OpenCL Kernels)"]
C --> E[Result]
D --> E
subgraph GPU_Acceleration[GPU Acceleration]
D --> F[Matrix Multiply]
D --> G[Log-Sum-Exp]
D --> H[Parallel Viterbi]
end
The library handles the hardware. You handle the math.
⚡ Performance: Benchmarks
Task: Viterbi Decoding (64 Hidden States, 5000 Days of Data).
Hardware: AMD Ryzen 680M (Integrated Graphics).
| Engine | Execution Time | Speedup |
|---|---|---|
| 🐢 CPU (NumPy Optimized) | 5.06s | 1x |
| 🚀 GPU (MarkovGPU) | 0.82s | 6.2x |
⚙️ Quick Start in 30 Seconds
Installation
# Production
pip install markovgpu-rane
# Or for local development
uv pip install markovgpu-rane
1. Market Regime Detection (Viterbi)
Identify hidden "Bull" vs. "Bear" markets from noisy stock returns.
import numpy as np
from markovgpu import MarkovEngine
# 1. Setup the Rules (Transition Matrix)
# "Bull markets tend to stay Bullish (95%)"
trans_mat = np.array([[0.95, 0.05],
[0.10, 0.90]], dtype=np.float32)
# 2. Feed the Data (Observation Likelihoods)
# Shape: (1000 Days, 2 States)
obs_probs = np.random.rand(1000, 2).astype(np.float32)
# 3. Ignite the Engine
engine = MarkovEngine()
predicted_states = engine.decode_regime(trans_mat, obs_probs)
print("Detected Regimes:", predicted_states)
# Output: [0, 0, 0, 1, 1, 1, 0 ...]
2. Unsupervised Learning (Baum-Welch)
Train the AI to discover the hidden rules from raw data.
# The engine learns the Transition Matrix automatically
learned_matrix = engine.fit(
obs_probs,
n_states=2,
n_iters=100,
tolerance=1e-4
)
print("Discovered Rules:")
print(learned_matrix)
🔬 Technical Brilliance
1. The Log-Sum-Exp Kernel
Standard HMMs crash on long sequences because probabilities like $0.9^{1000}$ vanish to zero. We solved this by rewriting the entire GPU kernel in Log-Space:
// Actual OpenCL Kernel snippet
float log_add(float log_a, float log_b) {
float max_val = max(log_a, log_b);
return max_val + log1p(exp(min(log_a, log_b) - max_val));
}
→ Result: You can process sequences of infinite length without numerical collapse.
2. Parallel Viterbi
Instead of a slow Python loop, we launch $N$ threads (one per state) for every time step on the GPU, calculating the optimal path in parallel.
🛠️ Project Structure
markovgpu/
├── src/markovgpu/
│ ├── backend.py # The Brain (Smart Dispatcher)
│ ├── kernels.cl # The Muscle (OpenCL C Code)
│ └── __init__.py
├── tests/ # Unit Tests
├── pyproject.toml # Modern Packaging Config
└── README.md
🌱 Contributing
We welcome forks, issues, and PRs!
git clone https://github.com/wizardwithcodehazard/markov.git
cd markov
uv sync --dev
uv run pytest
📄 License
MIT License — Free to use, modify, and ship in commercial products.
MarkovGPU doesn’t just crunch numbers.
It discovers the hidden structure of reality.
Made with 🧡 by Sahil Rane
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file markovgpu_rane-0.4.2.tar.gz.
File metadata
- Download URL: markovgpu_rane-0.4.2.tar.gz
- Upload date:
- Size: 88.6 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
62f5054a7d3ab830da46173a23d7cd3e34d14f6dd7d487a39dba3a9c5bfc0596
|
|
| MD5 |
7a81a8c4732fc18729f0e58a908bcb48
|
|
| BLAKE2b-256 |
6773248b9f420e56cb0d9ce2c488b7302d8e6ed464b4a4fa7615eebbf4e540e8
|
Provenance
The following attestation bundles were made for markovgpu_rane-0.4.2.tar.gz:
Publisher:
publish.yml on wizardwithcodehazard/markov
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
markovgpu_rane-0.4.2.tar.gz -
Subject digest:
62f5054a7d3ab830da46173a23d7cd3e34d14f6dd7d487a39dba3a9c5bfc0596 - Sigstore transparency entry: 910539973
- Sigstore integration time:
-
Permalink:
wizardwithcodehazard/markov@79a5833cb6d2a2cf2d5103407fa86d49483d4c75 -
Branch / Tag:
refs/tags/v0.4.2 - Owner: https://github.com/wizardwithcodehazard
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@79a5833cb6d2a2cf2d5103407fa86d49483d4c75 -
Trigger Event:
push
-
Statement type:
File details
Details for the file markovgpu_rane-0.4.2-py3-none-any.whl.
File metadata
- Download URL: markovgpu_rane-0.4.2-py3-none-any.whl
- Upload date:
- Size: 9.0 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
ac3ff24998bd62bd17542ba4732fefdf1b3fbd0e5865c3291e9b4dda66d8da7f
|
|
| MD5 |
7de77b9d829d4d80e6219b9b5cc023b3
|
|
| BLAKE2b-256 |
df1ea1815dd992d82dbd35c8f25968e73e07ba4efe0befe1feb10aedad910a80
|
Provenance
The following attestation bundles were made for markovgpu_rane-0.4.2-py3-none-any.whl:
Publisher:
publish.yml on wizardwithcodehazard/markov
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
markovgpu_rane-0.4.2-py3-none-any.whl -
Subject digest:
ac3ff24998bd62bd17542ba4732fefdf1b3fbd0e5865c3291e9b4dda66d8da7f - Sigstore transparency entry: 910539977
- Sigstore integration time:
-
Permalink:
wizardwithcodehazard/markov@79a5833cb6d2a2cf2d5103407fa86d49483d4c75 -
Branch / Tag:
refs/tags/v0.4.2 - Owner: https://github.com/wizardwithcodehazard
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@79a5833cb6d2a2cf2d5103407fa86d49483d4c75 -
Trigger Event:
push
-
Statement type: