Skip to main content

High-performance Markov Chains & HMMs using OpenCL

Project description

MarkovGPU

Massive Scale Markov Models on Consumer Hardware

image

Run million-state HMMs on your laptop GPU.
No CUDA required • Hybrid CPU/GPU Backend • Production Ready

PyPI version Python 3.9+ License: MIT Build Status


🌟 The Engine for Stochastic Intelligence

MarkovGPU is a high-performance probabilistic modeling library built for speed. It breaks the "NVIDIA Monopoly" by using OpenCL to accelerate Hidden Markov Models (HMM) and Markov Chains on any GPU—including AMD Radeon, Intel Arc, and Apple Silicon.

It doesn't just run; it thinks. The Smart Hybrid Backend automatically routes small tasks to the CPU (NumPy) and massive workloads to the GPU, giving you optimal performance at every scale.


🚀 Core Superpowers

Feature Magic Behind It
Hardware Agnostic Built on OpenCL — runs on AMD, Intel, NVIDIA, and Apple M1/M2/M3 chips.
🧠 Smart Hybrid Backend Auto-detects problem size ($N$). Uses NumPy for speed on small data, GPU for massive throughput.
📉 Log-Space Stability Implements Log-Sum-Exp kernels to prevent underflow on long time-series (1M+ steps).
🕵️ Viterbi Decoding Finds the "Hidden Truth" in noisy data (e.g., market regimes, DNA sequences) in milliseconds.
🎓 Unsupervised Learning Baum-Welch (EM) algorithm trains models directly on the GPU, learning rules from raw data.
📦 Zero-Config Install pip install markovgpu-rane. No driver hell. No CUDA toolkit nightmares.

🏗️ Architecture: The Hybrid Pipeline

graph LR
    A[User Code] -->|Request Fit/Predict| B{Smart Dispatcher}
    B -->|Small N < 64| C["CPU Engine
    (NumPy AVX2)"]
    B -->|Large N >= 64| D["GPU Engine
    (OpenCL Kernels)"]
    C --> E[Result]
    D --> E
    subgraph GPU_Acceleration[GPU Acceleration]
    D --> F[Matrix Multiply]
    D --> G[Log-Sum-Exp]
    D --> H[Parallel Viterbi]
    end

The library handles the hardware. You handle the math.

⚡ Performance: Benchmarks

Task: Viterbi Decoding (64 Hidden States, 5000 Days of Data).
Hardware: AMD Ryzen 680M (Integrated Graphics).

Engine Execution Time Speedup
🐢 CPU (NumPy Optimized) 5.06s 1x
🚀 GPU (MarkovGPU) 0.82s 6.2x

⚙️ Quick Start in 30 Seconds

Installation

# Production
pip install markovgpu-rane

# Or for local development
uv pip install markovgpu-rane

1. Market Regime Detection (Viterbi)

Identify hidden "Bull" vs. "Bear" markets from noisy stock returns.

import numpy as np
from markovgpu import MarkovEngine

# 1. Setup the Rules (Transition Matrix)
# "Bull markets tend to stay Bullish (95%)"
trans_mat = np.array([[0.95, 0.05], 
                      [0.10, 0.90]], dtype=np.float32)

# 2. Feed the Data (Observation Likelihoods)
# Shape: (1000 Days, 2 States)
obs_probs = np.random.rand(1000, 2).astype(np.float32) 

# 3. Ignite the Engine
engine = MarkovEngine()
predicted_states = engine.decode_regime(trans_mat, obs_probs)

print("Detected Regimes:", predicted_states) 
# Output: [0, 0, 0, 1, 1, 1, 0 ...]

2. Unsupervised Learning (Baum-Welch)

Train the AI to discover the hidden rules from raw data.

# The engine learns the Transition Matrix automatically
learned_matrix = engine.fit(
    obs_probs, 
    n_states=2, 
    n_iters=100, 
    tolerance=1e-4
)

print("Discovered Rules:")
print(learned_matrix)

🔬 Technical Brilliance

1. The Log-Sum-Exp Kernel

Standard HMMs crash on long sequences because probabilities like $0.9^{1000}$ vanish to zero. We solved this by rewriting the entire GPU kernel in Log-Space:

// Actual OpenCL Kernel snippet
float log_add(float log_a, float log_b) {
    float max_val = max(log_a, log_b);
    return max_val + log1p(exp(min(log_a, log_b) - max_val));
}

Result: You can process sequences of infinite length without numerical collapse.

2. Parallel Viterbi

Instead of a slow Python loop, we launch $N$ threads (one per state) for every time step on the GPU, calculating the optimal path in parallel.


🛠️ Project Structure

markovgpu/
├── src/markovgpu/
│   ├── backend.py       # The Brain (Smart Dispatcher)
│   ├── kernels.cl       # The Muscle (OpenCL C Code)
│   └── __init__.py
├── tests/               # Unit Tests
├── pyproject.toml       # Modern Packaging Config
└── README.md

🌱 Contributing

We welcome forks, issues, and PRs!

git clone https://github.com/wizardwithcodehazard/markov.git
cd markov
uv sync --dev
uv run pytest

📄 License

MIT License — Free to use, modify, and ship in commercial products.

MarkovGPU doesn’t just crunch numbers.

It discovers the hidden structure of reality.

Made with 🧡 by Sahil Rane

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

markovgpu_rane-0.4.1.tar.gz (58.1 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

markovgpu_rane-0.4.1-py3-none-any.whl (9.2 kB view details)

Uploaded Python 3

File details

Details for the file markovgpu_rane-0.4.1.tar.gz.

File metadata

  • Download URL: markovgpu_rane-0.4.1.tar.gz
  • Upload date:
  • Size: 58.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for markovgpu_rane-0.4.1.tar.gz
Algorithm Hash digest
SHA256 69f649b4bd7d2b8d7af99ec5f97042d523ff649c6610261fe3c58f69ecc3062a
MD5 e0cf1f3452d1ec3802f39e0e3c684a0b
BLAKE2b-256 8be3150e4401eacd793b006dd816508f0c771625eff3afa52bae36a908e8b842

See more details on using hashes here.

Provenance

The following attestation bundles were made for markovgpu_rane-0.4.1.tar.gz:

Publisher: publish.yml on wizardwithcodehazard/markov

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file markovgpu_rane-0.4.1-py3-none-any.whl.

File metadata

  • Download URL: markovgpu_rane-0.4.1-py3-none-any.whl
  • Upload date:
  • Size: 9.2 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for markovgpu_rane-0.4.1-py3-none-any.whl
Algorithm Hash digest
SHA256 8902582b0cbf27f032fc984425151d85c96663c22caedd854d2696d60f2e5f42
MD5 2249a3b1e8d11846671bc572eeed4616
BLAKE2b-256 18dd46a3c2ad6925cd6a86ea3aef6aa06448fafb8e38f8e01d8435584fec8e78

See more details on using hashes here.

Provenance

The following attestation bundles were made for markovgpu_rane-0.4.1-py3-none-any.whl:

Publisher: publish.yml on wizardwithcodehazard/markov

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page