Skip to main content

High-performance Markov Chains & HMMs using OpenCL

Project description

MarkovGPU

Massive Scale Markov Models on Consumer Hardware

image

Run million-state HMMs on your laptop GPU.
No CUDA required • Hybrid CPU/GPU Backend • Production Ready

PyPI version Python 3.9+ License: MIT Build Status


🌟 The Engine for Stochastic Intelligence

MarkovGPU is a high-performance probabilistic modeling library built for speed. It breaks the "NVIDIA Monopoly" by using OpenCL to accelerate Hidden Markov Models (HMM) and Markov Chains on any GPU—including AMD Radeon, Intel Arc, and Apple Silicon.

It doesn't just run; it thinks. The Smart Hybrid Backend automatically routes small tasks to the CPU (NumPy) and massive workloads to the GPU, giving you optimal performance at every scale.


🚀 Core Superpowers

Feature Magic Behind It
Hardware Agnostic Built on OpenCL — runs on AMD, Intel, NVIDIA, and Apple M1/M2/M3 chips.
🧠 Smart Hybrid Backend Auto-detects problem size ($N$). Uses NumPy for speed on small data, GPU for massive throughput.
📉 Log-Space Stability Implements Log-Sum-Exp kernels to prevent underflow on long time-series (1M+ steps).
🕵️ Viterbi Decoding Finds the "Hidden Truth" in noisy data (e.g., market regimes, DNA sequences) in milliseconds.
🎓 Unsupervised Learning Baum-Welch (EM) algorithm trains models directly on the GPU, learning rules from raw data.
📦 Zero-Config Install pip install markovgpu-rane. No driver hell. No CUDA toolkit nightmares.

🏗️ Architecture: The Hybrid Pipeline

graph LR
    A[User Code] -->|Request Fit/Predict| B{Smart Dispatcher}
    B -->|Small N < 64| C["CPU Engine
    (NumPy AVX2)"]
    B -->|Large N >= 64| D["GPU Engine
    (OpenCL Kernels)"]
    C --> E[Result]
    D --> E
    subgraph GPU_Acceleration[GPU Acceleration]
    D --> F[Matrix Multiply]
    D --> G[Log-Sum-Exp]
    D --> H[Parallel Viterbi]
    end

The library handles the hardware. You handle the math.

⚡ Performance: Benchmarks

Task: Viterbi Decoding (64 Hidden States, 5000 Days of Data).
Hardware: AMD Ryzen 680M (Integrated Graphics).

Engine Execution Time Speedup
🐢 CPU (NumPy Optimized) 5.06s 1x
🚀 GPU (MarkovGPU) 0.82s 6.2x

⚙️ Quick Start in 30 Seconds

Installation

# Production
pip install markovgpu-rane

# Or for local development
uv pip install markovgpu-rane

1. Market Regime Detection (Viterbi)

Identify hidden "Bull" vs. "Bear" markets from noisy stock returns.

import numpy as np
from markovgpu import MarkovEngine

# 1. Setup the Rules (Transition Matrix)
# "Bull markets tend to stay Bullish (95%)"
trans_mat = np.array([[0.95, 0.05], 
                      [0.10, 0.90]], dtype=np.float32)

# 2. Feed the Data (Observation Likelihoods)
# Shape: (1000 Days, 2 States)
obs_probs = np.random.rand(1000, 2).astype(np.float32) 

# 3. Ignite the Engine
engine = MarkovEngine()
predicted_states = engine.decode_regime(trans_mat, obs_probs)

print("Detected Regimes:", predicted_states) 
# Output: [0, 0, 0, 1, 1, 1, 0 ...]

2. Unsupervised Learning (Baum-Welch)

Train the AI to discover the hidden rules from raw data.

# The engine learns the Transition Matrix automatically
learned_matrix = engine.fit(
    obs_probs, 
    n_states=2, 
    n_iters=100, 
    tolerance=1e-4
)

print("Discovered Rules:")
print(learned_matrix)

🔬 Technical Brilliance

1. The Log-Sum-Exp Kernel

Standard HMMs crash on long sequences because probabilities like $0.9^{1000}$ vanish to zero. We solved this by rewriting the entire GPU kernel in Log-Space:

// Actual OpenCL Kernel snippet
float log_add(float log_a, float log_b) {
    float max_val = max(log_a, log_b);
    return max_val + log1p(exp(min(log_a, log_b) - max_val));
}

Result: You can process sequences of infinite length without numerical collapse.

2. Parallel Viterbi

Instead of a slow Python loop, we launch $N$ threads (one per state) for every time step on the GPU, calculating the optimal path in parallel.


🛠️ Project Structure

markovgpu/
├── src/markovgpu/
│   ├── backend.py       # The Brain (Smart Dispatcher)
│   ├── kernels.cl       # The Muscle (OpenCL C Code)
│   └── __init__.py
├── tests/               # Unit Tests
├── pyproject.toml       # Modern Packaging Config
└── README.md

🌱 Contributing

We welcome forks, issues, and PRs!

git clone https://github.com/wizardwithcodehazard/markov.git
cd markov
uv sync --dev
uv run pytest

📄 License

MIT License — Free to use, modify, and ship in commercial products.

MarkovGPU doesn’t just crunch numbers.

It discovers the hidden structure of reality.

Made with 🧡 by Sahil Rane

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

markovgpu_rane-0.4.2.tar.gz (88.6 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

markovgpu_rane-0.4.2-py3-none-any.whl (9.0 kB view details)

Uploaded Python 3

File details

Details for the file markovgpu_rane-0.4.2.tar.gz.

File metadata

  • Download URL: markovgpu_rane-0.4.2.tar.gz
  • Upload date:
  • Size: 88.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for markovgpu_rane-0.4.2.tar.gz
Algorithm Hash digest
SHA256 62f5054a7d3ab830da46173a23d7cd3e34d14f6dd7d487a39dba3a9c5bfc0596
MD5 7a81a8c4732fc18729f0e58a908bcb48
BLAKE2b-256 6773248b9f420e56cb0d9ce2c488b7302d8e6ed464b4a4fa7615eebbf4e540e8

See more details on using hashes here.

Provenance

The following attestation bundles were made for markovgpu_rane-0.4.2.tar.gz:

Publisher: publish.yml on wizardwithcodehazard/markov

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file markovgpu_rane-0.4.2-py3-none-any.whl.

File metadata

  • Download URL: markovgpu_rane-0.4.2-py3-none-any.whl
  • Upload date:
  • Size: 9.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for markovgpu_rane-0.4.2-py3-none-any.whl
Algorithm Hash digest
SHA256 ac3ff24998bd62bd17542ba4732fefdf1b3fbd0e5865c3291e9b4dda66d8da7f
MD5 7de77b9d829d4d80e6219b9b5cc023b3
BLAKE2b-256 df1ea1815dd992d82dbd35c8f25968e73e07ba4efe0befe1feb10aedad910a80

See more details on using hashes here.

Provenance

The following attestation bundles were made for markovgpu_rane-0.4.2-py3-none-any.whl:

Publisher: publish.yml on wizardwithcodehazard/markov

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page