Skip to main content

A lightweight GPU runtime for Python with NVRTC JIT compilation and NumPy-like API

Project description

PyGPUkit — Lightweight GPU Runtime for Python

A minimal, modular GPU runtime with NVRTC JIT compilation, GPU scheduling, and a clean NumPy-like API.


🚀 Overview

PyGPUkit is a lightweight GPU runtime for Python that provides:

  • NVRTC-based JIT kernel compilation
  • A NumPy-like GPUArray type
  • Kubernetes-inspired GPU scheduler (bandwidth + memory guarantees)
  • Extensible operator set (add/mul/matmul, custom kernels)
  • Minimal dependencies and embeddable runtime

PyGPUkit aims to be the “micro-runtime for GPU computing”: small, fast, and ideal for research, inference tooling, DSP, and real-time systems.


✨ Features

  • Lightweight — no PyTorch/CuPy overhead
  • 🧩 Modular — runtime / memory / scheduler / JIT / ops
  • 📦 GPUArray with NumPy interop
  • 🛠 NVRTC JIT for CUDA kernels
  • 🎼 Advanced Scheduler with memory & bandwidth guarantees
  • 🔌 Optional Triton backend (planned)
  • 🧪 Test-friendly runtime

🔧 Installation

(Available after first PyPI release)

pip install pygpukit

From source:

git clone https://github.com/m96-chan/PyGPUkit
cd PyGPUkit
pip install -e .

Requirements:

  • Python 3.9+
  • CUDA 11+
  • NVRTC available
  • NVIDIA GPU

🧭 Project Goals

  1. Provide the smallest usable GPU runtime for Python
  2. Expose GPU scheduling (bandwidth, memory, partitioning)
  3. Make writing custom GPU kernels easy
  4. Serve as a building block for inference engines, DSP systems, and real-time workloads

📚 Usage Examples

Allocate Arrays

import pygpukit as gp

x = gp.zeros((1024, 1024), dtype="float32")
y = gp.ones((1024, 1024), dtype="float32")

Basic Operations

z = gp.add(x, y)
w = gp.matmul(x, y)

CPU ↔ GPU Transfer

arr = z.to_numpy()
garr = gp.from_numpy(arr)

Custom NVRTC Kernel

extern "C" __global__
void scale(float* x, float factor, int n) {
    int idx = blockIdx.x * blockDim.x + threadIdx.x;
    if (idx < n) x[idx] *= factor;
}
kernel = gp.jit(src, func="scale")
kernel(x, factor=0.5, n=x.size)

🎼 Scheduler — Kubernetes‑Inspired GPU Orchestration

PyGPUkit includes an experimental scheduler that treats a single GPU as a multi-tenant compute node, similar to how Kubernetes orchestrates CPU workloads. The goal is to provide resource isolation, guarantees, and fair sharing across multiple GPU tasks.

Core Capabilities


1. GPU Memory Reservation

Tasks may request a guaranteed block of GPU memory.

  • Hard guarantees → task is rejected if memory cannot be allocated
  • Soft guarantees → best‑effort allocation
  • Overcommit strategies (evict to host when pressure is high)
  • Reclaim policies (LRU GPUArray eviction)

Example:

task = scheduler.submit(
    fn,
    memory="512MB",
)

2. GPU Bandwidth Guarantees / Throttling

Tasks may request a specific percentage of GPU compute bandwidth.

Bandwidth control is implemented via:

  • Stream priority
  • Kernel pacing (launch intervals)
  • Micro‑slicing large kernels
  • Cooperative time‑quantized scheduling
  • Persistent dispatcher kernels (planned)

Example:

task = scheduler.submit(
    fn,
    bandwidth=0.20,   # 20% GPU compute share
)

3. Logical GPU Partitioning

PyGPUkit implements software‑defined GPU slicing, similar in spirit to Kubernetes device plugin resource partitioning.

Slices may define:

  • Memory quota
  • Bandwidth share
  • Stream priority band
  • Isolation level

Useful for:

  • Multi‑tenant inference servers
  • Real‑time audio/DSP workloads
  • Background/foreground GPU task separation

4. Scheduling Policies

The scheduler supports multiple policies:

  • Guaranteed — exclusive reservation, strict QoS
  • Burstable — partial guarantees, opportunistic bandwidth
  • BestEffort — uses leftover GPU cycles
  • Priority scheduling
  • Deadline scheduling (planned)
  • Weighted fair sharing

Example:

task = scheduler.submit(
    fn,
    policy="guaranteed",
    memory="1GB",
    bandwidth=0.10,
)

5. Admission Control

Before executing a task, the scheduler performs:

  • Resource validation
  • Quota check
  • QoS matching
  • Scheduling feasibility

Results in:

  • admitted
  • queued
  • rejected

6. Monitoring & Introspection

PyGPUkit exposes live metrics:

  • Memory usage per task
  • SM occupancy and GPU utilization
  • Throttling / pacing logs
  • Queue position / execution state
  • Reclaim/eviction count

Example:

stats = scheduler.stats(task_id)

7. Soft Isolation Model

While not OS‑level isolation, each GPU task is provided:

  • Dedicated stream groups
  • Guaranteed memory pools
  • Kernel pacing to enforce bandwidth
  • Optional sandboxed GPUArray region

This provides practical multi‑tenant safety without MIG/MPS.


🏗 Proposed Directory Structure

PyGPUkit/
  core/         # NVRTC wrapper, device info
  memory/       # GPUArray, allocators
  scheduler/    # orchestration, partitioning, throttling
  ops/          # built-in kernels
  jit/          # JIT compiler + cache
  python/       # high-level Python API
  examples/
  tests/

🧪 Roadmap

v0.1 (MVP)

  • GPUArray
  • NVRTC JIT
  • add/mul/matmul ops
  • Basic stream manager
  • Packaging + wheels

v0.2

  • Scheduler (memory + bandwidth guarantees)
  • Kernel cache
  • NumPy interop
  • Benchmarks

v0.3

  • Triton optional backend
  • Advanced ops (softmax, layernorm)
  • Inference‑oriented plugin system

🤝 Contributing

Contributions and discussions are welcome!
Please open Issues for feature requests, bugs, or design proposals.


📄 License

MIT License


⭐ Acknowledgements

Inspired by:

  • CUDA Runtime
  • NVRTC
  • PyCUDA
  • CuPy
  • Triton

PyGPUkit aims to fill the gap for a tiny, embeddable GPU runtime for Python.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

pygpukit-0.1.0.tar.gz (28.0 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

pygpukit-0.1.0-py3-none-any.whl (18.6 kB view details)

Uploaded Python 3

File details

Details for the file pygpukit-0.1.0.tar.gz.

File metadata

  • Download URL: pygpukit-0.1.0.tar.gz
  • Upload date:
  • Size: 28.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for pygpukit-0.1.0.tar.gz
Algorithm Hash digest
SHA256 d7f709a756718162c0525a318eeb51a13ff027a72f05b4b3f73e5bdb1a6edf22
MD5 ed48f54143f036da2b69cc6a765ed114
BLAKE2b-256 65843558b97d6168f8738a03f31e1e62ef4ffe7ef4e70bacec4094443806fbd7

See more details on using hashes here.

Provenance

The following attestation bundles were made for pygpukit-0.1.0.tar.gz:

Publisher: release.yml on m96-chan/PyGPUkit

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file pygpukit-0.1.0-py3-none-any.whl.

File metadata

  • Download URL: pygpukit-0.1.0-py3-none-any.whl
  • Upload date:
  • Size: 18.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for pygpukit-0.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 9fafa738e2593c3f9dade947ed4d2318bd37f78d074f5151ccd61a1ccf75160c
MD5 649221afafef827f1316b93802025070
BLAKE2b-256 f752ba7d1de39b74fa8e759a47a73771072ef9ef3de05e33b3c55e5b17528a0f

See more details on using hashes here.

Provenance

The following attestation bundles were made for pygpukit-0.1.0-py3-none-any.whl:

Publisher: release.yml on m96-chan/PyGPUkit

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page