Skip to main content

GMICloud media provider adapters for genblaze (video, image, audio)

Project description

genblaze-gmicloud

GMICloud multi-provider video / image / audio adapters for genblaze — access Seedance, Kling, Veo, Sora, Wan, Seedream, FLUX, Gemini image, ElevenLabs, MiniMax and more through one API with SHA-256 provenance manifests.

genblaze-gmicloud wraps GMICloud's request-queue API, giving you one-call access to a large catalog of video, image, and audio models — including Kling, Veo, Sora, Wan, Seedream, FLUX-Kontext-Pro, Gemini-2.5-Flash-Image, ElevenLabs TTS, MiniMax TTS, and MiniMax Music — via three genblaze provider classes. Compose into multi-step AI pipelines, persist outputs to Backblaze B2 or any S3-compatible store, and emit a tamper-evident provenance manifest on every run.

Why genblaze-gmicloud

  • One API, dozens of models — text-to-video (Seedance, Kling, Veo, Sora, Wan), text-to-image (Seedream, FLUX, Gemini, Reve), audio (ElevenLabs, MiniMax TTS/Music).
  • LLM access too — standalone chat() wrapper for Llama, DeepSeek, Qwen over GMICloud's OpenAI-compatible inference endpoint (see below).
  • Provenance by default — SHA-256-verified manifest with provider, model, prompt, params, cost.
  • Cost trackingstep.cost_usd is populated from GMICloud's response.
  • Production-ready — retries, timeouts, progress streaming, step caching.
  • Durable storage — plug genblaze-s3 in for Backblaze B2 / AWS S3 / R2 / MinIO persistence.

Providers + models

Provider class Modality Example models
GMICloudVideoProvider video kling-text2video-v1.6-pro, kling-image2video-v2.1-master, veo3, wan2.6-t2v, seedance-1-0-pro-250528, sora-2-pro
GMICloudImageProvider image seedream-5.0-lite, gemini-2.5-flash-image, reve-edit-fast-20251030, flux-kontext-pro
GMICloudAudioProvider audio ElevenLabs-TTS-v3, MiniMax-TTS-Speech-2.6-Turbo, MiniMax-Music-2.5

Registered via entry points as gmicloud, gmicloud-image, and gmicloud-audio. Any model on GMICloud's queue is supported — pass the exact model slug.

Slug casing — GMICloud's request queue is case-sensitive. Model ids are the lowercase slugs shown above. Pre-0.3 PascalCase ids (e.g. Seedream-5.0-Lite, Veo3, Wan-2.6-I2V) still resolve via ModelSpec.deprecated_aliases but emit a DeprecationWarning and will be removed in 0.4 — migrate early.

Install

pip install genblaze-gmicloud

Quickstart — video (Kling)

pip install genblaze-core genblaze-gmicloud
export GMI_API_KEY="..."
from genblaze_core import Modality, Pipeline
from genblaze_gmicloud import GMICloudVideoProvider

run, manifest = (
    Pipeline("gmicloud-video-demo")
    .step(GMICloudVideoProvider(), model="kling-text2video-v1.6-pro",
          prompt="A drone shot flying over a misty mountain valley at sunrise, cinematic",
          modality=Modality.VIDEO, duration=10, aspect_ratio="16:9")
    .run(timeout=600)
)
print(run.steps[0].assets[0].url, f"${run.steps[0].cost_usd:.3f}")

Quickstart — image (Seedream)

from genblaze_gmicloud import GMICloudImageProvider

run, manifest = (
    Pipeline("gmicloud-image-demo")
    .step(GMICloudImageProvider(), model="seedream-5.0-lite",
          prompt="A photorealistic macro shot of morning dew on a spider web, soft bokeh",
          modality=Modality.IMAGE, aspect_ratio="16:9")
    .run(timeout=120)
)

Quickstart — audio (ElevenLabs via GMICloud)

from genblaze_gmicloud import GMICloudAudioProvider

run, manifest = (
    Pipeline("gmicloud-audio-demo")
    .step(GMICloudAudioProvider(), model="ElevenLabs-TTS-v3",
          prompt="Welcome to Genblaze — the fastest way to build generative AI pipelines.",
          modality=Modality.AUDIO)
    .run(timeout=120)
)

Persist to Backblaze B2

from genblaze_core import KeyStrategy, ObjectStorageSink
from genblaze_s3 import S3StorageBackend

storage = ObjectStorageSink(
    S3StorageBackend.for_backblaze("my-bucket"),
    key_strategy=KeyStrategy.HIERARCHICAL,
)
# pass sink=storage to .run(…)

Backblaze B2 is the recommended default sink — cost-efficient, S3-compatible, Object Lock for immutable manifests.

LLM access — standalone chat()

For callers driving a media pipeline from an LLM — caption expansion, prompt rewriting, scene description — genblaze-gmicloud ships a chat() callable over GMICloud's OpenAI-compatible inference endpoint. It sits outside the Pipeline / Step machinery (text generation doesn't benefit from the polling / manifest / asset machinery built for media).

from genblaze_gmicloud import chat

resp = chat("deepseek-ai/DeepSeek-V3", prompt="A cinematic sunset over Tokyo")
print(resp.text, resp.tokens_out)

Full signature and ChatResponse shape: docs/features/llm-calls.md.

Credentials

Only API-key auth is supported. Set GMI_API_KEY (obtain from https://console.gmicloud.ai/) or pass api_key= to any provider ctor or to chat().

Configuring the endpoint (staging, proxies, VPC)

All three provider classes and chat() accept a base_url= ctor kwarg (or GMI_BASE_URL env var) to override the default endpoint, and an http_client= kwarg for injecting a pre-built httpx.Client — useful for shared connection pools across multi-modality pipelines or for mocking in tests.

import httpx
from genblaze_gmicloud import GMICloudVideoProvider, GMICloudImageProvider

shared = httpx.Client(
    base_url="https://my-vpc-proxy.example/gmi",
    headers={"Authorization": f"Bearer {key}"},
    timeout=120,
)
video = GMICloudVideoProvider(http_client=shared)
image = GMICloudImageProvider(http_client=shared)
# Caller owns `shared` — providers never close externally-supplied clients.

Naming reference

GMICloud surfaces five related names; they look interchangeable but come from different namespaces:

Surface Value
PyPI package genblaze-gmicloud
Python import import genblaze_gmicloud
Provider class prefix GMICloud* (e.g. GMICloudVideoProvider)
Entry-point slug gmicloud, gmicloud-image, gmicloud-audio
Env vars GMI_API_KEY, GMI_BASE_URL

The GMI_ env prefix is short on purpose; the class / import / PyPI names use the full gmicloud for precision and to leave room for future genblaze-gmi* packages if needed.

Reading outputs safely

step.assets[0] is only valid when the step succeeded. Always check step.status first — especially in fan-out runs where one step may fail and others succeed:

for step in run.steps:
    if step.status == "succeeded" and step.assets:
        print(step.assets[0].url)
    elif step.status == "failed":
        print(f"failed ({step.error_code}): {step.error}")

Documentation

Related packages

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

genblaze_gmicloud-0.2.3.tar.gz (22.1 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

genblaze_gmicloud-0.2.3-py3-none-any.whl (21.1 kB view details)

Uploaded Python 3

File details

Details for the file genblaze_gmicloud-0.2.3.tar.gz.

File metadata

  • Download URL: genblaze_gmicloud-0.2.3.tar.gz
  • Upload date:
  • Size: 22.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.5

File hashes

Hashes for genblaze_gmicloud-0.2.3.tar.gz
Algorithm Hash digest
SHA256 61ceedaa728fab47efdb8c83cc2b90019e54434a08a5dca695abaa0fb2d69b53
MD5 691c5a187d8a41a4c31f8c015661b188
BLAKE2b-256 fff0d8bb8b0dea5eca99291eafe673681a77a171c7a16501d87dcafc34366669

See more details on using hashes here.

File details

Details for the file genblaze_gmicloud-0.2.3-py3-none-any.whl.

File metadata

File hashes

Hashes for genblaze_gmicloud-0.2.3-py3-none-any.whl
Algorithm Hash digest
SHA256 6d1ca01dc9d88154b2cfe730ce191c4b1fc043bc0ad6c334f916d0c729309d19
MD5 6619315d8d09166280cb25b452672717
BLAKE2b-256 dc0621fc00e00d6218266b9ee3ae4ecedb5327905fa0bc9a9a278273c17b4a73

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page