Skip to main content

Official Python SDK for WhiteBoxAI - ML Monitoring and Observability

Project description

WhiteBoxAI Python SDK

Official Python SDK for integrating WhiteBoxAI monitoring into your ML applications.

Features

  • ๐Ÿš€ Easy Integration - Monitor models with just a few lines of code
  • ๐Ÿ“Š Framework Support - Native integrations for Scikit-learn, PyTorch, TensorFlow, XGBoost, and more
  • ๐ŸŽฏ Decorator-based Monitoring - Zero-code-change monitoring with decorators
  • โšก Async/Sync Interfaces - Support for both synchronous and asynchronous workflows
  • ๐Ÿ”’ Privacy-First - Built-in PII detection and data masking
  • ๐Ÿ’พ Local Caching - TTL-based caching to reduce API calls
  • ๐Ÿ“ˆ Drift Detection - Automatic model and data drift monitoring
  • ๐ŸŽจ Flexible Configuration - Extensive configuration options and feature flags

Installation

pip install whiteboxai-sdk

# With specific framework support
pip install whiteboxai-sdk[sklearn]
pip install whiteboxai-sdk[pytorch]
pip install whiteboxai-sdk[all]  # All integrations

Quick Start

Basic Usage

from whiteboxai import WhiteBoxAI, ModelMonitor

# Initialize client
client = WhiteBoxAI(api_key="your-api-key")

# Create monitor
monitor = ModelMonitor(client)

# Register model
model_id = monitor.register_model(
    name="fraud_detection",
    model_type="classification",
    framework="sklearn"
)

# Log predictions
monitor.log_prediction(
    inputs={"amount": 100.0, "merchant": "store_123"},
    output={"fraud_probability": 0.15, "prediction": "legitimate"}
)

Scikit-learn Integration

from sklearn.ensemble import RandomForestClassifier
from whiteboxai import WhiteBoxAI
from whiteboxai.integrations.sklearn import SklearnMonitor

# Train model
model = RandomForestClassifier()
model.fit(X_train, y_train)

# Setup monitoring
client = WhiteBoxAI(api_key="your-api-key")
monitor = SklearnMonitor(client, model=model)
monitor.register_from_model(model_type="classification")

# Wrap model for automatic monitoring
monitored_model = monitor.wrap_model(model)

# Predictions are automatically logged
predictions = monitored_model.predict(X_test)

PyTorch Integration

import torch
import torch.nn as nn
from whiteboxai import WhiteBoxAI
from whiteboxai.integrations.pytorch import TorchMonitor

# Define model
model = nn.Sequential(
    nn.Linear(10, 64),
    nn.ReLU(),
    nn.Linear(64, 2)
)

# Setup monitoring
client = WhiteBoxAI(api_key="your-api-key")
monitor = TorchMonitor(client, model=model)
monitor.register_from_model(model_type="classification")

# Wrap model
monitored_model = monitor.wrap_model(model)

# Predictions are automatically logged
with torch.no_grad():
    outputs = monitored_model(inputs)

TensorFlow/Keras Integration

from tensorflow import keras
from whiteboxai import WhiteBoxAI
from whiteboxai.integrations.tensorflow import KerasMonitor, WhiteBoxAICallback

# Build model
model = keras.Sequential([
    keras.layers.Dense(64, activation='relu', input_shape=(20,)),
    keras.layers.Dense(1)
])
model.compile(optimizer='adam', loss='mse')

# Setup monitoring
client = WhiteBoxAI(api_key="your-api-key")
monitor = KerasMonitor(client, model=model, model_name="keras_model")
monitor.register_from_model(model_type="regression")

# Train with monitoring callback
callback = WhiteBoxAICallback(monitor, log_frequency=1)
model.fit(X_train, y_train,
          validation_split=0.2,
          callbacks=[callback],
          epochs=50)

# Make predictions with automatic logging
predictions = monitor.predict(X_test, log=True)

Hugging Face Transformers Integration

from transformers import pipeline
from whiteboxai import WhiteBoxAI
from whiteboxai.integrations.transformers import TransformersMonitor, wrap_transformers_pipeline

# Load model
classifier = pipeline("sentiment-analysis")

# Setup monitoring
client = WhiteBoxAI(api_key="your-api-key")
monitor = TransformersMonitor(
    client=client,
    pipeline=classifier,
    model_name="sentiment_classifier"
)

# Register model
monitor.register_from_model(name="Sentiment Classifier", version="1.0.0")

# Make predictions with automatic logging
result = monitor.predict("I love this product!", log=True)

# Or wrap pipeline for auto-logging
wrapped = wrap_transformers_pipeline(classifier, monitor)
result = wrapped("Great service!")  # Automatically logged

LangChain Integration

from langchain.chains import LLMChain
from langchain.llms import OpenAI
from langchain.prompts import PromptTemplate
from whiteboxai import WhiteBoxAI
from whiteboxai.integrations.langchain import LangChainMonitor, wrap_langchain_chain

# Setup monitoring
client = WhiteBoxAI(api_key="your-api-key")
monitor = LangChainMonitor(
    client=client,
    application_name="qa_bot",
    track_tokens=True,
    track_cost=True
)

# Register application
monitor.register_application(name="Q&A Bot", version="1.0.0")

# Create chain
llm = OpenAI(temperature=0.7)
prompt = PromptTemplate(input_variables=["question"], template="Answer: {question}")
chain = LLMChain(llm=llm, prompt=prompt)

# Option 1: Use callback handler
callback = monitor.create_callback_handler()
result = chain.run(question="What is AI?", callbacks=[callback])

# Option 2: Wrap chain for auto-logging
wrapped_chain = wrap_langchain_chain(chain, monitor)
result = wrapped_chain.run(question="What is AI?")  # Automatically logged

XGBoost/LightGBM Monitoring

import xgboost as xgb
import lightgbm as lgb
from whiteboxai import WhiteBoxAI
from whiteboxai.integrations.boosting import XGBoostMonitor, LightGBMMonitor, wrap_xgboost_model

client = WhiteBoxAI(api_key="your-api-key")

# XGBoost monitoring
xgb_monitor = XGBoostMonitor(
    client=client,
    model_name="fraud_detector",
    track_feature_importance=True,
    importance_type="gain"  # or 'weight', 'cover', 'total_gain', 'total_cover'
)

# Train and register model
model = xgb.XGBClassifier(n_estimators=100, max_depth=5)
model.fit(X_train, y_train)
xgb_monitor.register_from_model(model, X_train, y_train)

# Make predictions with monitoring
predictions = xgb_monitor.predict(model, X_test, y_test)

# Or wrap model for automatic logging
wrapped_model = wrap_xgboost_model(model, xgb_monitor)
predictions = wrapped_model.predict(X_test)  # Auto-logged

# LightGBM monitoring
lgb_monitor = LightGBMMonitor(
    client=client,
    model_name="churn_predictor",
    track_feature_importance=True,
    importance_type="gain"  # or 'split'
)

model = lgb.LGBMClassifier(n_estimators=100)
model.fit(X_train, y_train)
lgb_monitor.register_from_model(model, X_train, y_train)
predictions = lgb_monitor.predict(model, X_test, y_test)

Decorator-based Monitoring

from whiteboxai import WhiteBoxAI, ModelMonitor, monitor_model

client = WhiteBoxAI(api_key="your-api-key")
monitor = ModelMonitor(client, model_id=123)

@monitor_model(monitor, input_keys=["features"], explain=True)
def predict(features):
    # Your prediction logic
    return model.predict(features)

# Predictions are automatically logged
result = predict(features=[1.0, 2.0, 3.0])

Async Support

import asyncio
from whiteboxai import WhiteBoxAI, ModelMonitor

async def main():
    async with WhiteBoxAI(api_key="your-api-key") as client:
        monitor = ModelMonitor(client)

        # Register model
        model_id = await monitor.aregister_model(
            name="async_model",
            model_type="classification"
        )

        # Log prediction
        await monitor.alog_prediction(
            inputs={"feature1": 1.0},
            output={"prediction": 0.85}
        )

asyncio.run(main())

Advanced Features

Offline Mode

Enable robust operation with unreliable network connectivity. Operations are queued locally and synced automatically.

from whiteboxai import WhiteBoxAI

# Enable offline mode with auto-sync
client = WhiteBoxAI(
    api_key="your-api-key",
    enable_offline=True,
    offline_dir="./whiteboxai_offline",
    offline_auto_sync=True,
    offline_sync_interval=60  # Sync every 60 seconds
)

# Operations are automatically queued when API is unavailable
# Check queue status
status = client.get_offline_status()
print(f"Queued operations: {status['queue_size']}")

# Manually trigger sync
result = client.sync_offline_queue()
print(f"Synced: {result['synced']}, Failed: {result['failed']}")

# Cleanup old operations
client.cleanup_offline_queue(older_than_days=7)

Key Features:

  • Persistent Queue: SQLite-based storage survives restarts
  • Auto-Sync: Background synchronization every 60s (configurable)
  • Priority-Based: CRITICAL > HIGH > NORMAL > LOW
  • Retry Logic: Automatic retry with exponential backoff (max 3 attempts)
  • Thread-Safe: Supports concurrent operations

Configuration:

client = WhiteBoxAI(
    api_key="your-api-key",
    enable_offline=True,
    offline_dir="./offline_queue",        # Storage directory
    offline_max_queue_size=10000,         # Max operations (0 = unlimited)
    offline_auto_sync=True,               # Enable auto-sync
    offline_sync_interval=60,             # Sync interval (seconds)
)

See Offline Mode Guide for complete documentation.

Privacy Filters

from whiteboxai import WhiteBoxAI
from whiteboxai.privacy import mask_data

client = WhiteBoxAI(
    api_key="your-api-key",
    enable_privacy_filters=True
)

# Data is automatically masked before sending
data = {
    "email": "user@example.com",
    "phone": "555-123-4567",
    "amount": 100.0
}

# Mask sensitive data
masked = mask_data(data)
# {"email": "***MASKED***", "phone": "***MASKED***", "amount": 100.0}

Local Caching

client = WhiteBoxAI(
    api_key="your-api-key",
    enable_caching=True,
    cache_ttl=3600,
    cache_max_size=1000
)

Sampling

# Monitor 10% of predictions
monitor = ModelMonitor(
    client,
    model_id=123,
    sampling_rate=0.1
)

Drift Detection

import numpy as np

# Set baseline data
baseline = np.random.randn(1000, 10)
monitor.set_baseline(baseline)

# Detect drift
current_data = np.random.randn(100, 10)
drift_report = monitor.detect_drift(current_data)

Configuration

The SDK can be configured via constructor parameters or environment variables:

from whiteboxai import WhiteBoxAI

client = WhiteBoxAI(
    api_key="your-api-key",              # or EXPLAINAI_API_KEY env var
    base_url="https://api.whiteboxai.io", # Custom API endpoint
    timeout=30,                           # Request timeout (seconds)
    max_retries=3,                        # Retry attempts

    # Offline mode
    enable_offline=True,                  # Enable offline queueing
    offline_dir="./whiteboxai_offline",  # Queue storage directory
    offline_max_queue_size=10000,        # Max queued operations
    offline_auto_sync=True,              # Auto-sync in background
    offline_sync_interval=60,            # Sync interval (seconds)

    # Other features
    enable_caching=True,                  # Enable local caching
    enable_privacy_filters=True,          # Enable PII masking
    enable_sampling=True,                 # Enable prediction sampling
    sampling_rate=1.0                     # Sample 100% of predictions
)

API Reference

WhiteBoxAI Client

Main client for API interaction.

Methods:

  • models - Models resource
  • predictions - Predictions resource
  • explanations - Explanations resource
  • drift - Drift detection resource
  • alerts - Alerts resource

ModelMonitor

Simplified monitoring interface.

Methods:

  • register_model() - Register a new model
  • log_prediction() - Log a single prediction
  • log_batch() - Log multiple predictions
  • set_baseline() - Set baseline data for drift detection
  • detect_drift() - Detect model drift

Decorators

  • @monitor_model - Monitor all predictions from a function
  • @monitor_prediction - Monitor individual predictions with custom extractors

Framework Integrations

  • whiteboxai.integrations.sklearn - Scikit-learn integration
  • whiteboxai.integrations.pytorch - PyTorch integration

Examples

See the examples/ directory for more examples:

  • basic_monitoring.py - Basic monitoring example
  • sklearn_integration.py - Scikit-learn integration
  • pytorch_integration.py - PyTorch integration
  • async_monitoring.py - Async API usage
  • batch_logging.py - Batch prediction logging
  • drift_detection.py - Drift detection example
  • offline_mode_example.py - Offline mode with queue management
  • boosting_example.py - XGBoost/LightGBM integration

Support

whiteboxai-python-sdk/
โ”œโ”€โ”€ src/
โ”‚   โ””โ”€โ”€ whiteboxai/
โ”‚       โ”œโ”€โ”€ __init__.py
โ”‚       โ”œโ”€โ”€ __version__.py
โ”‚       โ”œโ”€โ”€ client.py
โ”‚       โ”œโ”€โ”€ monitor.py
โ”‚       โ”œโ”€โ”€ decorators.py
โ”‚       โ”œโ”€โ”€ privacy.py
โ”‚       โ”œโ”€โ”€ offline.py
โ”‚       โ”œโ”€โ”€ integrations/
โ”‚       โ”‚   โ”œโ”€โ”€ sklearn.py
โ”‚       โ”‚   โ”œโ”€โ”€ pytorch.py
โ”‚       โ”‚   โ”œโ”€โ”€ tensorflow.py
โ”‚       โ”‚   โ”œโ”€โ”€ transformers.py
โ”‚       โ”‚   โ”œโ”€โ”€ langchain.py
โ”‚       โ”‚   โ””โ”€โ”€ boosting.py
โ”‚       โ””โ”€โ”€ models/          # Pydantic models
โ”œโ”€โ”€ tests/
โ”‚   โ”œโ”€โ”€ unit/
โ”‚   โ”œโ”€โ”€ integration/
โ”‚   โ””โ”€โ”€ e2e/
โ”œโ”€โ”€ examples/
โ”‚   โ”œโ”€โ”€ basic_monitoring.py
โ”‚   โ”œโ”€โ”€ sklearn_integration.py
โ”‚   โ”œโ”€โ”€ pytorch_integration.py
โ”‚   โ”œโ”€โ”€ offline_mode_example.py
โ”‚   โ””โ”€โ”€ ...
โ”œโ”€โ”€ docs/
โ”‚   โ”œโ”€โ”€ getting-started.md
โ”‚   โ”œโ”€โ”€ integrations.md
โ”‚   โ”œโ”€โ”€ offline-mode.md
โ”‚   โ””โ”€โ”€ api-reference.md
โ”œโ”€โ”€ pyproject.toml
โ”œโ”€โ”€ setup.py
โ”œโ”€โ”€ README.md
โ”œโ”€โ”€ CHANGELOG.md
โ”œโ”€โ”€ LICENSE
โ””โ”€โ”€ .github/
    โ””โ”€โ”€ workflows/
        โ”œโ”€โ”€ test.yml
        โ”œโ”€โ”€ publish.yml
        โ””โ”€โ”€ docs.yml

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

whiteboxai_sdk-0.1.0.tar.gz (93.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

whiteboxai_sdk-0.1.0-py3-none-any.whl (47.0 kB view details)

Uploaded Python 3

File details

Details for the file whiteboxai_sdk-0.1.0.tar.gz.

File metadata

  • Download URL: whiteboxai_sdk-0.1.0.tar.gz
  • Upload date:
  • Size: 93.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for whiteboxai_sdk-0.1.0.tar.gz
Algorithm Hash digest
SHA256 e3a80019bfdcb3eb7b2a6da2b1e19f192dda39b22b7ed5831c3ef7de15a17f10
MD5 8bd5165cc9ff964dac58c6b3d00ea218
BLAKE2b-256 7d8ddd8db2d7291acdc1a0e6846aef7afc598acab8e44bbc0a4028b3ef23301a

See more details on using hashes here.

Provenance

The following attestation bundles were made for whiteboxai_sdk-0.1.0.tar.gz:

Publisher: publish.yml on AgentaFlow/whitebox-python-sdk

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file whiteboxai_sdk-0.1.0-py3-none-any.whl.

File metadata

  • Download URL: whiteboxai_sdk-0.1.0-py3-none-any.whl
  • Upload date:
  • Size: 47.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for whiteboxai_sdk-0.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 fe0078fc6d06adffa23ebda20c0c2e6beade18893955407775f63cab15322431
MD5 5234d17dacdabb32c8d48585dbe44f6c
BLAKE2b-256 398d62c0ae4a4bb080e28d52f60d6c0da47492e1dd6c35af143fc1211ae89f35

See more details on using hashes here.

Provenance

The following attestation bundles were made for whiteboxai_sdk-0.1.0-py3-none-any.whl:

Publisher: publish.yml on AgentaFlow/whitebox-python-sdk

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page