Skip to main content

Python client for alphainfo.io — Structural Regime Detection API

Project description

alphainfo

Python client for the alphainfo Structural Intelligence API.

Detect structural regime changes in time series — biomedical signals, financial markets, energy grids, seismic data, IoT sensors, and more. No model training required.

from alphainfo import AlphaInfo

client = AlphaInfo(api_key="ai_your_key")
result = client.analyze(signal=ecg_data, sampling_rate=360.0, domain="biomedical")

print(result.confidence_band)   # 'stable', 'transition', or 'unstable'
print(result.structural_score)  # 0.0 to 1.0
print(result.analysis_id)       # UUID for audit trail

Installation

pip install alphainfo

Requires Python 3.8+. Only dependency: httpx.

Quick Start

1. Get your API key

Sign up at alphainfo.io/register — free tier includes 50 analyses/month.

2. Analyze a signal

from alphainfo import AlphaInfo

client = AlphaInfo(api_key="ai_your_key")

# Any time series: ECG, market prices, sensor readings, power grid...
result = client.analyze(
    signal=[1.2, 1.3, 1.1, 2.8, 3.1, 3.0, ...],
    sampling_rate=250.0,
    domain="biomedical",
)

if result.change_detected:
    print(f"Regime change detected! Band: {result.confidence_band}")
    print(f"Structural score: {result.structural_score:.3f}")
    print(f"Audit ID: {result.analysis_id}")

3. Analyze market data

# The API fetches market data automatically
market = client.analyze_market("AAPL", interval="1d")
print(f"AAPL regime: {market.confidence_band}")
print(f"Score: {market.structural_score:.3f}")

4. Batch analysis

# Analyze up to 100 signals in one call
batch = client.analyze_batch(
    signals=[signal_1, signal_2, signal_3],
    sampling_rate=1000.0,
    domain="sensors",
)

for item in batch.results:
    if item.success:
        print(f"Signal {item.index}: {item.confidence_band} ({item.structural_score:.3f})")
    else:
        print(f"Signal {item.index}: error — {item.error}")

5. Multi-channel (vector) analysis

# Multi-lead ECG, multi-axis accelerometer, etc.
vector = client.analyze_vector(
    channels={
        "lead_I": ecg_lead_1,
        "lead_II": ecg_lead_2,
        "lead_III": ecg_lead_3,
    },
    sampling_rate=360.0,
    domain="biomedical",
)

print(f"Aggregated score: {vector.structural_score:.3f}")
for name, ch in vector.channels.items():
    print(f"  {name}: {ch.confidence_band}")

6. Audit trail

# Replay any past analysis
replay = client.audit_replay("550e8400-e29b-41d4-a716-446655440000")
print(f"Original score: {replay.output['structural_score']}")

# List recent analyses
history = client.audit_list(limit=10)
for entry in history:
    print(f"{entry.analysis_id}{entry.structural_score}")

Async Support

from alphainfo import AsyncAlphaInfo

async with AsyncAlphaInfo(api_key="ai_your_key") as client:
    result = await client.analyze(signal=data, sampling_rate=250.0)
    market = await client.analyze_market("BTC-USD")

All methods available on AlphaInfo are also available on AsyncAlphaInfo.

Error Handling

from alphainfo import AlphaInfo, AuthError, RateLimitError, ValidationError

client = AlphaInfo(api_key="ai_your_key")

try:
    result = client.analyze(signal=data, sampling_rate=250.0)
except AuthError:
    print("Invalid API key")
except RateLimitError as e:
    print(f"Rate limited. Retry after {e.retry_after}s")
except ValidationError as e:
    print(f"Invalid input: {e.message}")

Exception hierarchy:

Exception HTTP Code When
AuthError 401 Invalid or missing API key
ValidationError 400, 413 Bad input or signal too large
RateLimitError 429 Quota or concurrency limit exceeded
NotFoundError 404 Analysis ID not found (audit)
APIError 5xx Server error
TimeoutError Request timed out after retries
NetworkError Connection failed

All inherit from AlphaInfoError.

Configuration

client = AlphaInfo(
    api_key="ai_your_key",
    base_url="https://alphainfo.io",  # default
    timeout=30.0,                      # seconds (default)
    max_retries=3,                     # automatic retry on transient errors
)

The client automatically retries on:

  • Network timeouts and connection errors
  • HTTP 429 (rate limits) — respects Retry-After header
  • HTTP 5xx (server errors)

Non-retryable errors (401, 400, 404) are raised immediately.

Rate Limit Info

result = client.analyze(signal=data, sampling_rate=250.0)
info = client.rate_limit_info
if info:
    print(f"Remaining: {info.remaining}/{info.limit}")

Domains

Domain Use case
generic Default — works for any signal
biomedical ECG, EEG, EMG, SpO2
finance Market prices, returns, volume
energy Power grid frequency, load
seismic Earthquake, vibration sensors
sensors IoT, industrial sensors
mlops Model drift, data quality
security Network traffic, intrusion
industrial Machinery, SCADA

Links

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

alphainfo-1.0.1.tar.gz (21.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

alphainfo-1.0.1-py3-none-any.whl (14.3 kB view details)

Uploaded Python 3

File details

Details for the file alphainfo-1.0.1.tar.gz.

File metadata

  • Download URL: alphainfo-1.0.1.tar.gz
  • Upload date:
  • Size: 21.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.9.6

File hashes

Hashes for alphainfo-1.0.1.tar.gz
Algorithm Hash digest
SHA256 0b500d54c00bbc929b500319a3db2e886270deb499c8193fb6c441b1ad87af31
MD5 b0dac7c18a6ca68fd5767027001abcff
BLAKE2b-256 18ef99ca24232a03fd82c7ef626561da37a0d2b290ad35f8c27ae3ebb31bd2fc

See more details on using hashes here.

File details

Details for the file alphainfo-1.0.1-py3-none-any.whl.

File metadata

  • Download URL: alphainfo-1.0.1-py3-none-any.whl
  • Upload date:
  • Size: 14.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.9.6

File hashes

Hashes for alphainfo-1.0.1-py3-none-any.whl
Algorithm Hash digest
SHA256 e26f20ad81124eae8181cf1e61688d20a1fb84cf454e405679035355bf64f015
MD5 0a8310c533f33fa6c81e77c21a0b76e5
BLAKE2b-256 f4a16c53dc07a42d5bd344a86c78b0a4c4ca340c25b9cd878ca20f14e391adf5

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page