The official Embedl Hub Python client library.
Project description
Embedl Hub Python library
Optimize and deploy your model on any edge device with the Embedl Hub Python library:
- Compile your model for execution on CPU, GPU, NPU or other AI accelerators using ONNX Runtime, TensorRT, or TFLite backends.
- Profile your model's latency and memory usage on real edge devices in the cloud.
- Invoke your compiled model to run inference with real input data.
The library logs your metrics, parameters, and results on the Embedl Hub website, allowing you to inspect, compare, and reproduce your results.
For comprehensive getting started guides and API reference, visit the Embedl Hub documentation.
Create a free Embedl Hub account to get started.
Installation
Install embedl-hub with pip:
pip install embedl-hub
Usage
The embedl-hub library can be used in two ways:
CLI
The embedl-hub (or ehub) command provides an end-to-end workflow for
compiling, profiling, and invoking models from the terminal:
Usage: embedl-hub [OPTIONS] COMMAND [ARGS]...
embedl-hub end-to-end Edge-AI workflow CLI
╭─ Options ────────────────────────────────────────────────────────────────╮
│ --version -V Print embedl-hub version and exit. │
│ --verbose -v INTEGER Increase verbosity (-v, -vv, -vvv). │
│ --help Show this message and exit. │
╰──────────────────────────────────────────────────────────────────────────╯
╭─ Commands ───────────────────────────────────────────────────────────────╮
│ auth Store the API key for embedl-hub CLI. │
│ init Configure persistent CLI context. │
│ show Print the active project name and artifact directory. │
│ compile Compile a model for on-device deployment. │
│ profile Profile a compiled model on a target device. │
│ invoke Run inference on a compiled model. │
│ log Show past runs from the artifact directory. │
│ list-devices List available devices. │
╰──────────────────────────────────────────────────────────────────────────╯
Python API
For programmatic use, import from the embedl_hub package. The API provides
compiler, profiler, and invoker components for each supported backend
(ONNX Runtime, TensorRT, TFLite):
from embedl_hub.compile import OnnxRuntimeCompiler
from embedl_hub.profile import OnnxRuntimeProfiler
from embedl_hub.invoke import OnnxRuntimeInvoker
See the Embedl Hub documentation for detailed guides and examples.
License
Copyright (C) 2025, 2026 Embedl AB
This software is subject to the Embedl Hub Software License Agreement.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file embedl_hub-2026.4.1.dev0.tar.gz.
File metadata
- Download URL: embedl_hub-2026.4.1.dev0.tar.gz
- Upload date:
- Size: 128.2 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.11.15
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
18751db36ef7292048be09c36f1bbc302a64a7ff719f35292459a00736f902f4
|
|
| MD5 |
62e375817b773955e72e958dcdf7509f
|
|
| BLAKE2b-256 |
240164a2af8b0752b79f88a1d4fc96a11ffed56a2eef31f0b5d2ecaf70f3df78
|
File details
Details for the file embedl_hub-2026.4.1.dev0-py3-none-any.whl.
File metadata
- Download URL: embedl_hub-2026.4.1.dev0-py3-none-any.whl
- Upload date:
- Size: 166.7 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.11.15
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
d23bf3c1f8f1a0763916d0343272e38be0261f60f06616ccf99872b645ec2d7a
|
|
| MD5 |
63ae71a85a828eff23136c6cfbd2f0e8
|
|
| BLAKE2b-256 |
d733fa9457ea4cba4be211cf515c8b336e8f2914fe99a26c4aba2db31f2c1e64
|