DHB Extended Representations - SE(3) invariant trajectory encoding for robotics and VLA
Project description
dhb_xr
DHB Extended Representations — SE(3) invariant trajectory encoding for robotics, VLAs, and motion data management.
Overview
This library implements the double-reflection (DHB-DR) and quaternion-relative (DHB-QR) invariant representations for rigid-body motion trajectories on SE(3), as described in the manuscript "Double-Reflection DHB Invariant Representation on SE(3)". It provides:
- Encoding/Decoding: DHB-DR (Euler) and DHB-QR (quaternion) invariant computation and reconstruction
- DHB-TI (time-invariant): Reparameterize by geometric progress (translational arc-length, angular, or hybrid) and resample at uniform progress knots so invariants are approximately independent of execution speed and sampling rate; then encode with DHB-DR or DHB-QR
- Trajectory adaptation: Constrained optimization for retargeting demos to new start/goal poses
- GPU acceleration: PyTorch batched operations and optional Cusadi for large-scale optimization
- VLA support: VQ-VAE/RVQ tokenization for streaming action representation
- Motion database: Similarity search, DTW alignment, and retrieval
- Imitation learning: Invariant-space and geodesic losses
Installation
End users (pip)
# Basic installation
pip install dhb_xr
# With optimization (CasADi)
pip install dhb_xr[optimization]
# With GPU (PyTorch)
pip install dhb_xr[gpu]
# With examples and notebooks
pip install dhb_xr[examples]
# Full installation
pip install dhb_xr[all]
Developers (pixi)
# Install pixi: https://pixi.sh
curl -fsSL https://pixi.sh/install.sh | bash
# Clone and setup
cd dhb_xr
pixi install # installs default env (dev tools, jupyter, casadi, examples, build tools)
# Run tests
pixi run test
# Editable install (includes examples package)
pixi run build
# Run notebooks (CPU-only PyTorch)
pixi run notebook
# Copy examples for local development
pixi run dhb_xr-examples --copy ./local_examples
# Run examples programmatically
pixi run python -c "import dhb_xr_examples; dhb_xr_examples.run_basic_encoding()"
# Build for PyPI
pixi run build-dist
# Setup PyPI credentials
pixi run setup-pypirc
# Publish to TestPyPI
pixi run upload-testpypi
# Publish to PyPI (production)
pixi run upload-pypi
# Version management
pixi run version # Show current version
pixi run version --bump patch # Bump patch version
Publishing to PyPI
Automated (GitHub Actions)
- Update version:
pixi run version --bump patch - Commit and push changes
- Create and push a git tag:
git tag v0.2.1 && git push origin v0.2.1 - The
release.ymlworkflow will automatically build and publish to PyPI
Manual Publishing
# Set up PyPI credentials (run once)
pixi run setup-pypirc
# Edit ~/.pypirc with your API tokens
# Build distributions
pixi run build-dist
# Test on TestPyPI first
pixi run upload-testpypi
# Test installation from TestPyPI
pip install -i https://test.pypi.org/simple/ dhb_xr
# Publish to production PyPI
pixi run upload-pypi
Version Management
# Show current version
pixi run version
# Set specific version
pixi run version 0.2.1
# Bump version components
pixi run version --bump patch # 0.2.0 -> 0.2.1
pixi run version --bump minor # 0.2.0 -> 0.3.0
pixi run version --bump major # 0.2.0 -> 1.0.0
API Token Setup:
- TestPyPI: https://test.pypi.org/manage/account/#api-tokens
- PyPI: https://pypi.org/manage/account/#api-tokens
- Credentials stored in
~/.pypircor environment variables
CUDA Environment (GPU acceleration)
For GPU features (CusADi, VLA tokenization, faster PyTorch):
# Install the cuda environment (requires NVIDIA GPU with driver)
pixi install -e cuda
# Verify CUDA is available
pixi run -e cuda check-cuda
# Output: PyTorch 2.5.1, CUDA available: True, CUDA version: 12.4
# Run notebooks with CUDA
pixi run -e cuda notebook-cuda
# Run tests with CUDA
pixi run -e cuda test
Performance (GPU position decode):
- 1000 trajectories: 6.8 ms (146k traj/s)
- Per-trajectory: 6.8 µs
Technical notes on pixi + PyTorch CUDA setup
Getting CUDA-enabled PyTorch to work with pixi required careful configuration. Here are the key insights:
Problem: By default, pixi's dependency solver picks PyTorch from conda-forge, which is CPU-only (pytorch-2.x.x-cpu_mkl_*). Simply adding pytorch-cuda doesn't make it pick the CUDA build.
Solution: The cuda feature in pyproject.toml uses these techniques:
-
Channel priority: The cuda feature specifies
channels = ["pytorch", "nvidia", "conda-forge"]withchannel-priority = "strict"so PyTorch comes from the pytorch channel (which has CUDA builds), not conda-forge. -
Explicit channel specification: Dependencies use
{ version = ">=2.0", channel = "pytorch" }to force the pytorch channel:[tool.pixi.feature.cuda] channels = ["pytorch", "nvidia", "conda-forge"] channel-priority = "strict" [tool.pixi.feature.cuda.target.linux-64.dependencies] pytorch = { version = ">=2.0", channel = "pytorch" } pytorch-cuda = { version = ">=12.1", channel = "pytorch" }
-
Platform-specific:
pytorch-cudaonly exists forlinux-64, so we usetarget.linux-64.dependenciesto avoid solve failures on macOS. -
Separate solve group: The cuda environment uses
solve-group = "cuda"to avoid conflicts with the default CPU environment.
Verification:
# Check which pytorch package is installed
pixi list -e cuda | grep pytorch
# Should show: pytorch 2.x.x from pytorch channel (not conda-forge)
# Should show: pytorch-cuda 12.x from pytorch channel
# Verify CUDA is actually available
pixi run -e cuda python -c "import torch; print(torch.version.cuda)"
# Should print: 12.4 (not None)
Common pitfalls:
- Using
pytorch-cuda = "12.4"fails (ambiguous version); use"12.4.*"or">=12.1" - Not specifying channel priority causes conda-forge's CPU pytorch to be picked
- Forgetting
target.linux-64causes solve failures on non-Linux platforms
Examples Package
DHB-XR includes a comprehensive examples package that can be installed separately or cloned locally.
Option 1: Install Examples Package
pip install dhb_xr[examples]
Then run examples programmatically:
import dhb_xr_examples as examples
# Run basic encoding example
examples.run_basic_encoding()
# Or run individual examples
from dhb_xr_examples.basic_encoding import run_example
run_example()
Option 2: Copy Examples Locally
For development and experimentation, copy examples to a local directory:
# Copy to default location (./dhb_xr_examples)
dhb_xr-examples --copy
# Copy to specific directory
dhb_xr-examples --copy ~/my_dhb_examples
# List available examples
dhb_xr-examples --list
# Show examples location
dhb_xr-examples
This creates a local copy you can modify and experiment with.
The examples package includes:
- Core examples: Basic encoding/decoding, trajectory adaptation, DHB-DR vs QR
- Advanced examples: GPU batch optimization, VLA tokenization, motion databases
- VLA integration: Full LIBERO simulation, perturbation robustness demos
- Research examples: Imitation learning losses, time-invariant reparameterization
- Tutorial notebooks: Interactive Jupyter notebooks for learning DHB-XR concepts
Quick start
import numpy as np
from dhb_xr import encode_dhb_dr, decode_dhb_dr
from dhb_xr.core.types import DHBMethod
# Create or load trajectory: N poses (position + quaternion wxyz)
positions = np.cumsum(np.random.randn(50, 3) * 0.01, axis=0)
quaternions = np.tile(np.array([1.0, 0, 0, 0]), (50, 1)) # identity orientation
# Encode to invariants (DHB-DR: double reflection + Euler)
from dhb_xr.core.types import EncodingMethod
result = encode_dhb_dr(
positions, quaternions,
method=EncodingMethod.POSITION,
use_default_initial_frames=True,
dhb_method=DHBMethod.DOUBLE_REFLECTION,
)
linear_inv = result["linear_motion_invariants"]
angular_inv = result["angular_motion_invariants"]
# Decode back to trajectory
decoded = decode_dhb_dr(
linear_inv, angular_inv,
result["initial_pose"],
method=EncodingMethod.POSITION,
dhb_method=DHBMethod.DOUBLE_REFLECTION,
drop_padded=True,
)
print(decoded["positions"].shape, decoded["quaternions"].shape)
Time-invariant reparameterization (DHB-TI)
To reduce sensitivity to execution speed and sampling rate, reparameterize by a geometric progress variable and resample at uniform progress knots before encoding:
from dhb_xr.encoder.dhb_ti import compute_progress, resample_by_progress, encode_dhb_dr_ti
# Progress: translation (arc-length), angular, or hybrid σ = α||Δp|| + (1-α)||Δr||
progress = compute_progress(positions, quaternions, kind="hybrid", alpha=0.5)
pos_m, quat_m = resample_by_progress(positions, quaternions, M=30, progress_kind="hybrid", alpha=0.5)
# Time-invariant encode
out = encode_dhb_dr_ti(positions, quaternions, M=30, progress_kind="hybrid", alpha=0.5, ...)
See examples/08_dhb_ti_time_invariant.py.
Documentation
📚 Read the Docs - Complete API documentation with examples
The documentation is built with MkDocs and can be deployed to GitHub Pages on pushes to main
when Pages is enabled and set to build using GitHub Actions.
Build locally
# Install development dependencies (includes MkDocs)
pixi install
# Build documentation
pixi run docs # or: pixi run build-docs
# Serve locally for development
pixi run serve-docs # opens http://127.0.0.1:8000/
Without pixi
pip install mkdocs mkdocs-material mkdocstrings mkdocstrings-python
mkdocs build
mkdocs serve # opens http://127.0.0.1:8000/
CusADi GPU Acceleration (optional)
For batch processing thousands of trajectories, CusADi provides up to 387x speedup on GPU:
| Batch Size | CPU (ms) | GPU (ms) | Speedup |
|---|---|---|---|
| 100 | 34 | 0.8 | 43x |
| 1000 | 342 | 1.7 | 199x |
| 2000 | 685 | 1.8 | 387x |
Requirements:
- NVIDIA GPU with CUDA toolkit (nvcc)
- PyTorch with CUDA support
Setup with pixi:
cd dhb_xr
# 1. Install pixi environment
pixi install
# 2. Install PyTorch with CUDA (one-time)
pixi run install-cuda
# 3. Clone cusadi (if not already)
git clone https://github.com/se-hwan/cusadi /path/to/cusadi
cd /path/to/cusadi && pip install -e .
# 4. Build CasADi functions (use pixi python for version compatibility)
cd dhb_xr
pixi run python3 << 'EOF'
import casadi as ca
import numpy as np
def euler_to_rot(rx, ry, rz):
cx, sx = ca.cos(rx), ca.sin(rx)
cy, sy = ca.cos(ry), ca.sin(ry)
cz, sz = ca.cos(rz), ca.sin(rz)
return ca.vertcat(
ca.horzcat(cy*cz, -cy*sz, sy),
ca.horzcat(cx*sz + cz*sx*sy, cx*cz - sx*sy*sz, -cy*sx),
ca.horzcat(sx*sz - cx*cz*sy, cz*sx + cx*sy*sz, cx*cy))
T = 50
lin_inv = ca.SX.sym("lin_inv", T * 4)
init_pos = ca.SX.sym("init_pos", 3)
init_rot = ca.SX.sym("init_rot", 9)
rotm = ca.reshape(init_rot, 3, 3)
pos = init_pos
out = [init_pos]
for k in range(T):
mag, rx, ry, rz = lin_inv[k*4], lin_inv[k*4+1], lin_inv[k*4+2], lin_inv[k*4+3]
rotm = rotm @ euler_to_rot(rx, ry, rz)
pos = pos + rotm @ ca.vertcat(mag, 0, 0)
out.append(pos)
fn = ca.Function("fn_dhb_decode_linear", [lin_inv, init_pos, init_rot], [ca.horzcat(*out).T])
fn.save("/path/to/cusadi/src/casadi_functions/fn_dhb_decode_linear.casadi")
print(f"Saved: {fn}")
EOF
# 5. Compile CUDA kernel (use pixi python for same CasADi version)
cd /path/to/cusadi
/path/to/dhb_xr/.pixi/envs/default/bin/python3 run_codegen.py --fn=fn_dhb_decode_linear
Usage:
import sys
sys.path.insert(0, "/path/to/cusadi")
sys.path.insert(0, "/path/to/cusadi/src")
sys.path.insert(0, "/path/to/cusadi/build")
import torch
import casadi as ca
from src.CusadiFunction import CusadiFunction
fn = ca.Function.load("/path/to/cusadi/src/casadi_functions/fn_dhb_decode_linear.casadi")
cusadi_fn = CusadiFunction(fn, batch_size=1000)
# GPU tensors (batch_size, features)
lin_inv_gpu = torch.from_numpy(invariants).cuda().contiguous()
init_pos_gpu = torch.from_numpy(positions).cuda().contiguous()
init_rot_gpu = torch.from_numpy(rotations).cuda().contiguous()
cusadi_fn.evaluate([lin_inv_gpu, init_pos_gpu, init_rot_gpu])
positions = cusadi_fn.getDenseOutput(0).cpu().numpy() # (batch, T+1, 3)
Important: The CasADi .casadi files must be saved with the same CasADi version that loads them. Use pixi python for both building and running to ensure version compatibility.
See the CusADi paper for details on the parallelization framework.
Fatrop Fast Optimization (optional)
For single trajectory optimization with constraints, Fatrop provides ~10x speedup over IPOPT:
| Solver | Use Case | Speed |
|---|---|---|
| IPOPT | General NLP | ~50-100ms |
| Fatrop | Structured OCP | ~5-10ms |
Installation:
# Rockit (required for OCP formulation)
pip install rockit-meco
# or with pixi:
pixi run install-rockit
# Fatrop is bundled with conda casadi (no separate install needed)
# The pixi environment includes casadi with Fatrop support
Usage:
from dhb_xr.optimization import generate_trajectory_fatrop
result = generate_trajectory_fatrop(
demo_positions, demo_quaternions,
start_pose={'position': start_pos, 'quaternion': start_quat},
goal_pose={'position': goal_pos, 'quaternion': goal_quat},
traj_length=50,
use_fatrop=True, # False for IPOPT fallback
)
print(f"Solved in {result['solve_time']*1000:.1f} ms")
Use cases:
- Real-time MPC (100+ Hz replanning)
- Constrained trajectory generation (obstacles, joint limits)
- Online trajectory adaptation
CusADi vs Fatrop:
- CusADi: Best for batch evaluation (1000 trajectories in 2ms)
- Fatrop: Best for single optimization with constraints (5-10ms)
C++ extension (optional)
- Build (from repo root, with pixi):
pixi run build-cpp(requires nanobind in dev feature). This builds the nanobind module intosrc/dhb_xr/soimport dhb_xr._dhb_xr_cppworks. - Use:
from dhb_xr import cpp_version(returnsNoneif not built). Seesrc/dhb_xr/_cpp/README.mdfor extending with encode/decode.
VLA Integration (LIBERO-PRO / LIBERO / RoboCASA)
DHB-XR includes adapters for loading trajectory data from popular VLA benchmarks, with full support for LIBERO-PRO — the extended LIBERO benchmark that tests policy robustness under spatial, object, semantic, task, and environment perturbations.
Why DHB-XR for VLA: Current VLA models (RT-2, Octo, OpenVLA) map (vision + language) → actions, but struggle when the scene layout changes. DHB-XR provides a structured trajectory representation that decouples motion shape from spatial context:
| Without DHB-XR | With DHB-XR |
|---|---|
| Retrain or add data augmentation | Re-decode from new pose (~7ms, Fatrop solver) |
| Actions tied to absolute positions | Invariants are SE(3)-invariant by construction |
| 100s of demos per spatial arrangement | 1 demo + DHB adaptation covers spatial variations |
When LIBERO-PRO perturbs object positions or swaps objects, DHB can adapt the demonstration trajectory to the new configuration while perfectly preserving the original motion geometry (0.000 mm shape error).
Quick Start: DHB Encoding Only
No simulation required - just load and process trajectory data:
# 1. Download LIBERO-Spatial dataset (smallest, ~2.8GB compressed)
mkdir -p ~/Projects/data/libero && cd ~/Projects/data/libero
wget -O libero_spatial.zip "https://utexas.box.com/shared/static/04k94hyizn4huhbv5sz4ev9p2h1p6s7f.zip"
unzip libero_spatial.zip
# 2. Test DHB encoding (works with pixi environment)
pixi run python examples/integration/test_libero_adapter.py
pixi run python examples/integration/test_libero_encoding.py
# 3. Run full demo (DHB-only mode, no simulation, saves plot to /tmp/dhb_demo_plot.png)
pixi run python examples/integration/libero_full_demo.py --dhb-only
# 4. Motion retrieval demo
pixi run python examples/integration/libero_full_demo.py --retrieval
# 5. View generated plot
xdg-open /tmp/dhb_demo_plot.png # Linux
Programmatic Usage
from dhb_xr.integration.vla.libero import LiberoAdapter
from dhb_xr.encoder.dhb_dr import encode_dhb_dr
from dhb_xr.core.types import EncodingMethod, DHBMethod
# Load episodes from LIBERO HDF5
adapter = LiberoAdapter()
for episode in adapter.load_dataset("/path/to/libero_task.hdf5"):
positions = episode["positions"] # (N, 3) end-effector positions
quaternions = episode["quaternions"] # (N, 4) quaternions (x, y, z, w)
# Encode to SE(3)-invariant representation
result = encode_dhb_dr(
positions, quaternions,
method=EncodingMethod.POSITION,
dhb_method=DHBMethod.DOUBLE_REFLECTION,
)
invariants = result["linear_motion_invariants"] # Shape: (N+2, 4)
Full LIBERO / LIBERO-PRO Simulation
For running LIBERO tasks in simulation with DHB-XR trajectory adaptation and perturbation robustness testing:
# 1. Install Miniforge (if conda/mamba not available)
curl -L -O "https://github.com/conda-forge/miniforge/releases/latest/download/Miniforge3-$(uname)-$(uname -m).sh"
bash Miniforge3-$(uname)-$(uname -m).sh -b -p ~/miniforge3
# 2. Create and configure libero environment
~/miniforge3/bin/mamba create -n libero python=3.10 -y
~/miniforge3/bin/mamba run -n libero pip install robosuite==1.4.0 mujoco bddl==1.0.1 robomimic==0.2.0
~/miniforge3/bin/mamba run -n libero pip install future easydict hydra-core cloudpickle 'gym==0.25.2'
# 3. Clone and install LIBERO-PRO (drop-in replacement for LIBERO with perturbation support)
git clone https://github.com/Zxy-MLlab/LIBERO-PRO.git ~/Projects/repos/LIBERO-PRO
~/miniforge3/bin/mamba run -n libero pip install -e ~/Projects/repos/LIBERO-PRO --config-settings editable_mode=compat
# 4. Configure LIBERO paths (creates ~/.libero/config.yaml)
mkdir -p ~/.libero
cat > ~/.libero/config.yaml << 'EOF'
benchmark_root: ~/Projects/repos/LIBERO-PRO/libero/libero
bddl_files: ~/Projects/repos/LIBERO-PRO/libero/libero/bddl_files
init_states: ~/Projects/repos/LIBERO-PRO/libero/libero/init_files
datasets: ~/Projects/data/libero
assets: ~/Projects/repos/LIBERO-PRO/libero/libero/assets
EOF
# 5. Install dhb_xr and visualization dependencies
~/miniforge3/bin/mamba run -n libero pip install dhb_xr opencv-python imageio imageio-ffmpeg
# 6. Run simulation demo
~/miniforge3/bin/mamba run -n libero python examples/integration/libero_full_demo.py
Note: LIBERO-PRO is a drop-in replacement for LIBERO with identical core dependencies. It adds perturbation test suites (spatial swap, object replacement, language, task, environment) for evaluating policy robustness. All original LIBERO benchmarks (
libero_spatial,libero_goal, etc.) work unchanged.
Viewing Simulations
# Option 1: Real-time display with OpenCV (requires X11 display)
~/miniforge3/bin/mamba run -n libero python examples/integration/libero_full_demo.py --render
# Option 2: Save video for later viewing (works headless)
~/miniforge3/bin/mamba run -n libero python examples/integration/libero_full_demo.py --save-video demo.mp4
# Option 3: Both display and save
~/miniforge3/bin/mamba run -n libero python examples/integration/libero_full_demo.py --render --save-video demo.mp4
# Play saved video
vlc demo.mp4 # or: ffplay demo.mp4
For remote servers without display, use --save-video and download the video locally.
Key version requirements:
- robosuite==1.4.0 (LIBERO is incompatible with robosuite 1.5+)
- Python 3.10 recommended
- bddl==1.0.1, robomimic==0.2.0
DHB-XR vs Naive Replay — Swap Demo
The most compelling showcase of DHB-XR's value — directly comparing naive replay vs solver-adapted trajectory under spatial perturbation:
# Object positions swap (~17cm shift) — naive replay fails, DHB adapts
~/miniforge3/bin/mamba run -n libero python examples/integration/libero_swap_demo.py
# Results:
# Naive replay: 11.1 cm from NEW plate (wrong target)
# DHB-adapted: 4.6 cm from NEW plate (correct target, Fatrop ~7ms)
# Improvement: 6.5 cm closer to correct target
LIBERO-PRO Perturbation Robustness Demo
The libero_pro_dhb_demo.py script demonstrates how DHB's SE(3)-invariance enables robust trajectory adaptation under LIBERO-PRO's perturbation types:
# DHB analysis: encode demo, apply spatial perturbations, verify shape preservation
pixi run python examples/integration/libero_pro_dhb_demo.py --analysis
# Batch evaluation across multiple tasks (generates comparison plots)
pixi run python examples/integration/libero_pro_dhb_demo.py --batch
# Simulation: run original + perturbed variants, compare invariants
~/miniforge3/bin/mamba run -n libero python examples/integration/libero_pro_dhb_demo.py --simulate
# With comparison video
~/miniforge3/bin/mamba run -n libero python examples/integration/libero_pro_dhb_demo.py --simulate --save-video comparison.mp4
Key results:
| Metric | Value |
|---|---|
| Reconstruction error | 0.000 mm |
| Shape error (20mm perturbation) | 0.000 mm |
| Shape error (50mm perturbation) | 0.000 mm |
| Shape error (100mm perturbation) | 0.000 mm |
| Invariant correlation (with_mug variant) | 0.990 |
| Invariant correlation (with_milk variant) | 0.975 |
DHB invariants are perfectly frame-independent: adapting a trajectory to any perturbed starting pose preserves the original motion shape with zero error. Even under LIBERO-PRO's object replacement perturbations, the invariant representation of the same motion achieves >0.97 correlation.
LIBERO-PRO perturbation types:
| Type | Description | LIBERO-PRO Benchmark |
|---|---|---|
| Position/Swap | Objects swap positions on the table | libero_spatial_swap |
| Object | Replace objects with visually different ones | libero_spatial_object |
| Semantic | Change language instructions | libero_spatial_lan |
| Task | Change goal/task entirely | libero_spatial_task |
| Environment | Change table/scene environment | libero_spatial_env |
See VLA Integration Guide for full documentation.
Testing
Full test suite (pixi)
cd dhb_xr
pixi install
pixi run test
Or without pixi: PYTHONPATH=src pytest tests/ -v.
C++ extension (nanobind)
-
Build the extension. Nanobind must be available to CMake (e.g. conda:
conda install -c conda-forge nanobind; or setnanobind_DIRto the nanobind install share path). With pixi (default env has nanobind from conda-forge):pixi run build-cpp
If pixi solve fails (e.g. CUDA), use a minimal env:
conda install -c conda-forge python cmake ninja nanobind, then from repo root:mkdir build && cd build cmake .. -DCMAKE_BUILD_TYPE=Release cmake --build . cp src/dhb_xr/_cpp/_dhb_xr_cpp*.so ../src/dhb_xr/
-
Run C++ tests (skip if extension not built):
pixi run test -- tests/test_cpp.py -v
Or run the checks manually:
PYTHONPATH=src python3 -c " from dhb_xr import cpp_version if cpp_version: print('C++ extension:', cpp_version()) from dhb_xr import _dhb_xr_cpp print('add(1,2)=', _dhb_xr_cpp.add(1.0, 2.0)) else: print('C++ extension not built (pixi run build-cpp)') "
Cusadi implementation
Cusadi tests cover batched_decode_dhb_dr and CusadiTrajectoryOptimizer (NumPy fallback; no cusadi package required):
pixi run test -- tests/test_cusadi.py -v
- batched_decode_dhb_dr: batch decode; test compares with single
decode_dhb_drfor consistency. - CusadiTrajectoryOptimizer.forward: same batch decode via the optimizer interface.
- export_casadi_decode (optional): if
casadiis installed (pip install dhb_xr[optimization]), one test exports a.casadidecode step.
To test the CasADi export script explicitly:
pip install dhb_xr[optimization]
python -m dhb_xr.optimization.export_casadi_decode --out /tmp/fn_dhb_decode_step.casadi
# Check: ls /tmp/fn_dhb_decode_step.casadi
References
- D. Lee, R. Soloperto, M. Saveriano, "Bidirectional invariant representation of rigid body motions and its application to gesture recognition and reproduction", Autonomous Robots, 2018.
- R. Soloperto, M. Saveriano, D. Lee, "A Bidirectional Invariant Representation of Motion for Gesture Recognition and Reproduction", ICRA, 2015.
- W. Wang et al., "Computation of rotation minimizing frames", ACM TOG, 2008.
License
MIT
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file dhb_xr-0.3.0.tar.gz.
File metadata
- Download URL: dhb_xr-0.3.0.tar.gz
- Upload date:
- Size: 177.1 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.14.2
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
20ce6a6779fe4bf73b3187651a1a61c30feaa4027a70ba053ebf0bf13b66b609
|
|
| MD5 |
86430963e7315908817340d721cc26e6
|
|
| BLAKE2b-256 |
e02fd93b58ee535724b4561b57f404c7333d1b3b21a380d287c44142b02117f0
|
File details
Details for the file dhb_xr-0.3.0-py3-none-any.whl.
File metadata
- Download URL: dhb_xr-0.3.0-py3-none-any.whl
- Upload date:
- Size: 185.8 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.14.2
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
a40e73f8936e60a9109c4a4e5a119b79e6d94d54d5aafd5c01e65c117cb35b48
|
|
| MD5 |
5fdb62d3e9420945fea2711a476b8527
|
|
| BLAKE2b-256 |
460429be6682ce3e2f832896dcdb731c2197c4d530bd49c4279a9dc0eadc8365
|