Skip to main content

Physics simulation and annotation tools for Blender

Project description

VibePhysics

VibePhysics Teaser

A lightweight Blender physics simulation framework for creating realistic robot animations, rigid body physics, water dynamics, and comprehensive annotation tools — all running efficiently on CPU.

⚙️ Installation (MacOS)

# 1. Create environment
conda create -n vibephysics python=3.11
conda activate vibephysics

# 2. Install core package (includes COLMAP mapping & Blender simulation)
pip install vibephysics

# 3. (Optional) Install GLOMAP backend
# Linux users: refer to "Linux System Dependencies" below first
pip install git+https://github.com/shamangary/glomap.git

🐧 Linux (Ubuntu) System Dependencies

If you are on Linux and want to use the GLOMAP or COLMAP backends, you must install the following C++ development libraries to enable successful compilation:

sudo apt-get update
sudo apt-get install -y \
    libeigen3-dev \
    libceres-dev \
    libgoogle-glog-dev \
    libboost-all-dev \
    libsuitesparse-dev \
    libsqlite3-dev \
    libgflags-dev \
    libfreeimage-dev \
    libmetis-dev

⚠️ Troubleshooting (Linker Errors)

If you are using Anaconda on Linux and see an error like relocation R_X86_64_TPOFF32 ... can not be used when making a shared object, it is due to a conflict with the Anaconda linker. Fix it by forcing the compiler to use the global-dynamic TLS model:

export CXXFLAGS="$CXXFLAGS -fPIC -ftls-model=global-dynamic"
export CFLAGS="$CFLAGS -fPIC"
pip install git+https://github.com/shamangary/glomap.git

🗺️ Mapping & Reconstruction

VibePhysics integrates high-performance Structure-from-Motion (SfM) engines to convert image sequences into 3D reconstructions.

  • GLOMAP Engine – Global SfM that is 1-2 orders of magnitude faster than traditional methods.
  • COLMAP Engine – Industry-standard incremental SfM for robust reconstruction.
  • GSplat Ready – Automatically generates standard output structures (sparse/0 and images/ symlink) ready for instant GSplat training.

💻 1. Run Reconstruction (SfM)

You can run the SfM pipeline via shell script or Python API:

Command Line:

# Run GLOMAP pipeline (Fastest - Default)
./run_glomap.sh --image_path path/to/images

# Run COLMAP pipeline (Most Robust)
./run_glomap.sh --image_path path/to/images --engine colmap

# Advanced options
./run_glomap.sh --image_path path/to/images --matcher sequential --camera_model PINHOLE

Python API:

from vibephysics import mapping

# Simple Usage (Directly run GLOMAP on a folder of images)
mapping.glomap_pipeline(image_path="path/to/images")

# COLMAP Incremental Pipeline
mapping.colmap_pipeline(image_path="path/to/images")

🎨 2. Visualize in Blender

Load your reconstruction into Blender for inspection with colored point clouds and correct camera poses.

Command Line:

# Visualize a sparse model folder (containing cameras.bin, points3D.bin etc.)
./run_glomap_visual.sh --sparse path/to/sparse/0 --output result.blend

Python API:

from vibephysics import mapping

# Load reconstruction directly into active Blender scene
mapping.load_colmap_reconstruction(
    input_path="output/mapping_output/sparse/0",
    point_size=0.03,
    rotation=(-90, 0, 0) # Optional: global rotation
)

🎬 Simulation Results (sh run_robot.sh)

Result Demo

Robot walking simulation with rigid body physics, interacting with uneven ground, puddles, and real-time annotation overlay.

📊 Annotation Tools Demo (sh run_basics.sh)

Annotation Demo

Comprehensive annotation system featuring bounding boxes, motion trails, and point cloud tracking for computer vision datasets.

🎯 Dynamic Frustum Culling Demo (sh run_basics.sh)

Frustum Demo

Per-point frustum culling with mounted camera. Points inside the camera frustum turn red in real-time as the camera moves.

💧 Water Simulation Demo (sh run_water.sh)

Water Float Demo

Water physics simulation with floating objects, buoyancy forces, dynamic ripples, and point cloud tracking.

🐕 Go2 Simulation Demo (python examples/go2/go2_waypoint_walk.py)

Go2 Demo

Unitree Go2 robot walking through a physics-enabled environment with water puddles and falling debris.

✨ Highlights

  • 🚀 No GPU Required – Runs efficiently on CPU-only machines (MacBook Pro, laptops, standard workstations). GPU accelerates rendering but is not mandatory.
  • 🤖 Robot Simulation – Realistic IK-based walking animations with Open Duck and Unitree Go2 robots
  • 💧 Water Physics – Dynamic water surfaces, puddles, ripples, and buoyancy simulation
  • 📊 Annotation Tools – Bounding boxes, motion trails, and point cloud tracking for vision datasets
  • 🎯 Production Ready – Clean API, modular architecture, and extensive examples
  • 🗺️ SfM Mapping – Integrated COLMAP and GLOMAP pipelines for high-speed 3D reconstruction
  • 🔧 Developer Friendly – Pure Python, works with Blender as a module (bpy), no GUI needed

Perfect for researchers, animators, and robotics engineers who need physics simulations without expensive GPU hardware.

Requirements

For Running Simulations

  • Python 3.11 (required for bpy compatibility - Python 3.12+ is not supported)
  • bpy (Blender as a Python module)

For Viewing Results (Optional)

  • Blender 5.0 - Free download from blender.org
  • Only needed to view/render the generated .blend files
  • Not required to run simulations

⚠️ Important: This package requires Python 3.11. Python 3.12 and later versions are not compatible with the current version of bpy.

Dependency

  • Open Duck: We use the Open Duck blender model as demo. We do not own the model. Please refer to the original github repo.
  • Unitree Go2: We use the Unitree Go2 USD model. The model is auto-downloaded when running Go2 examples. We do not own the model.

Quick Start

# Run basic annotation demos (bbox, motion trail, point tracking, frustum culling)
sh ./run_basics.sh

# Run Open Duck robot simulation (with mounted POV camera by default)
sh ./run_robot.sh

# Run robot simulation with different camera views
sh ./run_robot.sh mounted    # First-person POV (default)
sh ./run_robot.sh center     # Overview from multiple angles
sh ./run_robot.sh following  # Third-person tracking shot

# Run Unitree Go2 robot simulation (auto-downloads model on first run)
python examples/go2/go2_waypoint_walk.py

# Go2 with custom settings
python examples/go2/go2_waypoint_walk.py --end-frame 150 --num-spheres 50

# Run forest walk simulation (robot walking through dense forest)
sh ./run_forest.sh

# Run forest with frustum culling options
sh ./run_forest.sh --frustum-mode highlight    # In-frustum points turn red
sh ./run_forest.sh --frustum-mode frustum_only # Only show in-frustum points
sh ./run_forest.sh --no-physics                # Fastest playback

# Run water simulations
sh ./run_water.sh

Visualizing Results

All simulations generate .blend files in the output/ directory. To view and interact with these results:

Download Blender 5.0 (Free & Open Source)

  • 🔗 Download Blender
  • Compatible with Windows, macOS (Intel/Apple Silicon), and Linux
  • No installation required for VibePhysics to run – Blender is only needed to view results
  • GPU accelerates viewport and rendering performance, but CPU-only works fine

Opening Results:

# macOS
open output/robot_waypoint.blend

# Linux
blender output/robot_waypoint.blend

# Windows
start output/robot_waypoint.blend

Once in Blender, press Spacebar to play the animation and view your physics simulation!

Camera System

VibePhysics includes a flexible multi-camera system with three camera rig types:

Camera Type Description Best For
Center Multiple cameras arranged in a circle, pointing at scene center Overview shots, multi-angle captures
Mounted Cameras attached directly to an object (e.g., robot head) First-person POV, onboard views
Following Single camera that follows and tracks a target object Third-person view, tracking shots

Usage Example

from vibephysics.camera import CameraManager

cam_manager = CameraManager()

# Center-pointing cameras (fixed position, looking at origin)
center_rig = cam_manager.add_center_pointing('center', num_cameras=4, radius=25, height=12)
center_rig.create(target_location=(0, 0, 0))

# Mounted cameras (attached to robot head for POV shots)
mounted_rig = cam_manager.add_object_mounted('mounted', num_cameras=4, distance=0.15)
mounted_rig.create(parent_object=robot_head, lens=10)

# Following camera (tracks a moving object)
follow_rig = cam_manager.add_following('following', height=12, look_angle=60)
follow_rig.create(target=robot_armature)

# Activate a specific camera
cam_manager.activate_rig('mounted', camera_index=0)  # Front camera

Command Line Options

Robot simulations support camera selection via shell script or Python:

# Via shell script (recommended)
sh run_robot.sh mounted    # First-person POV (default)
sh run_robot.sh center     # Overview from multiple angles
sh run_robot.sh following  # Third-person tracking shot

# Via Python directly
python examples/robot/robot_waypoint_walk.py --active-camera mounted
python examples/robot/robot_waypoint_walk.py --active-camera center
python examples/robot/robot_waypoint_walk.py --active-camera following

Switching Cameras in Blender

All three camera rigs are created in every .blend file — the command line option only sets which one is active by default. You can manually switch between any camera directly in Blender:

  1. Open the .blend file in Blender
  2. Press Numpad 0 to view through the active camera
  3. Switch cameras using one of these methods:
    • Outliner (Easiest): In the top-right Outliner panel, find camera objects (e.g., MountedCam_0, CenterCam_0, FollowingCam) → Click the green camera icon 🎥 next to the camera name to make it active
    • Right-click Menu: Right-click a camera in Outliner → Set Active Camera
    • Keyboard: Select a camera → Press Ctrl + Numpad 0 to make it active
    • View Menu: View → Cameras → Set Active Object as Camera

💻 Mac Users: Simply click the green camera icon 🎥 in the Outliner (see above) to switch active cameras.

This means you can generate a single .blend file and render from any camera angle without re-running the simulation.

Setup Module

The setup module provides scene initialization, asset import/export, and viewport management:

from vibephysics import setup

# Initialize a simulation scene
setup.init_simulation(start_frame=1, end_frame=250)

# Load assets (auto-detects format from file extension)
setup.load_asset('robot.glb')           # GLB/GLTF
setup.load_asset('mesh.fbx')            # FBX
setup.load_asset('points.ply')          # PLY

# Save/export (auto-detects format)
setup.save_blend('output/scene.blend')  # Creates directories automatically

# For format-specific options, use submodules directly:
from vibephysics.setup import importer, exporter

objects = importer.load_glb('model.glb', transform={'scale': 0.5})
exporter.export_fbx('output.fbx', selected_only=True)

Supported Formats

Import Export
GLB/GLTF Blend
FBX GLB/GLTF
PLY FBX
OBJ OBJ
STL PLY
DAE (Collada) STL
USD/USDA/USDC USD
Blend (append)

Gaussian Splatting (3DGS) (BETA)

VibePhysics supports loading 3D Gaussian Splatting data. [Warning] Currently it's under development

sh run_3dgs_viewer.sh

License

© 2025 MIMI AI LTD, UK. All rights reserved.

Academic & Student Use (Free)

This software is free to use for:

  • Students
  • Academic research
  • Educational purposes

Commercial Use

For business or enterprise use, please contact: tsunyi@mimiaigen.com We have separate license for business/enterprise users.

Citation

@misc{VibePhysics,
  author = {Tsun-Yi Yang},
  title = {VibePhysics: Physics and Robotics Simulation in Blender Without GPU Requirements},
  month = {December},
  year = {2025},
  url = {https://github.com/mimiaigen/vibephysics}
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

vibephysics-0.2.4.tar.gz (112.3 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

vibephysics-0.2.4-py3-none-any.whl (113.6 kB view details)

Uploaded Python 3

File details

Details for the file vibephysics-0.2.4.tar.gz.

File metadata

  • Download URL: vibephysics-0.2.4.tar.gz
  • Upload date:
  • Size: 112.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.14

File hashes

Hashes for vibephysics-0.2.4.tar.gz
Algorithm Hash digest
SHA256 b978fe0436483ff63aa9118e1c93870d2c74c563ac66b31ccd6a4b4ef5bf7de7
MD5 18bd17715f1c3eefc0c7218ffb61e54a
BLAKE2b-256 5bb8a91a05eab38e116a1ce2da4fdfae7b24731c5638f54b2d035ab33116760f

See more details on using hashes here.

File details

Details for the file vibephysics-0.2.4-py3-none-any.whl.

File metadata

  • Download URL: vibephysics-0.2.4-py3-none-any.whl
  • Upload date:
  • Size: 113.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.14

File hashes

Hashes for vibephysics-0.2.4-py3-none-any.whl
Algorithm Hash digest
SHA256 9994d0010593b856440cc0050bad7a5f5d02d4146da375c1e5d69472f88cd503
MD5 805e27ee4df3a143f2709bce91604128
BLAKE2b-256 e94215ffedfad7ba947cc5a5f4da5dcf3f75d8868ccf6835b9ffe7e65e80e81a

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page