An Multi-Modality Foundation Model for Humanoid robots
Project description
NeoCortex: Multi-Modality Foundation Model for Humanoid Robots
NeoCortex is a cutting-edge, multi-modality foundation model engineered for humanoid robots. With the capability to process an array of modalities including images, text, videos, depth, heatmaps, LIDAR, 3D scenes, point clouds, and more, NeoCortex represents the zenith of robotic perception and cognition.
Appreciation
- All the creators in Agora, Join Agora the community of AI engineers changing the world with their creations.
- LucidRains for inspiring me to devote myself to open source AI
Installation
To integrate NeoCortex into your development environment:
pip install neox
Usage
import torch
from neox.model import NeoCortex
#usage
img = torch.randn(1, 3, 256, 256)
caption_tokens = torch.randint(0, 4)
model = NeoCortex()
output = model(img, caption_tokens)
Model Architecture
Building on the paradigms established by models like NeVA, NeoCortex expands its horizons to embrace a multi-faceted input system:
- Images: Processed through a frozen variant of the Hugging Face CLIP model, producing image embeddings.
- Text: Integrated seamlessly with an NVIDIA-trained GPT variant.
- LIDAR: Point clouds processed through PointNet or KPConv, extracting essential spatial features. These features, once captured, are projected to text embedding dimensions.
- Videos: Processed using 3D-CNNs to capture temporal information, then integrated with the main architecture.
- Heatmaps, 3D Scenes, and Depth: These are funneled through specialized modules tailored for each modality before being integrated.
Training comprises three stages:
- Pretraining: Specific modules (like PointNet for LIDAR) are pretrained, with the main model frozen.
- Inter-modality Training: The embeddings from different modalities are concatenated, projected, and trained to ensure inter-modality coherence.
- Finetuning: Utilizing synthetic data generated with advanced GPT versions to ensure the model understands the intricate relationship between various modalities.
Specifications
- Architecture Type: Multi-Modality Transformer
- Sub-Architectures: GPT, CLIP, PointNet, 3D-CNN
- Model versions: 10B, 30B, 50B
Input & Output
- Input Formats:
- Images: RGB
- Text: Tokens
- LIDAR: Point Clouds
- Videos: Frames
- And more...
- Output Format: Text or Action Tokens
Integration and Compatibility
- Supported Hardware Platforms: Hopper, Ampere/Turing, specialized robotic chipsets
- Supported Operating Systems: Linux, RoboOS
Training & Fine-tuning Data
- Image Dataset: [Link to Dataset]
- LIDAR Dataset: [Link to Dataset]
- Textual Data: Synthetically produced by GPT versions
- Licenses: Various, mainly MIT and CC-BY-NC 4.0
Inference
- Engine: Triton and specialized robotic inference engines
- Test Hardware: Humanoid robotic platforms
References and More
Licensing
This project is licensed under the MIT License.
For more details, contributions, and support, please refer to the official repository.
Todo
[] -
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file neox-0.1.3.tar.gz
.
File metadata
- Download URL: neox-0.1.3.tar.gz
- Upload date:
- Size: 22.9 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.3.2 CPython/3.11.0 Darwin/22.4.0
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 2aa8937c2d1cbd0002079fb6f8bb2a87f18f1a3efbb8bfbeb5340e4c3959bb48 |
|
MD5 | 2f8f64e84f5f8dcaec19e42608bcbc4a |
|
BLAKE2b-256 | 5ef2a1e8aaa28c73dcbd85481c9d5fa0c4747e5f691e2d6932fd7b8143fcedad |
File details
Details for the file neox-0.1.3-py3-none-any.whl
.
File metadata
- Download URL: neox-0.1.3-py3-none-any.whl
- Upload date:
- Size: 22.0 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.3.2 CPython/3.11.0 Darwin/22.4.0
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 6611add5b936c77f1ea2536a9305d47d0f408618d5ecd7ca974d60a22b21c6f1 |
|
MD5 | 27e06567f8af64edc1cb7683ae0a2359 |
|
BLAKE2b-256 | 59b8ebcbdc5bf166b145456a840eee1371fdc5be48b2532abadc6b1f6c43fd99 |