High-performance data ingestion tool for Milvus vector database with vectorized operations
Project description
Milvus Ingest - High-Performance Data Ingestion Tool
๐ Ultra-fast data ingestion tool for Milvus vector databases - Built for large-scale data generation and ingestion with vectorized operations, parallel processing, and optimized file I/O. Generate and ingest millions of rows in seconds with automatic file partitioning and intelligent memory management.
โก Performance Highlights
- ๐๏ธ 10,000-100,000+ rows/sec - Vectorized NumPy operations for maximum speed
- ๐ Large-scale optimized - Designed for datasets >100K rows with intelligent batching
- ๐ฅ Smart file partitioning - Automatic splitting (256MB chunks, 1M rows/file)
- ๐พ Memory efficient - Streaming generation prevents memory exhaustion
- โก Direct PyArrow I/O - Optimized Parquet writing with Snappy compression
- ๐ Parallel processing - Multi-core CPU utilization with configurable workers
โจ Key Features
- ๐ฏ Ready-to-use schemas - Pre-built schemas for e-commerce, documents, images, users, news, and videos
- ๐ Schema management - Add, organize, and reuse custom schemas with metadata
- ๐ High-performance generation - Vectorized operations optimized for large datasets
- ๐ง Complete Milvus support - All field types including vectors, arrays, JSON, and primitive types
- โ Smart validation - Pydantic-based validation with detailed error messages and suggestions
- ๐ High-performance formats - Parquet (fastest I/O), JSON (structured data)
- ๐ฑ Reproducible results - Seed support for consistent data generation
- ๐จ Rich customization - Field constraints, nullable fields, auto-generated IDs
- ๐ Schema exploration - Validation, help commands, and schema details
- ๐ Unified interface - Use custom and built-in schemas interchangeably
Installation
# Install from source (recommended for development)
git clone https://github.com/zilliz/milvus-ingest.git
cd milvus-ingest
pdm install # Installs with development dependencies
# For production use only
pdm install --prod
# After installation, the CLI tool is available as:
milvus-ingest --help
๐ Quick Start
1. Use Built-in Schemas (Recommended)
Get started instantly with pre-built schemas optimized for large-scale generation:
# List all available built-in schemas
milvus-ingest schema list
# Generate data using a built-in schema (high-performance by default)
milvus-ingest generate --builtin simple --rows 100000 --preview
# Generate large e-commerce dataset with automatic file partitioning
milvus-ingest generate --builtin ecommerce --rows 2500000 --out products/
Available Built-in Schemas:
| Schema | Description | Use Cases |
|---|---|---|
simple |
Basic example with common field types | Learning, testing |
ecommerce |
Product catalog with search embeddings | Online stores, recommendations |
documents |
Document search with semantic embeddings | Knowledge bases, document search |
images |
Image gallery with visual similarity | Media platforms, image search |
users |
User profiles with behavioral embeddings | User analytics, personalization |
videos |
Video library with multimodal embeddings | Video platforms, content discovery |
news |
News articles with sentiment analysis | News aggregation, content analysis |
audio_transcripts |
Audio transcription with FP16 embeddings | Speech-to-text search, podcasts |
ai_conversations |
AI chat history with BF16 embeddings | Chatbot analytics, conversation search |
face_recognition |
Facial recognition with binary vectors | Security systems, identity verification |
ecommerce_partitioned |
Partitioned e-commerce schema | Scalable product catalogs |
cardinality_demo |
Schema demonstrating cardinality features | Testing cardinality constraints |
2. Create Custom Schemas
Define your own collection structure with JSON or YAML:
{
"collection_name": "my_collection",
"fields": [
{
"name": "id",
"type": "Int64",
"is_primary": true
},
{
"name": "title",
"type": "VarChar",
"max_length": 256
},
{
"name": "embedding",
"type": "FloatVector",
"dim": 128
}
]
}
# Generate large dataset from custom schema with high-performance mode
milvus-ingest generate --schema my_schema.json --rows 1000000 --format parquet --preview
Note: Output is always a directory containing data files (in the specified format) and a meta.json file with collection metadata.
3. Schema Management
Store and organize your schemas for reuse:
# Add a custom schema to your library
milvus-ingest schema add my_products product_schema.json
# List all schemas (built-in + custom)
milvus-ingest schema list
# Use your custom schema like a built-in one (optimized for large datasets)
milvus-ingest generate --builtin my_products --rows 500000
# Show detailed schema information
milvus-ingest schema show my_products
4. Python API
from milvus_ingest.generator import generate_mock_data
from milvus_ingest.schema_manager import get_schema_manager
from milvus_ingest.builtin_schemas import load_builtin_schema
import tempfile
import json
# Use the schema manager to work with schemas
manager = get_schema_manager()
# List all available schemas
all_schemas = manager.list_all_schemas()
print("Available schemas:", list(all_schemas.keys()))
# Load any schema (built-in or custom)
schema = manager.get_schema("ecommerce") # Built-in
# schema = manager.get_schema("my_products") # Custom
# Generate data from schema file
with tempfile.NamedTemporaryFile(mode='w', suffix='.json', delete=False) as f:
json.dump(schema, f, indent=2)
f.flush()
# Generate data (returns dict with DataFrame and metadata)
result = generate_mock_data(f.name, rows=10000, seed=42, output_format="dict")
df = result["data"]
metadata = result["metadata"]
print(df.head())
print(f"Generated {len(df)} rows for collection: {metadata['collection_name']}")
# Add a custom schema programmatically
custom_schema = {
"collection_name": "my_collection",
"fields": [
{"name": "id", "type": "Int64", "is_primary": True},
{"name": "text", "type": "VarChar", "max_length": 100},
{"name": "vector", "type": "FloatVector", "dim": 256}
]
}
manager.add_schema("my_custom", custom_schema, "Custom schema", ["testing"])
print("Added custom schema!")
๐ Schema Reference
Supported Field Types
| Type | Description | Required Parameters | Optional Parameters |
|---|---|---|---|
| Numeric Types | |||
Int8, Int16, Int32, Int64 |
Integer types | - | min, max |
Float, Double |
Floating point | - | min, max |
Bool |
Boolean values | - | - |
| Text Types | |||
VarChar, String |
Variable length string | max_length |
- |
JSON |
JSON objects | - | - |
| Vector Types | |||
FloatVector |
32-bit float vectors | dim |
- |
BinaryVector |
Binary vectors | dim |
- |
Float16Vector |
16-bit float vectors | dim |
- |
BFloat16Vector |
Brain float vectors | dim |
- |
SparseFloatVector |
Sparse float vectors | dim |
- |
| Complex Types | |||
Array |
Array of elements | element_type, max_capacity |
max_length (for string elements) |
Field Properties
| Property | Description | Applicable Types |
|---|---|---|
is_primary |
Mark field as primary key (exactly one required) | All types |
auto_id |
Auto-generate ID values | Int64 primary keys only |
nullable |
Allow null values (10% probability) | All types |
min, max |
Value constraints | Numeric types |
max_length |
String/element length limit | String and Array types |
dim |
Vector dimension (1-32768) | Vector types |
element_type |
Array element type | Array type |
max_capacity |
Array capacity (1-4096) | Array type |
Complete Example
collection_name: "advanced_catalog"
fields:
# Primary key with auto-generated IDs
- name: "id"
type: "Int64"
is_primary: true
auto_id: true
# Text fields with constraints
- name: "title"
type: "VarChar"
max_length: 200
- name: "description"
type: "VarChar"
max_length: 1000
nullable: true
# Numeric fields with ranges
- name: "price"
type: "Float"
min: 0.01
max: 9999.99
- name: "rating"
type: "Int8"
min: 1
max: 5
# Vector for semantic search
- name: "embedding"
type: "FloatVector"
dim: 768
# Array of tags
- name: "tags"
type: "Array"
element_type: "VarChar"
max_capacity: 10
max_length: 50
# Structured metadata
- name: "metadata"
type: "JSON"
nullable: true
# Boolean flags
- name: "in_stock"
type: "Bool"
๐ CLI Reference
Command Structure
The CLI uses a clean grouped structure:
# Main command groups
milvus-ingest generate [options] # Data generation
milvus-ingest schema [command] # Schema management
milvus-ingest clean [options] # Utility commands
Data Generation Commands
| Command | Description | Example |
|---|---|---|
--schema PATH |
Generate from custom schema file | milvus-ingest generate --schema my_schema.json |
--builtin SCHEMA_ID |
Use built-in or managed schema | milvus-ingest generate --builtin ecommerce |
--rows INTEGER |
Number of rows to generate | milvus-ingest generate --rows 5000 |
--format FORMAT |
Output format (parquet, json) | milvus-ingest generate --format json |
--out DIRECTORY |
Output directory path | milvus-ingest generate --out my_data/ |
--preview |
Show first 5 rows | milvus-ingest generate --preview |
--seed INTEGER |
Random seed for reproducibility | milvus-ingest generate --seed 42 |
--validate-only |
Validate schema without generating | milvus-ingest generate --validate-only |
--no-progress |
Disable progress bar display | milvus-ingest generate --no-progress |
--batch-size INTEGER |
Batch size for memory efficiency (default: 50000) | milvus-ingest generate --batch-size 100000 |
--max-file-size INTEGER |
Maximum size per file in MB (default: 256) | milvus-ingest generate --max-file-size 100 |
--max-rows-per-file INTEGER |
Maximum rows per file (default: 1000000) | milvus-ingest generate --max-rows-per-file 500000 |
--force |
Force overwrite output directory | milvus-ingest generate --force |
Schema Management Commands
| Command | Description | Example |
|---|---|---|
schema list |
List all schemas (built-in + custom) | milvus-ingest schema list |
schema show SCHEMA_ID |
Show schema details | milvus-ingest schema show ecommerce |
schema add SCHEMA_ID FILE |
Add custom schema | milvus-ingest schema add products schema.json |
schema remove SCHEMA_ID |
Remove custom schema | milvus-ingest schema remove products |
schema help |
Show schema format help | milvus-ingest schema help |
Utility Commands
| Command | Description | Example |
|---|---|---|
clean |
Clean up generated output files | milvus-ingest clean --yes |
--help |
Show help message | milvus-ingest --help |
Common Usage Patterns
# Quick start with built-in schema (high-performance by default)
milvus-ingest generate --builtin simple --rows 100000 --preview
# Generate massive datasets with automatic file partitioning
milvus-ingest generate --builtin ecommerce --rows 5000000 --format parquet --out products/
# Test custom schema validation
milvus-ingest generate --schema my_schema.json --validate-only
# Reproducible large-scale data generation
milvus-ingest generate --builtin users --rows 2000000 --seed 42 --out users/
# Control file partitioning (smaller files for easier handling)
milvus-ingest generate --builtin ecommerce --rows 5000000 --max-file-size 128 --max-rows-per-file 500000
# Schema management workflow
milvus-ingest schema list
milvus-ingest schema show ecommerce
milvus-ingest schema add my_ecommerce ecommerce_base.json
# Clean up generated output files
milvus-ingest clean --yes
๐ Milvus Integration
Direct Insert to Milvus
Insert generated data directly into Milvus with automatic collection creation:
# Generate data first
milvus-ingest generate --builtin ecommerce --rows 100000 --out products/
# Insert to local Milvus (default: localhost:19530)
milvus-ingest to-milvus insert ./products/
# Insert to remote Milvus with authentication
milvus-ingest to-milvus insert ./products/ \
--uri http://192.168.1.100:19530 \
--token your-api-token \
--db-name custom_db
# Insert with custom settings
milvus-ingest to-milvus insert ./products/ \
--collection-name product_catalog \
--batch-size 5000 \
--drop-if-exists
Direct Insert Features:
- โ Automatic collection creation from metadata
- โ Smart index creation based on vector dimensions
- โ Progress tracking with batch processing
- โ Support for authentication and custom databases
- โ Connection testing before import
Bulk Import from S3/MinIO
For very large datasets, use bulk import with pre-uploaded files:
# First, upload to S3/MinIO
milvus-ingest upload ./products/ s3://bucket/data/ \
--endpoint-url http://minio:9000 \
--access-key-id minioadmin \
--secret-access-key minioadmin
# Then bulk import to Milvus
milvus-ingest to-milvus import product_catalog s3://bucket/data/file1.parquet
# Import multiple files
milvus-ingest to-milvus import product_catalog \
s3://bucket/data/file1.parquet \
s3://bucket/data/file2.parquet
# Import all files from directory
milvus-ingest to-milvus import product_catalog ./products/
# Import and wait for completion
milvus-ingest to-milvus import product_catalog ./products/ \
--wait \
--timeout 300
Bulk Import Features:
- โก High-performance import for millions of rows
- ๐ Support for single/multiple files or directories
- โณ Asynchronous operation with job tracking
- ๐ Wait for completion with timeout support
- ๐ Import job status monitoring
S3/MinIO Upload
Upload generated data to S3-compatible storage:
# Upload to AWS S3 (using default credentials)
milvus-ingest upload ./output s3://my-bucket/data/
# Upload to MinIO with custom endpoint
milvus-ingest upload ./output s3://my-bucket/data/ \
--endpoint-url http://localhost:9000 \
--access-key-id minioadmin \
--secret-access-key minioadmin
# Upload with environment variables
export AWS_ACCESS_KEY_ID=your-key
export AWS_SECRET_ACCESS_KEY=your-secret
milvus-ingest upload ./output s3://my-bucket/data/
# Disable SSL verification for local MinIO
milvus-ingest upload ./output s3://my-bucket/data/ \
--endpoint-url http://localhost:9000 \
--no-verify-ssl
Complete Workflow Example
# 1. Generate large dataset
milvus-ingest generate --builtin ecommerce --rows 5000000 --out products/
# 2. Option A: Direct insert (for smaller datasets)
milvus-ingest to-milvus insert ./products/ \
--uri http://milvus:19530 \
--collection-name ecommerce_products
# 2. Option B: Bulk import (for very large datasets)
# First upload to MinIO
milvus-ingest upload ./products/ s3://milvus-data/products/ \
--endpoint-url http://minio:9000
# Then bulk import
milvus-ingest to-milvus import ecommerce_products \
s3://milvus-data/products/ \
--wait
Import Method Comparison
| Method | Best For | Speed | Max Size | Features |
|---|---|---|---|---|
| Direct Insert | <1M rows | Moderate | Limited by memory | Automatic collection creation, progress bar |
| Bulk Import | >1M rows | Very Fast | 16GB per file | Async operation, job tracking |
Important Notes:
- Files must be uploaded to S3/MinIO before bulk import
- Maximum 1024 files per import request
- Each file should not exceed 16GB
- Collection must exist for bulk import (create with direct insert first if needed)
๐ ๏ธ Development
This project uses PDM for dependency management and follows modern Python development practices.
Setup Development Environment
# Clone and setup
git clone https://github.com/your-org/milvus-ingest.git
cd milvus-ingest
pdm install # Install development dependencies
Development Workflow
# Code formatting and linting
pdm run ruff format src tests # Format code
pdm run ruff check src tests # Check linting
pdm run mypy src # Type checking
# Testing
pdm run pytest # Run all tests
pdm run pytest --cov=src --cov-report=html # With coverage
pdm run pytest tests/test_generator.py # Specific test file
# Combined quality checks
make lint test # Run linting and tests together
Project Structure
src/milvus_fake_data/
โโโ cli.py # Click-based CLI interface
โโโ generator.py # Core data generation logic
โโโ optimized_writer.py # High-performance vectorized data generation
โโโ models.py # Pydantic schema validation models
โโโ schema_manager.py # Schema management system
โโโ builtin_schemas.py # Built-in schema definitions and metadata
โโโ rich_display.py # Rich terminal formatting and UI
โโโ logging_config.py # Loguru-based structured logging
โโโ exceptions.py # Custom exception classes
โโโ uploader.py # S3/MinIO upload functionality
โโโ milvus_inserter.py # Direct Milvus insertion
โโโ milvus_importer.py # Bulk import from S3/MinIO
โโโ schemas/ # Built-in schema JSON files (12 schemas)
โโโ simple.json
โโโ ecommerce.json
โโโ documents.json
โโโ images.json
โโโ users.json
โโโ videos.json
โโโ news.json
โโโ audio_transcripts.json
โโโ ai_conversations.json
โโโ face_recognition.json
โโโ ecommerce_partitioned.json
โโโ cardinality_demo.json
๐ Performance Benchmarks
The high-performance engine delivers exceptional speed for large-scale data generation:
| Dataset Size | Time | Throughput | Memory Usage | File Output |
|---|---|---|---|---|
| 100K rows | ~13s | 7,500 rows/sec | <1GB | Single file |
| 1M rows | ~87s | 11,500 rows/sec | <2GB | Single file |
| 2.5M rows | ~217s | 11,500 rows/sec | <3GB | 5 files (auto-partitioned) |
| 10M rows | ~870s | 11,500 rows/sec | <4GB | 10 files (auto-partitioned) |
Key Performance Features:
- Vectorized Operations: NumPy-based batch processing for maximum CPU efficiency
- Smart Memory Management: Streaming generation prevents memory exhaustion
- Automatic File Partitioning: Files split at 256MB/1M rows for optimal handling
- Optimized I/O: Direct PyArrow integration with Snappy compression
- Parallel Processing: Multi-core utilization for vector generation and normalization
Recommended Settings for Large Datasets:
- Use
--format parquetfor fastest I/O (default) - Batch size 50K-100K rows for optimal memory/speed balance
- Enable automatic file partitioning for datasets >1M rows
๐ค Contributing
We welcome contributions! Please follow these steps:
- Fork the repository
- Create a feature branch:
git checkout -b feature/amazing-feature - Make your changes with tests
- Ensure quality checks pass:
make lint test - Commit changes:
git commit -m 'Add amazing feature' - Push to branch:
git push origin feature/amazing-feature - Open a Pull Request
Contribution Guidelines
- Add tests for new functionality
- Update documentation for API changes
- Follow existing code style (ruff + mypy)
- Include helpful error messages for user-facing features
License
This project is licensed under the MIT License - see the LICENSE file for details.
Acknowledgments
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file milvus_ingest-0.1.1.tar.gz.
File metadata
- Download URL: milvus_ingest-0.1.1.tar.gz
- Upload date:
- Size: 94.5 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: pdm/2.22.3 CPython/3.13.4 Darwin/22.6.0
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
5d3e6aec7e02635cc75aa3636ad2cbcbc75fcaf34c001b6965b5a140a3a316ad
|
|
| MD5 |
3e151dd3067782def6f136c58f129e6f
|
|
| BLAKE2b-256 |
e0db9486c3ce29ab58fe236343fb0be81dde59a7ac1149275f9eb03cb68bece2
|
File details
Details for the file milvus_ingest-0.1.1-py3-none-any.whl.
File metadata
- Download URL: milvus_ingest-0.1.1-py3-none-any.whl
- Upload date:
- Size: 96.9 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: pdm/2.22.3 CPython/3.13.4 Darwin/22.6.0
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
5774d0ae0d962e3659ace48138e34f0049d85b6805c76f280a1563a27c02af94
|
|
| MD5 |
310f5ac6db0d49ecfb3285160f8660f4
|
|
| BLAKE2b-256 |
b33d7193ae12f4fe4c8d3a504a4084d8d9aeb8c6d5e2232d57ec1997f0d694a9
|