ML evaluation, validation, and test case generation toolkit with production monitoring.
Project description
ML Robust Eval
ML Eval Robust is a pure Python, object-oriented library for comprehensive machine learning model evaluation, validation, and robustness testing.
It's an all-in-one toolkit that features:
- 📊 Metrics for classification, regression, NLP, and computer vision tasks
- 🔁 Cross-validation and A/B testing helpers
- 📈 Visualization tools for confusion matrices and ROC curves (stdout-based, no dependencies!)
- 🦾 Automated test case generation: edge cases, adversarial samples, and boundary value tests
- 📡 Production Monitoring (optional): Real-time model degradation tracking with Prometheus & Grafana
- 🧩 Zero dependencies (core) – works anywhere Python runs!
🚀 Installation
Basic installation:
pip install ml_robust_eval
With production monitoring support (Prometheus/Grafana):
pip install ml_robust_eval[monitoring]
🧠 Features
- Classification, Regression, NLP, and CV Metrics
- Accuracy, Precision, Recall, F1, MAE, MSE, R², BLEU, IoU, and more!
- Cross-Validation & A/B Testing
- K-fold splitting, group comparison, and statistical difference calculation
- Visualization
- Confusion matrices and ROC curves printed directly to your console
- Robustness Test Case Generation
- Edge, boundary, and adversarial sample generation for any tabular data
- Production Monitoring (Optional)
- Real-time model performance tracking with Prometheus
- Automatic degradation detection compared to baseline
- Pre-configured Grafana dashboards
- Docker Compose setup for easy deployment
- Zero Dependencies (Core)
- Core library uses only standard library, OOP-based, and lightweight
- Monitoring module is optional and requires prometheus-client
📚 Documentation
- API Reference
- Examples & Tutorials
- Production Monitoring Guide - Real-time model degradation tracking
💡 Why ML Eval Robust?
- Universal: No dependencies, works in any Python environment
- Educational: Clear, readable OOP code for learning and teaching
- Robust: Covers the full ML evaluation and validation pipeline, including adversarial and edge testing
- Production-Ready: Optional monitoring module for real-time model tracking in production
🤝 Contributing
All contributions, bug reports, and suggestions are welcome!
See the contributing guide.
📜 License
📬 Contact
Questions? Open an issue or reach out at [vikhyathchoppa699@gmail.com].
Let your models earn their confidence. Test, validate, and challenge them with ML Robust Eval!
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file ml_robust_eval-0.1.7.tar.gz.
File metadata
- Download URL: ml_robust_eval-0.1.7.tar.gz
- Upload date:
- Size: 9.6 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/2.2.1 CPython/3.12.3 Linux/6.14.0-33-generic
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
978241135ab44dfdfc5b7dc8412a1ef0428f5151c9cc155a301a4f765af6b3ce
|
|
| MD5 |
94f046f356e3ac35e96314a0e7964979
|
|
| BLAKE2b-256 |
2666f672c2a8e282a68103782227cb34bfe4b0e171a5d44143a4810047233d35
|
File details
Details for the file ml_robust_eval-0.1.7-py3-none-any.whl.
File metadata
- Download URL: ml_robust_eval-0.1.7-py3-none-any.whl
- Upload date:
- Size: 10.2 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/2.2.1 CPython/3.12.3 Linux/6.14.0-33-generic
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
b8888035132d3ea988d077df3c0de7b3347938edcfaba3e908b2fd21f56c014b
|
|
| MD5 |
1108f578379669b7dcb43fc0f136db86
|
|
| BLAKE2b-256 |
9f8a97a5a119dd6bf76771d0c3cff0c3e38da4faf19a931e008a9ddaf76659b5
|