Skip to main content

No project description provided

Project description







Plug and play modules to boost the performances of your AI systems

Nebullvm is an ecosystem of plug and play modules to boost the performances of your AI systems. The optimization modules are stack-agnostic and work with any library. They are designed to be easily integrated into your system, providing a quick and seamless boost to its performance. Simply plug and play to start realizing the benefits of optimized performance right away.

If you like the idea, give us a star to show your support for the project ⭐

Documentation

Please find here the full documentation on:

  • Installation
  • Getting started (quick view and examples)
  • Notebooks
  • Ecosystem and integrations
  • Product structure

What can this help with?

There are multiple modules we actually provide to boost the performances of your AI systems:

✅ Speedster: Automatically apply the best set of SOTA optimization techniques to achieve the maximum inference speed-up on your hardware.

✅ OpenAlphaTensor: Increase the computational performances of an AI model with custom-generated matrix multiplication algorithms.

Nos: Automatically maximize the utilization of GPU resources in a Kubernetes cluster through real-time dynamic partitioning and elastic quotas - Effortless optimization at its finest!

✅ Forward-Forward: The Forward Forward algorithm is a method for training deep neural networks that replaces the backpropagation forward and backward passes with two forward passes.

Next modules and roadmap

We are actively working on incorporating the following modules, as requested by members of our community, in upcoming releases:

  • Promptify: Effortlessly personalize large APIs generative models from OpenAI, Cohere, HF to your specific context and requirements.
  • CloudSurfer: Automatically discover the optimal cloud configuration and hardware on AWS, GCP and Azure to run your AI models.
  • OptiMate: Interactive tool guiding savvy users in achieving the best inference performance out of a given model / hardware setup.
  • TrainingSim: Easily simulate the training of large AI models on a distributed infrastructure to predict training behaviours without actual implementation.

Contributing

As an open source project in a rapidly evolving field, we welcome contributions of all kinds, including new features, improved infrastructure, and better documentation. If you're interested in contributing, please see the linked page for more information on how to get involved.


Join the community | Contribute to the library

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

nebullvm-0.7.1.tar.gz (90.9 kB view hashes)

Uploaded Source

Built Distribution

nebullvm-0.7.1-py3-none-any.whl (157.7 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page