Neptune Client
Project description
neptune.ai
What is neptune.ai?
Neptune is a lightweight experiment tracker for ML teams that struggle with debugging and reproducing experiments, sharing results, and messy model handover. It offers a single place to track, compare, store, and collaborate on experiments and models.
With Neptune, Data Scientists can develop production-ready models faster, and ML Engineers can access model artifacts instantly in order to deploy them to production.
Watch a 3min explainer video →
Play with a live example project in the Neptune app →
Getting started
Step 1: Create a free account
Step 2: Install the Neptune client library
pip install neptune
Step 3: Add an experiment tracking snippet to your code
import neptune
run = neptune.init_run(project="workspace-name/project-name")
run["parameters"] = {"lr": 0.1, "dropout": 0.4}
run["test_accuracy"] = 0.84
Core features
Log and display
Add a snippet to any step of your ML pipeline once. Decide what and how you want to log. Run a million times.
-
Any framework: any code, fastai, PyTorch, Lightning, TensorFlow/Keras, scikit-learn, 🤗 Transformers, XGBoost, Optuna.
-
Any metadata type: metrics, parameters, dataset and model versions, images, interactive plots, videos, hardware (GPU, CPU, memory), code state.
-
From anywhere in your ML pipeline: multinode pipelines, distributed computing, log during or after execution, log offline, and sync when you are back online.
Organize experiments
Organize logs in a fully customizable nested structure. Display model metadata in user-defined dashboard templates.
-
Nested metadata structure: the flexible API lets you customize the metadata logging structure however you want. Organize nested parameter configs or the results on k-fold validation splits the way they should be.
-
Custom dashboards: combine different metadata types in one view. Define it for one run. Use anywhere. Look at GPU, memory consumption, and load times to debug training speed. See learning curves, image predictions, and confusion matrix to debug model quality.
-
Table views: create different views of the runs table and save them for later. You can have separate table views for debugging, comparing parameter sets, or best experiments.
Compare results
Visualize training live in the neptune.ai web app. See how different parameters and configs affect the results. Optimize models quicker.
-
Compare: learning curves, parameters, images, datasets.
-
Search, sort, and filter: experiments by any field you logged. Use our query language to filter runs based on parameter values, metrics, execution times, or anything else.
-
Visualize and display: runs table, interactive display, folder structure, dashboards.
-
Monitor live: hardware consumption metrics, GPU, CPU, memory.
-
Group by: dataset versions, parameters.
Version models
Version, review, and access production-ready models and metadata associated with them in a single place.
-
Version models: register models, create model versions, version external model artifacts.
-
Review and change stages: look at the validation, test metrics and other model metadata. You can move models between None/Staging/Production/Archived.
-
Access and share models: every model and model version is accessible via the neptune.ai web app or through the API.
Share results
Have a single place where your team can see the results and access all models and experiments.
-
Send a link: share every chart, dashboard, table view, or anything else you see in the neptune.ai app by copying and sending persistent URLs.
-
Query API: access all model metadata via neptune.ai API. Whatever you logged, you can query in a similar way.
-
Manage users and projects: create different projects, add users to them, and grant different permissions levels.
-
Add your entire org: you can collaborate with a team on every plan, even the Free one. So, invite your entire organization, including product managers and subject matter experts, to increase the visibility from the very beginning.
Integrate with any MLOps stack
neptune.ai integrates with 25+ frameworks: PyTorch, Lightning, TensorFlow/Keras, LightGBM, scikit-learn, XGBoost, Optuna, Kedro, 🤗 Transformers, fastai, Prophet, detectron2, Airflow, and more.
PyTorch Lightning
Example:
from pytorch_lightning import Trainer
from lightning.pytorch.loggers import NeptuneLogger
# Create NeptuneLogger instance
from neptune import ANONYMOUS_API_TOKEN
neptune_logger = NeptuneLogger(
api_key=ANONYMOUS_API_TOKEN,
project="common/pytorch-lightning-integration",
tags=["training", "resnet"], # optional
)
# Pass the logger to the Trainer
trainer = Trainer(max_epochs=10, logger=neptune_logger)
# Run the Trainer
trainer.fit(my_model, my_dataloader)
neptune.ai is trusted by great companies
Read how various customers use Neptune to improve their workflow.
Support
If you get stuck or simply want to talk to us about something, here are your options:
- Check our FAQ page.
- Take a look at our resource center.
- Chat! In the app, click the blue message icon in the bottom-right corner and send a message. A real person will talk to you ASAP (typically very ASAP).
- You can just shoot us an email at support@neptune.ai.
People behind
Created with :heart: by the neptune.ai team
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file neptune_client-1.13.0.tar.gz
.
File metadata
- Download URL: neptune_client-1.13.0.tar.gz
- Upload date:
- Size: 276.8 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.0.0 CPython/3.12.7
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | d6699c8f4368b6797cb52cd6e9ace8b5786bd7a8de1015df1e3b78d84058a1be |
|
MD5 | c6f7498feef20b8f42970eee1c1b2c03 |
|
BLAKE2b-256 | 9bec941f1ce1785d7f88ef88cb1e14daeefe0a72479973e446e5ac1e8891f841 |
File details
Details for the file neptune_client-1.13.0-py3-none-any.whl
.
File metadata
- Download URL: neptune_client-1.13.0-py3-none-any.whl
- Upload date:
- Size: 502.7 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.0.0 CPython/3.12.7
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | e048189b18e2c817eefb13208ce2f557553e4e56734b8cbb5645343eeea0460e |
|
MD5 | ee437c2f7b7f7388cd6c2ed6caad2712 |
|
BLAKE2b-256 | 4837630b397c689c335bc020905c534fc757e6b2dfbbd0711139f292a638e0a2 |