Skip to main content

Ultralytics YOLO 🚀 for SOTA object detection, multi-object tracking, instance segmentation, pose estimation and image classification.

Project description


Ultralytics creates cutting-edge, state-of-the-art (SOTA) YOLO models built on years of foundational research in computer vision and AI. Constantly updated for performance and flexibility, our models are fast, accurate, and easy to use. They excel at object detection, tracking, instance segmentation, image classification, and pose estimation tasks.

Find detailed documentation in the Ultralytics Docs. Get support via GitHub Issues. Join discussions on Discord, Reddit, and the Ultralytics Community Forums!

Request an Enterprise License for commercial use at Ultralytics Licensing.

YOLO11 performance plots
Ultralytics GitHub space Ultralytics LinkedIn space Ultralytics Twitter space Ultralytics YouTube space Ultralytics TikTok space Ultralytics BiliBili space Ultralytics Discord

📄 Documentation

See below for quickstart installation and usage examples. For comprehensive guidance on training, validation, prediction, and deployment, refer to our full Ultralytics Docs.

Install

Install the ultralytics package, including all requirements, in a Python>=3.8 environment with PyTorch>=1.8.

PyPI - Version Ultralytics Downloads PyPI - Python Version

pip install ultralytics

For alternative installation methods, including Conda, Docker, and building from source via Git, please consult the Quickstart Guide.

Conda Version Docker Image Version Ultralytics Docker Pulls

Usage

CLI

You can use Ultralytics YOLO directly from the Command Line Interface (CLI) with the yolo command:

# Predict using a pretrained YOLO model (e.g., YOLO26n) on an image
yolo predict model=yolo26n.pt source='https://ultralytics.com/images/bus.jpg'

The yolo command supports various tasks and modes, accepting additional arguments like imgsz=640. Explore the YOLO CLI Docs for more examples.

Python

Ultralytics YOLO can also be integrated directly into your Python projects. It accepts the same configuration arguments as the CLI:

from ultralytics import YOLO

# Load a pretrained YOLO26n model
model = YOLO("yolo26n.pt")

# Train the model on the COCO8 dataset for 100 epochs
train_results = model.train(
    data="coco8.yaml",  # Path to dataset configuration file
    epochs=100,  # Number of training epochs
    imgsz=640,  # Image size for training
    device="cpu",  # Device to run on (e.g., 'cpu', 0, [0,1,2,3])
)

# Evaluate the model's performance on the validation set
metrics = model.val()

# Perform object detection on an image
results = model("path/to/image.jpg")  # Predict on an image
results[0].show()  # Display results

# Export the model to ONNX format for deployment
path = model.export(format="onnx")  # Returns the path to the exported model

Discover more examples in the YOLO Python Docs.

✨ Models

Ultralytics supports a wide range of YOLO models, from early versions like YOLOv3 to the latest YOLO26. The tables below showcase YOLO26 models pretrained on the COCO dataset for Detection, Segmentation, and Pose Estimation. Additionally, Classification models pretrained on the ImageNet dataset are available. Tracking mode is compatible with all Detection, Segmentation, and Pose models. All Models are automatically downloaded from the latest Ultralytics release upon first use.

Ultralytics YOLO supported tasks

Detection (COCO)

Explore the Detection Docs for usage examples. These models are trained on the COCO dataset, featuring 80 object classes.

Model size
(pixels)
mAPval
50-95
Speed
CPU ONNX
(ms)
Speed
T4 TensorRT10
(ms)
params
(M)
FLOPs
(B)
YOLO26n 640 40.9 38.9 ± 0.7 1.7 ± 0.0 2.4 5.4
YOLO26s 640 48.6 87.2 ± 0.9 2.5 ± 0.0 9.5 20.7
YOLO26m 640 53.1 220.0 ± 1.4 4.7 ± 0.1 20.4 68.2
YOLO26l 640 55.0 286.2 ± 2.0 6.2 ± 0.2 24.8 86.4
YOLO26x 640 57.5 525.8 ± 4.0 11.8 ± 0.2 55.7 193.9
  • mAPval values refer to single-model single-scale performance on the COCO val2017 dataset. See YOLO Performance Metrics for details.
    Reproduce with yolo val detect data=coco.yaml device=0
  • Speed metrics are averaged over COCO val images using an Amazon EC2 P4d instance. CPU speeds measured with ONNX export. GPU speeds measured with TensorRT export.
    Reproduce with yolo val detect data=coco.yaml batch=1 device=0|cpu
Segmentation (COCO)

Refer to the Segmentation Docs for usage examples. These models are trained on COCO-Seg, including 80 classes.

Model size
(pixels)
mAPbox
50-95
mAPmask
50-95
Speed
CPU ONNX
(ms)
Speed
T4 TensorRT10
(ms)
params
(M)
FLOPs
(B)
YOLO26n-seg 640 39.6 33.9 53.3 ± 0.5 2.1 ± 0.0 2.7 9.1
YOLO26s-seg 640 47.3 40.0 118.4 ± 0.9 3.3 ± 0.0 10.4 34.2
YOLO26m-seg 640 52.5 44.1 328.2 ± 2.4 6.7 ± 0.1 23.6 121.5
YOLO26l-seg 640 54.4 45.5 387.0 ± 3.7 8.0 ± 0.1 28.0 139.8
YOLO26x-seg 640 56.5 47.0 787.0 ± 6.8 16.4 ± 0.1 62.8 313.5
  • mAPval values are for single-model single-scale on the COCO val2017 dataset. See YOLO Performance Metrics for details.
    Reproduce with yolo val segment data=coco.yaml device=0
  • Speed metrics are averaged over COCO val images using an Amazon EC2 P4d instance. CPU speeds measured with ONNX export. GPU speeds measured with TensorRT export.
    Reproduce with yolo val segment data=coco.yaml batch=1 device=0|cpu
Classification (ImageNet)

Consult the Classification Docs for usage examples. These models are trained on ImageNet, covering 1000 classes.

Model size
(pixels)
acc
top1
acc
top5
Speed
CPU ONNX
(ms)
Speed
T4 TensorRT10
(ms)
params
(M)
FLOPs
(B) at 224
YOLO26n-cls 224 71.4 90.1 5.0 ± 0.3 1.1 ± 0.0 2.8 0.5
YOLO26s-cls 224 76.0 92.9 7.9 ± 0.2 1.3 ± 0.0 6.7 1.6
YOLO26m-cls 224 78.1 94.2 17.2 ± 0.4 2.0 ± 0.0 11.6 4.9
YOLO26l-cls 224 79.0 94.6 23.2 ± 0.3 2.8 ± 0.0 14.1 6.2
YOLO26x-cls 224 79.9 95.0 41.4 ± 0.9 3.8 ± 0.0 29.6 13.6
  • acc values represent model accuracy on the ImageNet dataset validation set.
    Reproduce with yolo val classify data=path/to/ImageNet device=0
  • Speed metrics are averaged over ImageNet val images using an Amazon EC2 P4d instance. CPU speeds measured with ONNX export. GPU speeds measured with TensorRT export.
    Reproduce with yolo val classify data=path/to/ImageNet batch=1 device=0|cpu
Pose (COCO)

See the Pose Estimation Docs for usage examples. These models are trained on COCO-Pose, focusing on the 'person' class.

Model size
(pixels)
mAPpose
50-95
mAPpose
50
Speed
CPU ONNX
(ms)
Speed
T4 TensorRT10
(ms)
params
(M)
FLOPs
(B)
YOLO26n-pose 640 57.2 83.3 40.3 ± 0.5 1.8 ± 0.0 2.9 7.5
YOLO26s-pose 640 63.0 86.6 85.3 ± 0.9 2.7 ± 0.0 10.4 23.9
YOLO26m-pose 640 68.8 89.6 218.0 ± 1.5 5.0 ± 0.1 21.5 73.1
YOLO26l-pose 640 70.4 90.5 275.4 ± 2.4 6.5 ± 0.1 25.9 91.3
YOLO26x-pose 640 71.6 91.6 565.4 ± 3.0 12.2 ± 0.2 57.6 201.7
  • mAPval values are for single-model single-scale on the COCO Keypoints val2017 dataset. See YOLO Performance Metrics for details.
    Reproduce with yolo val pose data=coco-pose.yaml device=0
  • Speed metrics are averaged over COCO val images using an Amazon EC2 P4d instance. CPU speeds measured with ONNX export. GPU speeds measured with TensorRT export.
    Reproduce with yolo val pose data=coco-pose.yaml batch=1 device=0|cpu
Oriented Bounding Boxes (DOTAv1)

Check the OBB Docs for usage examples. These models are trained on DOTAv1, including 15 classes.

Model size
(pixels)
mAPtest
50
Speed
CPU ONNX
(ms)
Speed
T4 TensorRT10
(ms)
params
(M)
FLOPs
(B)
YOLO26n-obb 1024 78.9 97.7 ± 0.9 2.8 ± 0.0 2.5 14.0
YOLO26s-obb 1024 80.9 218.0 ± 1.4 4.9 ± 0.1 9.8 55.1
YOLO26m-obb 1024 81.0 579.2 ± 3.8 10.2 ± 0.3 21.2 183.3
YOLO26l-obb 1024 81.6 735.6 ± 3.1 13.0 ± 0.2 25.6 230.0
YOLO26x-obb 1024 81.7 1485.7 ± 11.5 30.5 ± 0.9 57.6 516.5
  • mAPtest values are for single-model multiscale performance on the DOTAv1 test set.
    Reproduce by yolo val obb data=DOTAv1.yaml device=0 split=test and submit merged results to the DOTA evaluation server.
  • Speed metrics are averaged over DOTAv1 val images using an Amazon EC2 P4d instance. CPU speeds measured with ONNX export. GPU speeds measured with TensorRT export.
    Reproduce by yolo val obb data=DOTAv1.yaml batch=1 device=0|cpu

🧩 Integrations

Our key integrations with leading AI platforms extend the functionality of Ultralytics' offerings, enhancing tasks like dataset labeling, training, visualization, and model management. Discover how Ultralytics, in collaboration with partners like Weights & Biases, Comet ML, Roboflow, and Intel OpenVINO, can optimize your AI workflow. Explore more at Ultralytics Integrations.

Ultralytics active learning integrations

Ultralytics HUB 🌟 Weights & Biases Comet Neural Magic
Streamline YOLO workflows: Label, train, and deploy effortlessly with Ultralytics HUB. Try now! Track experiments, hyperparameters, and results with Weights & Biases. Free forever, Comet ML lets you save YOLO models, resume training, and interactively visualize predictions. Run YOLO inference up to 6x faster with Neural Magic DeepSparse.

🤝 Contribute

We thrive on community collaboration! Ultralytics YOLO wouldn't be the SOTA framework it is without contributions from developers like you. Please see our Contributing Guide to get started. We also welcome your feedback—share your experience by completing our Survey. A huge Thank You 🙏 to everyone who contributes!

Ultralytics open-source contributors

We look forward to your contributions to help make the Ultralytics ecosystem even better!

📜 License

Ultralytics offers two licensing options to suit different needs:

  • AGPL-3.0 License: This OSI-approved open-source license is perfect for students, researchers, and enthusiasts. It encourages open collaboration and knowledge sharing. See the LICENSE file for full details.
  • Ultralytics Enterprise License: Designed for commercial use, this license allows for the seamless integration of Ultralytics software and AI models into commercial products and services, bypassing the open-source requirements of AGPL-3.0. If your use case involves commercial deployment, please contact us via Ultralytics Licensing.

📞 Contact

For bug reports and feature requests related to Ultralytics software, please visit GitHub Issues. For questions, discussions, and community support, join our active communities on Discord, Reddit, and the Ultralytics Community Forums. We're here to help with all things Ultralytics!


Ultralytics GitHub space Ultralytics LinkedIn space Ultralytics Twitter space Ultralytics YouTube space Ultralytics TikTok space Ultralytics BiliBili space Ultralytics Discord

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

ultralytics-8.4.2.tar.gz (1.0 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

ultralytics-8.4.2-py3-none-any.whl (1.2 MB view details)

Uploaded Python 3

File details

Details for the file ultralytics-8.4.2.tar.gz.

File metadata

  • Download URL: ultralytics-8.4.2.tar.gz
  • Upload date:
  • Size: 1.0 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for ultralytics-8.4.2.tar.gz
Algorithm Hash digest
SHA256 4b4b4b32409e393b1285270411e27cd9dc0141f595f490536a130482501b1f2f
MD5 e91444327ef3f66fa2a74246f8e71fdf
BLAKE2b-256 1e85e61f22ec01ef982332e206e0d23cf88667969ac707fdbc7228a23243c8ba

See more details on using hashes here.

Provenance

The following attestation bundles were made for ultralytics-8.4.2.tar.gz:

Publisher: publish.yml on ultralytics/ultralytics

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file ultralytics-8.4.2-py3-none-any.whl.

File metadata

  • Download URL: ultralytics-8.4.2-py3-none-any.whl
  • Upload date:
  • Size: 1.2 MB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for ultralytics-8.4.2-py3-none-any.whl
Algorithm Hash digest
SHA256 50ef362f4ba45e2f6681428688e898ebd9a8c4a40e08f82491ad8da9aab26644
MD5 815654fb4e59e27c285cf02da34ac0fd
BLAKE2b-256 d58a67a93d087b3d5ee0ae5ff1d817a862d2fb211a419c19d4ed68b45ab43d52

See more details on using hashes here.

Provenance

The following attestation bundles were made for ultralytics-8.4.2-py3-none-any.whl:

Publisher: publish.yml on ultralytics/ultralytics

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page