Ultralytics YOLO: Object detection and computer vision models
PyTorch library for YOLO-based real-time computer vision.
Learn more about ultralytics
Ultralytics YOLO is a PyTorch-based computer vision library that implements successive versions of the YOLO (You Only Look Once) object detection architecture. The codebase provides model definitions, training pipelines, inference engines, and utilities for tasks including object detection, instance segmentation, image classification, pose estimation, and multi-object tracking. Models are distributed through the Ultralytics Hub and can be deployed via command-line interface or Python API. The library supports various hardware configurations and includes integration with popular deployment platforms.
Unified Multi-Task Interface
Single codebase handles detection, segmentation, classification, pose estimation, and tracking through consistent model APIs. Eliminates the need for separate specialized implementations or framework switching across vision tasks.
Versioned Model Lineage
Multiple YOLO versions (v8, v10, v11) with documented architectural differences and benchmarked performance characteristics. Enables explicit accuracy-latency trade-offs based on deployment constraints rather than guessing optimal models.
CLI and Python API
Offers both command-line interface for quick experiments and a comprehensive Python API for integration. Train, validate, and deploy models using simple commands or programmatic workflows with identical capabilities.
from ultralytics import YOLO
# Load a pre-trained model
model = YOLO('yolo11n.pt')
# Run inference on an image
results = model('path/to/image.jpg')
# Display results
results[0].show()v8.4.7 - `ultralytics 8.4.7` AVIF training and new COCO12-Formats dataset (#23358)
- –AVIF is now a supported training image format** ✅📷
- –`IMG_FORMATS` expanded to include `avif`
- –More robust image decoding for modern formats** 🛡️
- –Adds a Pillow-based fallback image reader (`imreadpil`) for cases where OpenCV can’t decode AVIF/HEIC
- –New “COCO12-Formats” dataset + generator script** 🧰
v8.4.6 - `ultralytics 8.4.6` Add missing PosixPath import in DDP train file gen (#23301)
- –✅ Fixed DDP multi-GPU training crash (PR #23301 by @pfabreu)**
- –Added missing `PosixPath` import to the generated temporary DDP training script (`ultralytics/utils/dist.py`).
- –Prevents `NameError: name 'PosixPath' is not defined` when model paths are passed as `PosixPath` objects (common on Linux) during Distributed Data Parallel training 🧩🧯
- –⏱️ More robust dataset URI resolution for large Ultralytics HUB/Platform datasets (PR #23305 by @glenn-jocher)**
- –`ul://...` dataset URLs now allow a much longer server-side preparation time (NDJSON generation) to avoid premature timeouts 🐢📦
v8.4.5 - `ultralytics 8.4.5` 2D Pose `Result.summary()` support (#23293)
- –🔖 Version bump: `8.4.4` → `8.4.5`.
- –📚 Docs & examples shift to YOLO26: Kaggle links, Ultralytics Platform docs, and multiple notebooks now point to YOLO26 as the recommended current model family 🚀.
- –🗂️ NDJSON dataset docs improved: clearer per-task examples/tabs for Detect/Segment/Pose/OBB/Classify, including pose visibility explanations 🧾👀.
- –🧹 Dependency cleanup: removed optional `hub-sdk` extra from `pyproject.toml` (simplifies installs for users who don’t need it) 📦✂️.
- –🔄 More robust downstream integrations: safer summaries improve reliability for APIs, analytics, logging, and dataset export workflows that depend on consistent serialization.
See how people are using ultralytics
Top in AI & ML
Related Repositories
Discover similar tools and frameworks used by developers
DALL-E
Official PyTorch package implementing the discrete VAE component for image tokenization used in OpenAI's DALL-E system.
Magenta
Google Brain research project using ML to generate music, images, and creative content with TensorFlow.
higgsfield
Cluster manager for multi-node PyTorch model training.
gym
Standard API for reinforcement learning environment interfaces.
OpenAI.fm
Web demo showcasing OpenAI's Speech API text-to-speech capabilities with an interactive Next.js interface.