Skip to content

Computer Vision Tasks supported by Ultralytics YOLO11

Ultralytics YOLO supported computer vision tasks

Ultralytics YOLO11 is a versatile AI framework that supports multiple computer vision tasks. The framework can be used to perform detection, segmentation, obb, classification, and pose estimation. Each of these tasks has a different objective and use case, allowing you to address various computer vision challenges with a single framework.



Watch: Explore Ultralytics YOLO Tasks: Object Detection, Segmentation, OBB, Tracking, and Pose Estimation.

Detection

Detection is the primary task supported by YOLO11. It involves identifying objects in an image or video frame and drawing bounding boxes around them. The detected objects are classified into different categories based on their features. YOLO11 can detect multiple objects in a single image or video frame with high accuracy and speed, making it ideal for real-time applications like surveillance systems and autonomous vehicles.

Detection Examples

Image segmentation

Segmentation takes object detection further by segmenting an image into different regions based on content. Each region is assigned a label, providing pixel-level precision for applications such as medical imaging, agricultural analysis, and manufacturing quality control. YOLO11 implements a variant of the U-Net architecture to perform efficient and accurate segmentation.

Segmentation Examples

Classification

Classification involves categorizing entire images based on their content. YOLO11's classification capabilities leverage a variant of the EfficientNet architecture to deliver high-performance image classification. This task is essential for applications like product categorization in e-commerce, content moderation, and wildlife monitoring.

Classification Examples

Pose estimation

Pose estimation detects specific keypoints in images or video frames to track movements or estimate poses. These keypoints can represent human joints, facial features, or other significant points of interest. YOLO11 excels at keypoint detection with high accuracy and speed, making it valuable for fitness applications, sports analytics, and human-computer interaction.

Pose Examples

OBB

Oriented Bounding Box (OBB) detection enhances traditional object detection by adding an orientation angle to better locate rotated objects. This capability is particularly valuable for aerial imagery analysis, document processing, and industrial applications where objects appear at various angles. YOLO11 delivers high accuracy and speed for detecting rotated objects in diverse scenarios.

Oriented Detection

Conclusion

Ultralytics YOLO11 supports multiple computer vision tasks, including detection, segmentation, classification, oriented object detection, and keypoint detection. Each task addresses specific needs in the computer vision landscape, from basic object identification to detailed pose analysis. By understanding the capabilities and applications of each task, you can select the most appropriate approach for your specific computer vision challenges and leverage YOLO11's powerful features to build effective solutions.

FAQ

What computer vision tasks can Ultralytics YOLO11 perform?

Ultralytics YOLO11 is a versatile AI framework capable of performing various computer vision tasks with high accuracy and speed. These tasks include:

  • Object Detection: Identifying and localizing objects in images or video frames by drawing bounding boxes around them.
  • Image segmentation: Segmenting images into different regions based on their content, useful for applications like medical imaging.
  • Classification: Categorizing entire images based on their content, leveraging variants of the EfficientNet architecture.
  • Pose estimation: Detecting specific keypoints in an image or video frame to track movements or poses.
  • Oriented Object Detection (OBB): Detecting rotated objects with an added orientation angle for enhanced accuracy.

How do I use Ultralytics YOLO11 for object detection?

To use Ultralytics YOLO11 for object detection, follow these steps:

  1. Prepare your dataset in the appropriate format.
  2. Train the YOLO11 model using the detection task.
  3. Use the model to make predictions by feeding in new images or video frames.

Example

from ultralytics import YOLO

# Load a pre-trained YOLO model (adjust model type as needed)
model = YOLO("yolo11n.pt")  # n, s, m, l, x versions available

# Perform object detection on an image
results = model.predict(source="image.jpg")  # Can also use video, directory, URL, etc.

# Display the results
results[0].show()  # Show the first image results
# Run YOLO detection from the command line
yolo detect model=yolo11n.pt source="image.jpg"  # Adjust model and source as needed

For more detailed instructions, check out our detection examples.

What are the benefits of using YOLO11 for segmentation tasks?

Using YOLO11 for segmentation tasks provides several advantages:

  1. High Accuracy: The segmentation task leverages a variant of the U-Net architecture to achieve precise segmentation.
  2. Speed: YOLO11 is optimized for real-time applications, offering quick processing even for high-resolution images.
  3. Multiple Applications: It is ideal for medical imaging, autonomous driving, and other applications requiring detailed image segmentation.

Learn more about the benefits and use cases of YOLO11 for segmentation in the image segmentation section.

Can Ultralytics YOLO11 handle pose estimation and keypoint detection?

Yes, Ultralytics YOLO11 can effectively perform pose estimation and keypoint detection with high accuracy and speed. This feature is particularly useful for tracking movements in sports analytics, healthcare, and human-computer interaction applications. YOLO11 detects keypoints in an image or video frame, allowing for precise pose estimation.

For more details and implementation tips, visit our pose estimation examples.

Why should I choose Ultralytics YOLO11 for oriented object detection (OBB)?

Oriented Object Detection (OBB) with YOLO11 provides enhanced precision by detecting objects with an additional angle parameter. This feature is beneficial for applications requiring accurate localization of rotated objects, such as aerial imagery analysis and warehouse automation.

  • Increased Precision: The angle component reduces false positives for rotated objects.
  • Versatile Applications: Useful for tasks in geospatial analysis, robotics, etc.

Check out the Oriented Object Detection section for more details and examples.

📅 Created 1 year ago ✏️ Updated 1 day ago

Comments