Skip to content

Model Training

Ultralytics Platform provides comprehensive tools for training YOLO models, from organizing experiments to running cloud training jobs with real-time metrics streaming.

Overview

The Training section helps you:

  • Organize models into projects for easier management
  • Train on cloud GPUs with a single click
  • Monitor real-time metrics during training
  • Compare model performance across experiments
  • Export to 17+ deployment formats (see supported formats)

Ultralytics Platform Train Overview

Workflow

graph LR
    A[📁 Project] --> B[⚙️ Configure]
    B --> C[🚀 Train]
    C --> D[📈 Monitor]
    D --> E[📦 Export]

    style A fill:#4CAF50,color:#fff
    style B fill:#2196F3,color:#fff
    style C fill:#FF9800,color:#fff
    style D fill:#9C27B0,color:#fff
    style E fill:#00BCD4,color:#fff
StageDescription
ProjectCreate a workspace to organize related models
ConfigureSelect dataset, base model, and training parameters
TrainRun on cloud GPUs or your local hardware
MonitorView real-time loss curves and metrics
ExportConvert to 17+ deployment formats (details)

Training Options

Ultralytics Platform supports multiple training approaches:

MethodDescriptionBest For
Cloud TrainingTrain on Ultralytics Cloud GPUsNo local GPU, scalability
Local TrainingTrain locally, stream metrics to the platformExisting hardware, privacy
Colab TrainingUse Google Colab with platform integrationFree GPU access

GPU Options

Available GPUs for cloud training on Ultralytics Cloud:

GPUVRAMCost/HourBest For
RTX 2000 Ada16 GB$0.24Small datasets, testing
RTX A450020 GB$0.24Small-medium datasets
RTX A500024 GB$0.26Medium datasets
RTX 4000 Ada20 GB$0.38Medium datasets
L424 GB$0.39Inference optimized
A4048 GB$0.40Larger batch sizes
RTX 309024 GB$0.46Great price/performance
RTX A600048 GB$0.49Large models
RTX 409024 GB$0.59Best price/performance
RTX 6000 Ada48 GB$0.77Large batch training
L40S48 GB$0.86Large batch training
RTX 509032 GB$0.89Latest generation
L4048 GB$0.99Large models
A100 PCIe80 GB$1.39Production training
A100 SXM80 GB$1.49Production training
RTX PRO 600096 GB$1.89Recommended default
H100 PCIe80 GB$2.39High-performance training
H100 SXM80 GB$2.69Fastest training
H100 NVL94 GB$3.07Maximum performance
H200 NVL143 GB$3.39Maximum memory
H200 SXM141 GB$3.59Maximum performance
B200180 GB$4.99Largest models

Signup Credits

New accounts receive signup credits for training. Check Billing for details.

Real-Time Metrics

During training, view live metrics across three subtabs:

graph LR
    A[Charts] --> B[Loss Curves]
    A --> C[Performance Metrics]
    D[Console] --> E[Live Logs]
    D --> F[Error Detection]
    G[System] --> H[GPU Utilization]
    G --> I[Memory & Temp]

    style A fill:#2196F3,color:#fff
    style D fill:#FF9800,color:#fff
    style G fill:#9C27B0,color:#fff
SubtabMetrics
ChartsBox/class/DFL loss, mAP50, mAP50-95, precision, recall
ConsoleLive training logs with ANSI color and error detection
SystemGPU utilization, memory, temperature, CPU, disk

Automatic Checkpoints

The Platform automatically saves checkpoints at every epoch. The best model (highest mAP) and final model are always preserved.

Quick Start

Get started with cloud training in under a minute:

  1. Create a project in the sidebar
  2. Click New Model
  3. Select a model, dataset, and GPU
  4. Click Start Training
export ULTRALYTICS_API_KEY="your_api_key"
yolo train model=yolo26n.pt data=ul://username/datasets/my-dataset \
  epochs=100 project=username/my-project name=exp1
from ultralytics import YOLO

model = YOLO("yolo26n.pt")
model.train(
    data="ul://username/datasets/my-dataset",
    epochs=100,
    project="username/my-project",
    name="exp1",
)

FAQ

How long does training take?

Training time depends on:

  • Dataset size (number of images)
  • Model size (n, s, m, l, x)
  • Number of epochs
  • GPU type selected

A typical training run with 1000 images, YOLO26n, 100 epochs on RTX PRO 6000 takes about 2-3 hours. Smaller runs (500 images, 50 epochs on RTX 4090) complete in under an hour. See cost examples for detailed estimates.

Can I train multiple models simultaneously?

Yes. Concurrent cloud training limits depend on your plan: Free allows 3, Pro allows 10, and Enterprise is unlimited. For additional parallel training, use remote training from multiple machines.

What happens if training fails?

If training fails:

  1. Checkpoints are saved at each epoch
  2. You can resume from the last checkpoint
  3. Credits are only charged for completed compute time

How do I choose the right GPU?

ScenarioRecommended GPU
Most training jobsRTX PRO 6000
Large datasets or batch sizesH100 SXM or H200
Budget-consciousRTX 4090


📅 Created 1 month ago ✏️ Updated 6 days ago
glenn-jochersergiuwaxmann

Comments