Skip to content

Model YAML Configuration

The model YAML configuration file serves as the architectural blueprint for Ultralytics neural networks. It defines how layers connect, what parameters each module uses, and how the entire network scales across different model sizes.

Configuration Structure

Model YAML files are organized into three main sections that work together to define the architecture.

Parameters Section

The parameters section specifies the model's global characteristics and scaling behavior:

# Parameters
nc: 80 # number of classes
scales: # compound scaling constants [depth, width, max_channels]
    n: [0.50, 0.25, 1024] # nano: shallow layers, narrow channels
    s: [0.50, 0.50, 1024] # small: shallow depth, standard width
    m: [0.50, 1.00, 512] # medium: moderate depth, full width
    l: [1.00, 1.00, 512] # large: full depth and width
    x: [1.00, 1.50, 512] # extra-large: maximum performance
kpt_shape: [17, 3] # pose models only
  • nc sets the number of classes the model predicts.
  • scales define compound scaling factors that adjust model depth, width, and maximum channels to produce different size variants (nano through extra-large).
  • kpt_shape applies to pose models. It can be [N, 2] for (x, y) keypoints or [N, 3] for (x, y, visibility).

Reduce redundancy with scales

The scales parameter lets you generate multiple model sizes from a single base YAML. For instance, when you load yolo11n.yaml, Ultralytics reads the base yolo11.yaml and applies the n scaling factors (depth=0.50, width=0.25) to build the nano variant.

nc and kpt_shape are dataset-dependent

If your dataset specifies a different nc or kpt_shape, Ultralytics will automatically override the model config at runtime to match the dataset YAML.

Backbone and Head Architecture

The model architecture consists of backbone (feature extraction) and head (task-specific) sections:

backbone:
    # [from, repeats, module, args]
    - [-1, 1, Conv, [64, 3, 2]] # 0: Initial convolution
    - [-1, 1, Conv, [128, 3, 2]] # 1: Downsample
    - [-1, 3, C2f, [128, True]] # 2: Feature processing

head:
    - [-1, 1, nn.Upsample, [None, 2, nearest]] # 6: Upsample
    - [[-1, 2], 1, Concat, [1]] # 7: Skip connection
    - [-1, 3, C2f, [256]] # 8: Process features
    - [[8], 1, Detect, [nc]] # 9: Detection layer

Layer Specification Format

Every layer follows the consistent pattern: [from, repeats, module, args]

Component Purpose Examples
from Input connections -1 (previous), 6 (layer 6), [4, 6, 8] (multi-input)
repeats Number of repetitions 1 (single), 3 (repeat 3 times)
module Module type Conv, C2f, TorchVision, Detect
args Module arguments [64, 3, 2] (channels, kernel, stride)

Connection Patterns

The from field creates flexible data flow patterns throughout your network:

- [-1, 1, Conv, [64, 3, 2]]    # Takes input from previous layer
- [[-1, 6], 1, Concat, [1]]    # Combines current layer with layer 6
- [[4, 6, 8], 1, Detect, [nc]] # Detection head using 3 feature scales

Layer Indexing

Layers are indexed starting from 0. Negative indices reference previous layers (-1 = previous layer), while positive indices reference specific layers by their position.

Module Repetition

The repeats parameter creates deeper network sections:

- [-1, 3, C2f, [128, True]] # Creates 3 consecutive C2f blocks
- [-1, 1, Conv, [64, 3, 2]] # Single convolution layer

The actual repetition count gets multiplied by the depth scaling factor from your model size configuration.

Available Modules

Modules are organized by functionality and defined in the Ultralytics modules directory. The following tables show commonly used modules by category, with many more available in the source code:

Basic Operations

Module Purpose Source Arguments
Conv Convolution + BatchNorm + Activation conv.py [out_ch, kernel, stride, pad, groups]
nn.Upsample Spatial upsampling PyTorch [size, scale_factor, mode]
nn.Identity Pass-through operation PyTorch []

Composite Blocks

Module Purpose Source Arguments
C2f CSP bottleneck with 2 convolutions block.py [out_ch, shortcut, expansion]
SPPF Spatial Pyramid Pooling (fast) block.py [out_ch, kernel_size]
Concat Channel-wise concatenation conv.py [dimension]

Specialized Modules

Module Purpose Source Arguments
TorchVision Load any torchvision model block.py [out_ch, model_name, weights, unwrap, truncate, split]
Index Extract specific tensor from list block.py [out_ch, index]
Detect YOLO detection head head.py [nc, anchors, ch]

Complete Module List

This represents a subset of available modules. For the full list of modules and their parameters, explore the modules directory.

Advanced Features

TorchVision Integration

The TorchVision module enables seamless integration of any TorchVision model as a backbone:

from ultralytics import YOLO

# Model with ConvNeXt backbone
model = YOLO("convnext_backbone.yaml")
results = model.train(data="coco8.yaml", epochs=100)
backbone:
  - [-1, 1, TorchVision, [768, convnext_tiny, DEFAULT, True, 2, False]]
head:
  - [-1, 1, Classify, [nc]]

Parameter Breakdown:

  • 768: Expected output channels
  • convnext_tiny: Model architecture (available models)
  • DEFAULT: Use pretrained weights
  • True: Remove classification head
  • 2: Truncate last 2 layers
  • False: Return single tensor (not list)

Multi-Scale Features

Set the last parameter to True to get intermediate feature maps for multi-scale detection.

Index Module for Feature Selection

When using models that output multiple feature maps, the Index module selects specific outputs:

backbone:
    - [-1, 1, TorchVision, [768, convnext_tiny, DEFAULT, True, 2, True]] # Multi-output
head:
    - [0, 1, Index, [192, 4]] # Select 4th feature map (192 channels)
    - [0, 1, Index, [384, 6]] # Select 6th feature map (384 channels)
    - [0, 1, Index, [768, 8]] # Select 8th feature map (768 channels)
    - [[1, 2, 3], 1, Detect, [nc]] # Multi-scale detection

Module Resolution System

Understanding how Ultralytics locates and imports modules is crucial for customization:

Module Lookup Process

Ultralytics uses a three-tier system in parse_model:

# Core resolution logic
m = getattr(torch.nn, m[3:]) if "nn." in m else getattr(torchvision.ops, m[4:]) if "ops." in m else globals()[m]
  1. PyTorch modules: Names starting with 'nn.'torch.nn namespace
  2. TorchVision operations: Names starting with 'ops.'torchvision.ops namespace
  3. Ultralytics modules: All other names → global namespace via imports

Module Import Chain

Standard modules become available through imports in tasks.py:

from ultralytics.nn.modules import (  # noqa: F401, E501
    SPPF,
    C2f,
    Conv,
    Detect,
    # ... many more modules
    Index,
    TorchVision,
)

Custom Module Integration

Source Code Modification

Modifying the source code is the most versatile way to integrate your custom modules, but it can be tricky. To define and use a custom module, follow these steps:

  1. Define your module in ultralytics/nn/modules/block.py:

    class CustomBlock(nn.Module):
        def __init__(self, c1, c2):
            super().__init__()
            self.layers = nn.Sequential(nn.Conv2d(c1, c2, 3, 1, 1), nn.BatchNorm2d(c2), nn.ReLU())
    
        def forward(self, x):
            return self.layers(x)
    
  2. Expose your module at the package level in ultralytics/nn/modules/__init__.py:

    from .block import CustomBlock  # noqa makes CustomBlock available as ultralytics.nn.modules.CustomBlock
    
  3. Add to imports in ultralytics/nn/tasks.py:

    from ultralytics.nn.modules import CustomBlock  # noqa
    
  4. Handle special arguments (if needed) inside parse_model() in ultralytics/nn/tasks.py:

    elif m is CustomBlock:
        c1, c2 = ch[f], args[0]  # input channels, output channels
        args = [c1, c2, *args[1:]]
    
  5. Use the module in your model YAML:

    # custom_model.yaml
    nc: 1
    backbone:
        - [-1, 1, CustomBlock, [64]]
    head:
        - [-1, 1, Classify, [nc]]
    
  6. Check FLOPs to ensure the forward pass works:

    from ultralytics import YOLO
    
    model = YOLO("custom_model.yaml", task="classify")
    model.info()  # should print non-zero FLOPs if working
    

Example Configurations

Basic Detection Model

# Simple YOLO detection model
nc: 80
scales:
    n: [0.33, 0.25, 1024]

backbone:
    - [-1, 1, Conv, [64, 3, 2]] # 0-P1/2
    - [-1, 1, Conv, [128, 3, 2]] # 1-P2/4
    - [-1, 3, C2f, [128, True]] # 2
    - [-1, 1, Conv, [256, 3, 2]] # 3-P3/8
    - [-1, 6, C2f, [256, True]] # 4
    - [-1, 1, SPPF, [256, 5]] # 5

head:
    - [-1, 1, Conv, [256, 3, 1]] # 6
    - [[6], 1, Detect, [nc]] # 7

TorchVision Backbone Model

# ConvNeXt backbone with YOLO head
nc: 80

backbone:
    - [-1, 1, TorchVision, [768, convnext_tiny, DEFAULT, True, 2, True]]

head:
    - [0, 1, Index, [192, 4]] # P3 features
    - [0, 1, Index, [384, 6]] # P4 features
    - [0, 1, Index, [768, 8]] # P5 features
    - [[1, 2, 3], 1, Detect, [nc]] # Multi-scale detection

Classification Model

# Simple classification model
nc: 1000

backbone:
    - [-1, 1, Conv, [64, 7, 2, 3]]
    - [-1, 1, nn.MaxPool2d, [3, 2, 1]]
    - [-1, 4, C2f, [64, True]]
    - [-1, 1, Conv, [128, 3, 2]]
    - [-1, 8, C2f, [128, True]]
    - [-1, 1, nn.AdaptiveAvgPool2d, [1]]

head:
    - [-1, 1, Classify, [nc]]

Best Practices

Architecture Design Tips

Start Simple: Begin with proven architectures before customizing. Use existing YOLO configurations as templates and modify incrementally rather than building from scratch.

Test Incrementally: Validate each modification step-by-step. Add one custom module at a time and verify it works before proceeding to the next change.

Monitor Channels: Ensure channel dimensions match between connected layers. The output channels (c2) of one layer must match the input channels (c1) of the next layer in the sequence.

Use Skip Connections: Leverage feature reuse with [[-1, N], 1, Concat, [1]] patterns. These connections help with gradient flow and allow the model to combine features from different scales.

Scale Appropriately: Choose model scales based on your computational constraints. Use nano (n) for edge devices, small (s) for balanced performance, and larger scales (m, l, x) for maximum accuracy.

Performance Considerations

Depth vs Width: Deep networks capture complex hierarchical features through multiple transformation layers, while wide networks process more information in parallel at each layer. Balance these based on your task complexity.

Skip Connections: Improve gradient flow during training and enable feature reuse throughout the network. They're particularly important in deeper architectures to prevent vanishing gradients.

Bottleneck Blocks: Reduce computational cost while maintaining model expressiveness. Modules like C2f use fewer parameters than standard convolutions while preserving feature learning capacity.

Multi-Scale Features: Essential for detecting objects at different sizes in the same image. Use Feature Pyramid Network (FPN) patterns with multiple detection heads at different scales.

Troubleshooting

Common Issues

Problem Cause Solution
KeyError: 'ModuleName' Module not imported Add to tasks.py imports
Channel dimension mismatch Incorrect args specification Verify input/output channel compatibility
AttributeError: 'int' object has no attribute Wrong argument type Check module documentation for correct argument types
Model fails to build Invalid from reference Ensure referenced layers exist

Debugging Tips

When developing custom architectures, systematic debugging helps identify issues early:

Use Identity Head for Testing

Replace complex heads with nn.Identity to isolate backbone issues:

nc: 1
backbone:
    - [-1, 1, CustomBlock, [64]]
head:
    - [-1, 1, nn.Identity, []] # Pass-through for debugging

This allows direct inspection of backbone outputs:

import torch

from ultralytics import YOLO

model = YOLO("debug_model.yaml")
output = model.model(torch.randn(1, 3, 640, 640))
print(f"Output shape: {output.shape}")  # Should match expected dimensions

Model Architecture Inspection

Checking the FLOPs count and printing out each layer can also help debug issues with your custom model config. FLOPs count should be non-zero for a valid model. If it's zero, then there's likely an issue with the forward pass. Running a simple forward pass should show the exact error being encountered.

from ultralytics import YOLO

# Build model with verbose output to see layer details
model = YOLO("debug_model.yaml", verbose=True)

# Check model FLOPs. Failed forward pass causes 0 FLOPs.
model.info()

# Inspect individual layers
for i, layer in enumerate(model.model.model):
    print(f"Layer {i}: {layer}")

Step-by-Step Validation

  1. Start minimal: Test with simplest possible architecture first
  2. Add incrementally: Build complexity layer by layer
  3. Check dimensions: Verify channel and spatial size compatibility
  4. Validate scaling: Test with different model scales (n, s, m)

FAQ

How do I change the number of classes in my model?

Set the nc parameter at the top of your YAML file to match your dataset's number of classes.

nc: 5 # 5 classes

Can I use a custom backbone in my model YAML?

Yes. You can use any supported module, including TorchVision backbones, or define your own custom module and import it as described in Custom Module Integration.

How do I scale my model for different sizes (nano, small, medium, etc.)?

Use the scales section in your YAML to define scaling factors for depth, width, and max channels. The model will automatically apply these when you load the base YAML file with the scale appended to the filename (e.g., yolo11n.yaml).

What does the [from, repeats, module, args] format mean?

This format specifies how each layer is constructed:

  • from: input source(s)
  • repeats: number of times to repeat the module
  • module: the layer type
  • args: arguments for the module

How do I troubleshoot channel mismatch errors?

Check that the output channels of one layer match the expected input channels of the next. Use print(model.model.model) to inspect your model's architecture.

Where can I find a list of available modules and their arguments?

Check the source code in the ultralytics/nn/modules directory for all available modules and their arguments.

How do I add a custom module to my YAML configuration?

Define your module in the source code, import it as shown in Source Code Modification, and reference it by name in your YAML file.

Can I use pretrained weights with a custom YAML?

Yes, you can use model.load("path/to/weights") to load weights from a pretrained checkpoint. However, only weights for layers that match would load successfully.

How do I validate my model configuration?

Use model.info() to check whether FLOPs count is non-zero. A valid model should show non-zero FLOPs count. If it's zero, follow the suggestions in Debugging Tips to find the issue.



📅 Created 0 days ago ✏️ Updated 0 days ago