Model YAML Configuration
The model YAML configuration file serves as the architectural blueprint for Ultralytics neural networks. It defines how layers connect, what parameters each module uses, and how the entire network scales across different model sizes.
Configuration Structure
Model YAML files are organized into three main sections that work together to define the architecture.
Parameters Section
The parameters section specifies the model's global characteristics and scaling behavior:
# Parameters
nc: 80 # number of classes
scales: # compound scaling constants [depth, width, max_channels]
n: [0.50, 0.25, 1024] # nano: shallow layers, narrow channels
s: [0.50, 0.50, 1024] # small: shallow depth, standard width
m: [0.50, 1.00, 512] # medium: moderate depth, full width
l: [1.00, 1.00, 512] # large: full depth and width
x: [1.00, 1.50, 512] # extra-large: maximum performance
kpt_shape: [17, 3] # pose models only
nc
sets the number of classes the model predicts.scales
define compound scaling factors that adjust model depth, width, and maximum channels to produce different size variants (nano through extra-large).kpt_shape
applies to pose models. It can be[N, 2]
for(x, y)
keypoints or[N, 3]
for(x, y, visibility)
.
Reduce redundancy with scales
The scales
parameter lets you generate multiple model sizes from a single base YAML. For instance, when you load yolo11n.yaml
, Ultralytics reads the base yolo11.yaml
and applies the n
scaling factors (depth=0.50
, width=0.25
) to build the nano variant.
nc
and kpt_shape
are dataset-dependent
If your dataset specifies a different nc
or kpt_shape
, Ultralytics will automatically override the model config at runtime to match the dataset YAML.
Backbone and Head Architecture
The model architecture consists of backbone (feature extraction) and head (task-specific) sections:
backbone:
# [from, repeats, module, args]
- [-1, 1, Conv, [64, 3, 2]] # 0: Initial convolution
- [-1, 1, Conv, [128, 3, 2]] # 1: Downsample
- [-1, 3, C2f, [128, True]] # 2: Feature processing
head:
- [-1, 1, nn.Upsample, [None, 2, nearest]] # 6: Upsample
- [[-1, 2], 1, Concat, [1]] # 7: Skip connection
- [-1, 3, C2f, [256]] # 8: Process features
- [[8], 1, Detect, [nc]] # 9: Detection layer
Layer Specification Format
Every layer follows the consistent pattern: [from, repeats, module, args]
Component | Purpose | Examples |
---|---|---|
from | Input connections | -1 (previous), 6 (layer 6), [4, 6, 8] (multi-input) |
repeats | Number of repetitions | 1 (single), 3 (repeat 3 times) |
module | Module type | Conv , C2f , TorchVision , Detect |
args | Module arguments | [64, 3, 2] (channels, kernel, stride) |
Connection Patterns
The from
field creates flexible data flow patterns throughout your network:
- [-1, 1, Conv, [64, 3, 2]] # Takes input from previous layer
- [[-1, 6], 1, Concat, [1]] # Combines current layer with layer 6
- [[4, 6, 8], 1, Detect, [nc]] # Detection head using 3 feature scales
Layer Indexing
Layers are indexed starting from 0. Negative indices reference previous layers (-1
= previous layer), while positive indices reference specific layers by their position.
Module Repetition
The repeats
parameter creates deeper network sections:
- [-1, 3, C2f, [128, True]] # Creates 3 consecutive C2f blocks
- [-1, 1, Conv, [64, 3, 2]] # Single convolution layer
The actual repetition count gets multiplied by the depth scaling factor from your model size configuration.
Available Modules
Modules are organized by functionality and defined in the Ultralytics modules directory. The following tables show commonly used modules by category, with many more available in the source code:
Basic Operations
Module | Purpose | Source | Arguments |
---|---|---|---|
Conv |
Convolution + BatchNorm + Activation | conv.py | [out_ch, kernel, stride, pad, groups] |
nn.Upsample |
Spatial upsampling | PyTorch | [size, scale_factor, mode] |
nn.Identity |
Pass-through operation | PyTorch | [] |
Composite Blocks
Module | Purpose | Source | Arguments |
---|---|---|---|
C2f |
CSP bottleneck with 2 convolutions | block.py | [out_ch, shortcut, expansion] |
SPPF |
Spatial Pyramid Pooling (fast) | block.py | [out_ch, kernel_size] |
Concat |
Channel-wise concatenation | conv.py | [dimension] |
Specialized Modules
Module | Purpose | Source | Arguments |
---|---|---|---|
TorchVision |
Load any torchvision model | block.py | [out_ch, model_name, weights, unwrap, truncate, split] |
Index |
Extract specific tensor from list | block.py | [out_ch, index] |
Detect |
YOLO detection head | head.py | [nc, anchors, ch] |
Complete Module List
This represents a subset of available modules. For the full list of modules and their parameters, explore the modules directory.
Advanced Features
TorchVision Integration
The TorchVision module enables seamless integration of any TorchVision model as a backbone:
from ultralytics import YOLO
# Model with ConvNeXt backbone
model = YOLO("convnext_backbone.yaml")
results = model.train(data="coco8.yaml", epochs=100)
backbone:
- [-1, 1, TorchVision, [768, convnext_tiny, DEFAULT, True, 2, False]]
head:
- [-1, 1, Classify, [nc]]
Parameter Breakdown:
768
: Expected output channelsconvnext_tiny
: Model architecture (available models)DEFAULT
: Use pretrained weightsTrue
: Remove classification head2
: Truncate last 2 layersFalse
: Return single tensor (not list)
Multi-Scale Features
Set the last parameter to True
to get intermediate feature maps for multi-scale detection.
Index Module for Feature Selection
When using models that output multiple feature maps, the Index module selects specific outputs:
backbone:
- [-1, 1, TorchVision, [768, convnext_tiny, DEFAULT, True, 2, True]] # Multi-output
head:
- [0, 1, Index, [192, 4]] # Select 4th feature map (192 channels)
- [0, 1, Index, [384, 6]] # Select 6th feature map (384 channels)
- [0, 1, Index, [768, 8]] # Select 8th feature map (768 channels)
- [[1, 2, 3], 1, Detect, [nc]] # Multi-scale detection
Module Resolution System
Understanding how Ultralytics locates and imports modules is crucial for customization:
Module Lookup Process
Ultralytics uses a three-tier system in parse_model
:
# Core resolution logic
m = getattr(torch.nn, m[3:]) if "nn." in m else getattr(torchvision.ops, m[4:]) if "ops." in m else globals()[m]
- PyTorch modules: Names starting with
'nn.'
→torch.nn
namespace - TorchVision operations: Names starting with
'ops.'
→torchvision.ops
namespace - Ultralytics modules: All other names → global namespace via imports
Module Import Chain
Standard modules become available through imports in tasks.py
:
from ultralytics.nn.modules import ( # noqa: F401, E501
SPPF,
C2f,
Conv,
Detect,
# ... many more modules
Index,
TorchVision,
)
Custom Module Integration
Source Code Modification
Modifying the source code is the most versatile way to integrate your custom modules, but it can be tricky. To define and use a custom module, follow these steps:
-
Define your module in
ultralytics/nn/modules/block.py
:class CustomBlock(nn.Module): def __init__(self, c1, c2): super().__init__() self.layers = nn.Sequential(nn.Conv2d(c1, c2, 3, 1, 1), nn.BatchNorm2d(c2), nn.ReLU()) def forward(self, x): return self.layers(x)
-
Expose your module at the package level in
ultralytics/nn/modules/__init__.py
:from .block import CustomBlock # noqa makes CustomBlock available as ultralytics.nn.modules.CustomBlock
-
Add to imports in
ultralytics/nn/tasks.py
:from ultralytics.nn.modules import CustomBlock # noqa
-
Handle special arguments (if needed) inside
parse_model()
inultralytics/nn/tasks.py
:elif m is CustomBlock: c1, c2 = ch[f], args[0] # input channels, output channels args = [c1, c2, *args[1:]]
-
Use the module in your model YAML:
# custom_model.yaml nc: 1 backbone: - [-1, 1, CustomBlock, [64]] head: - [-1, 1, Classify, [nc]]
-
Check FLOPs to ensure the forward pass works:
from ultralytics import YOLO model = YOLO("custom_model.yaml", task="classify") model.info() # should print non-zero FLOPs if working
Example Configurations
Basic Detection Model
# Simple YOLO detection model
nc: 80
scales:
n: [0.33, 0.25, 1024]
backbone:
- [-1, 1, Conv, [64, 3, 2]] # 0-P1/2
- [-1, 1, Conv, [128, 3, 2]] # 1-P2/4
- [-1, 3, C2f, [128, True]] # 2
- [-1, 1, Conv, [256, 3, 2]] # 3-P3/8
- [-1, 6, C2f, [256, True]] # 4
- [-1, 1, SPPF, [256, 5]] # 5
head:
- [-1, 1, Conv, [256, 3, 1]] # 6
- [[6], 1, Detect, [nc]] # 7
TorchVision Backbone Model
# ConvNeXt backbone with YOLO head
nc: 80
backbone:
- [-1, 1, TorchVision, [768, convnext_tiny, DEFAULT, True, 2, True]]
head:
- [0, 1, Index, [192, 4]] # P3 features
- [0, 1, Index, [384, 6]] # P4 features
- [0, 1, Index, [768, 8]] # P5 features
- [[1, 2, 3], 1, Detect, [nc]] # Multi-scale detection
Classification Model
# Simple classification model
nc: 1000
backbone:
- [-1, 1, Conv, [64, 7, 2, 3]]
- [-1, 1, nn.MaxPool2d, [3, 2, 1]]
- [-1, 4, C2f, [64, True]]
- [-1, 1, Conv, [128, 3, 2]]
- [-1, 8, C2f, [128, True]]
- [-1, 1, nn.AdaptiveAvgPool2d, [1]]
head:
- [-1, 1, Classify, [nc]]
Best Practices
Architecture Design Tips
Start Simple: Begin with proven architectures before customizing. Use existing YOLO configurations as templates and modify incrementally rather than building from scratch.
Test Incrementally: Validate each modification step-by-step. Add one custom module at a time and verify it works before proceeding to the next change.
Monitor Channels: Ensure channel dimensions match between connected layers. The output channels (c2
) of one layer must match the input channels (c1
) of the next layer in the sequence.
Use Skip Connections: Leverage feature reuse with [[-1, N], 1, Concat, [1]]
patterns. These connections help with gradient flow and allow the model to combine features from different scales.
Scale Appropriately: Choose model scales based on your computational constraints. Use nano (n
) for edge devices, small (s
) for balanced performance, and larger scales (m
, l
, x
) for maximum accuracy.
Performance Considerations
Depth vs Width: Deep networks capture complex hierarchical features through multiple transformation layers, while wide networks process more information in parallel at each layer. Balance these based on your task complexity.
Skip Connections: Improve gradient flow during training and enable feature reuse throughout the network. They're particularly important in deeper architectures to prevent vanishing gradients.
Bottleneck Blocks: Reduce computational cost while maintaining model expressiveness. Modules like C2f
use fewer parameters than standard convolutions while preserving feature learning capacity.
Multi-Scale Features: Essential for detecting objects at different sizes in the same image. Use Feature Pyramid Network (FPN) patterns with multiple detection heads at different scales.
Troubleshooting
Common Issues
Problem | Cause | Solution |
---|---|---|
KeyError: 'ModuleName' |
Module not imported | Add to tasks.py imports |
Channel dimension mismatch | Incorrect args specification |
Verify input/output channel compatibility |
AttributeError: 'int' object has no attribute |
Wrong argument type | Check module documentation for correct argument types |
Model fails to build | Invalid from reference |
Ensure referenced layers exist |
Debugging Tips
When developing custom architectures, systematic debugging helps identify issues early:
Use Identity Head for Testing
Replace complex heads with nn.Identity
to isolate backbone issues:
nc: 1
backbone:
- [-1, 1, CustomBlock, [64]]
head:
- [-1, 1, nn.Identity, []] # Pass-through for debugging
This allows direct inspection of backbone outputs:
import torch
from ultralytics import YOLO
model = YOLO("debug_model.yaml")
output = model.model(torch.randn(1, 3, 640, 640))
print(f"Output shape: {output.shape}") # Should match expected dimensions
Model Architecture Inspection
Checking the FLOPs count and printing out each layer can also help debug issues with your custom model config. FLOPs count should be non-zero for a valid model. If it's zero, then there's likely an issue with the forward pass. Running a simple forward pass should show the exact error being encountered.
from ultralytics import YOLO
# Build model with verbose output to see layer details
model = YOLO("debug_model.yaml", verbose=True)
# Check model FLOPs. Failed forward pass causes 0 FLOPs.
model.info()
# Inspect individual layers
for i, layer in enumerate(model.model.model):
print(f"Layer {i}: {layer}")
Step-by-Step Validation
- Start minimal: Test with simplest possible architecture first
- Add incrementally: Build complexity layer by layer
- Check dimensions: Verify channel and spatial size compatibility
- Validate scaling: Test with different model scales (
n
,s
,m
)
FAQ
How do I change the number of classes in my model?
Set the nc
parameter at the top of your YAML file to match your dataset's number of classes.
nc: 5 # 5 classes
Can I use a custom backbone in my model YAML?
Yes. You can use any supported module, including TorchVision backbones, or define your own custom module and import it as described in Custom Module Integration.
How do I scale my model for different sizes (nano, small, medium, etc.)?
Use the scales
section in your YAML to define scaling factors for depth, width, and max channels. The model will automatically apply these when you load the base YAML file with the scale appended to the filename (e.g., yolo11n.yaml
).
What does the [from, repeats, module, args]
format mean?
This format specifies how each layer is constructed:
from
: input source(s)repeats
: number of times to repeat the modulemodule
: the layer typeargs
: arguments for the module
How do I troubleshoot channel mismatch errors?
Check that the output channels of one layer match the expected input channels of the next. Use print(model.model.model)
to inspect your model's architecture.
Where can I find a list of available modules and their arguments?
Check the source code in the ultralytics/nn/modules
directory for all available modules and their arguments.
How do I add a custom module to my YAML configuration?
Define your module in the source code, import it as shown in Source Code Modification, and reference it by name in your YAML file.
Can I use pretrained weights with a custom YAML?
Yes, you can use model.load("path/to/weights")
to load weights from a pretrained checkpoint. However, only weights for layers that match would load successfully.
How do I validate my model configuration?
Use model.info()
to check whether FLOPs count is non-zero. A valid model should show non-zero FLOPs count. If it's zero, follow the suggestions in Debugging Tips to find the issue.