TransformerEncoderLayer
Bases: nn.Module
Transformer Encoder.
Source code in ultralytics/nn/modules/transformer.py
forward(src, src_mask=None, src_key_padding_mask=None, pos=None)
Forward propagates the input through the encoder module.
Source code in ultralytics/nn/modules/transformer.py
AIFI
Bases: TransformerEncoderLayer
Source code in ultralytics/nn/modules/transformer.py
TransformerLayer
Bases: nn.Module
Transformer layer https://arxiv.org/abs/2010.11929 (LayerNorm layers removed for better performance).
Source code in ultralytics/nn/modules/transformer.py
__init__(c, num_heads)
Initializes a self-attention mechanism using linear transformations and multi-head attention.
Source code in ultralytics/nn/modules/transformer.py
forward(x)
Apply a transformer block to the input x and return the output.
TransformerBlock
Bases: nn.Module
Vision Transformer https://arxiv.org/abs/2010.11929.
Source code in ultralytics/nn/modules/transformer.py
__init__(c1, c2, num_heads, num_layers)
Initialize a Transformer module with position embedding and specified number of heads and layers.
Source code in ultralytics/nn/modules/transformer.py
forward(x)
Forward propagates the input through the bottleneck module.
Source code in ultralytics/nn/modules/transformer.py
MLPBlock
Bases: nn.Module
Source code in ultralytics/nn/modules/transformer.py
MLP
Bases: nn.Module
Very simple multi-layer perceptron (also called FFN)
Source code in ultralytics/nn/modules/transformer.py
LayerNorm2d
Bases: nn.Module
Source code in ultralytics/nn/modules/transformer.py
MSDeformAttn
Bases: nn.Module
Original Multi-Scale Deformable Attention Module. https://github.com/fundamentalvision/Deformable-DETR/blob/main/models/ops/modules/ms_deform_attn.py
Source code in ultralytics/nn/modules/transformer.py
187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 |
|
forward(query, reference_points, value, value_spatial_shapes, value_mask=None)
https://github.com/PaddlePaddle/PaddleDetection/blob/develop/ppdet/modeling/transformers/deformable_transformer.py
Parameters:
Name | Type | Description | Default |
---|---|---|---|
query |
Tensor
|
[bs, query_length, C] |
required |
reference_points |
Tensor
|
[bs, query_length, n_levels, 2], range in [0, 1], top-left (0,0), bottom-right (1, 1), including padding area |
required |
value |
Tensor
|
[bs, value_length, C] |
required |
value_spatial_shapes |
List
|
[n_levels, 2], [(H_0, W_0), (H_1, W_1), ..., (H_{L-1}, W_{L-1})] |
required |
value_mask |
Tensor
|
[bs, value_length], True for non-padding elements, False for padding elements |
None
|
Returns:
Name | Type | Description |
---|---|---|
output |
Tensor
|
[bs, Length_{query}, C] |
Source code in ultralytics/nn/modules/transformer.py
DeformableTransformerDecoderLayer
Bases: nn.Module
https://github.com/PaddlePaddle/PaddleDetection/blob/develop/ppdet/modeling/transformers/deformable_transformer.py https://github.com/fundamentalvision/Deformable-DETR/blob/main/models/deformable_transformer.py
Source code in ultralytics/nn/modules/transformer.py
DeformableTransformerDecoder
Bases: nn.Module
https://github.com/PaddlePaddle/PaddleDetection/blob/develop/ppdet/modeling/transformers/deformable_transformer.py
Source code in ultralytics/nn/modules/transformer.py
Created 2023-05-11, Updated 2023-05-17
Authors: Glenn Jocher (3)