Reference for ultralytics/utils/dist.py
Note
This file is available at https://github.com/ultralytics/ultralytics/blob/main/ultralytics/utils/dist.py. If you spot a problem please help fix it by contributing a Pull Request 🛠️. Thank you 🙏!
ultralytics.utils.dist.find_free_network_port
find_free_network_port() -> int
Find a free port on localhost.
It is useful in single-node training when we don't want to connect to a real main node but have to set the
MASTER_PORT
environment variable.
Returns:
Type | Description |
---|---|
int
|
The available network port number. |
Source code in ultralytics/utils/dist.py
13 14 15 16 17 18 19 20 21 22 23 24 25 |
|
ultralytics.utils.dist.generate_ddp_file
generate_ddp_file(trainer)
Generate a DDP (Distributed Data Parallel) file for multi-GPU training.
This function creates a temporary Python file that enables distributed training across multiple GPUs. The file contains the necessary configuration to initialize the trainer in a distributed environment.
Parameters:
Name | Type | Description | Default |
---|---|---|---|
trainer
|
object
|
The trainer object containing training configuration and arguments. Must have args attribute and be a class instance. |
required |
Returns:
Type | Description |
---|---|
str
|
Path to the generated temporary DDP file. |
Notes
The generated file is saved in the USER_CONFIG_DIR/DDP directory and includes: - Trainer class import - Configuration overrides from the trainer arguments - Model path configuration - Training initialization code
Source code in ultralytics/utils/dist.py
28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 |
|
ultralytics.utils.dist.generate_ddp_command
generate_ddp_command(world_size, trainer)
Generate command for distributed training.
Parameters:
Name | Type | Description | Default |
---|---|---|---|
world_size
|
int
|
Number of processes to spawn for distributed training. |
required |
trainer
|
object
|
The trainer object containing configuration for distributed training. |
required |
Returns:
Name | Type | Description |
---|---|---|
cmd |
List[str]
|
The command to execute for distributed training. |
file |
str
|
Path to the temporary file created for DDP training. |
Source code in ultralytics/utils/dist.py
78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 |
|
ultralytics.utils.dist.ddp_cleanup
ddp_cleanup(trainer, file)
Delete temporary file if created during distributed data parallel (DDP) training.
This function checks if the provided file contains the trainer's ID in its name, indicating it was created as a temporary file for DDP training, and deletes it if so.
Parameters:
Name | Type | Description | Default |
---|---|---|---|
trainer
|
object
|
The trainer object used for distributed training. |
required |
file
|
str
|
Path to the file that might need to be deleted. |
required |
Examples:
>>> trainer = YOLOTrainer()
>>> file = "/tmp/ddp_temp_123456789.py"
>>> ddp_cleanup(trainer, file)
Source code in ultralytics/utils/dist.py
101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 |
|