| |
| import datetime |
| import logging |
| import os |
| import platform |
| import warnings |
|
|
| import cv2 |
| import torch.multiprocessing as mp |
| from mmengine import DefaultScope |
| from mmengine.logging import print_log |
| from mmengine.utils import digit_version |
|
|
|
|
| def setup_cache_size_limit_of_dynamo(): |
| """Setup cache size limit of dynamo. |
| |
| Note: Due to the dynamic shape of the loss calculation and |
| post-processing parts in the object detection algorithm, these |
| functions must be compiled every time they are run. |
| Setting a large value for torch._dynamo.config.cache_size_limit |
| may result in repeated compilation, which can slow down training |
| and testing speed. Therefore, we need to set the default value of |
| cache_size_limit smaller. An empirical value is 4. |
| """ |
|
|
| import torch |
| if digit_version(torch.__version__) >= digit_version('2.0.0'): |
| if 'DYNAMO_CACHE_SIZE_LIMIT' in os.environ: |
| import torch._dynamo |
| cache_size_limit = int(os.environ['DYNAMO_CACHE_SIZE_LIMIT']) |
| torch._dynamo.config.cache_size_limit = cache_size_limit |
| print_log( |
| f'torch._dynamo.config.cache_size_limit is force ' |
| f'set to {cache_size_limit}.', |
| logger='current', |
| level=logging.WARNING) |
|
|
|
|
| def setup_multi_processes(cfg): |
| """Setup multi-processing environment variables.""" |
| |
| if platform.system() != 'Windows': |
| mp_start_method = cfg.get('mp_start_method', 'fork') |
| current_method = mp.get_start_method(allow_none=True) |
| if current_method is not None and current_method != mp_start_method: |
| warnings.warn( |
| f'Multi-processing start method `{mp_start_method}` is ' |
| f'different from the previous setting `{current_method}`.' |
| f'It will be force set to `{mp_start_method}`. You can change ' |
| f'this behavior by changing `mp_start_method` in your config.') |
| mp.set_start_method(mp_start_method, force=True) |
|
|
| |
| opencv_num_threads = cfg.get('opencv_num_threads', 0) |
| cv2.setNumThreads(opencv_num_threads) |
|
|
| |
| |
| workers_per_gpu = cfg.data.get('workers_per_gpu', 1) |
| if 'train_dataloader' in cfg.data: |
| workers_per_gpu = \ |
| max(cfg.data.train_dataloader.get('workers_per_gpu', 1), |
| workers_per_gpu) |
|
|
| if 'OMP_NUM_THREADS' not in os.environ and workers_per_gpu > 1: |
| omp_num_threads = 1 |
| warnings.warn( |
| f'Setting OMP_NUM_THREADS environment variable for each process ' |
| f'to be {omp_num_threads} in default, to avoid your system being ' |
| f'overloaded, please further tune the variable for optimal ' |
| f'performance in your application as needed.') |
| os.environ['OMP_NUM_THREADS'] = str(omp_num_threads) |
|
|
| |
| if 'MKL_NUM_THREADS' not in os.environ and workers_per_gpu > 1: |
| mkl_num_threads = 1 |
| warnings.warn( |
| f'Setting MKL_NUM_THREADS environment variable for each process ' |
| f'to be {mkl_num_threads} in default, to avoid your system being ' |
| f'overloaded, please further tune the variable for optimal ' |
| f'performance in your application as needed.') |
| os.environ['MKL_NUM_THREADS'] = str(mkl_num_threads) |
|
|
|
|
| def register_all_modules(init_default_scope: bool = True) -> None: |
| """Register all modules in mmdet into the registries. |
| |
| Args: |
| init_default_scope (bool): Whether initialize the mmdet default scope. |
| When `init_default_scope=True`, the global default scope will be |
| set to `mmdet`, and all registries will build modules from mmdet's |
| registry node. To understand more about the registry, please refer |
| to https://github.com/open-mmlab/mmengine/blob/main/docs/en/tutorials/registry.md |
| Defaults to True. |
| """ |
| import mmdet.datasets |
| import mmdet.engine |
| import mmdet.evaluation |
| import mmdet.models |
| import mmdet.visualization |
|
|
| if init_default_scope: |
| never_created = DefaultScope.get_current_instance() is None \ |
| or not DefaultScope.check_instance_created('mmdet') |
| if never_created: |
| DefaultScope.get_instance('mmdet', scope_name='mmdet') |
| return |
| current_scope = DefaultScope.get_current_instance() |
| if current_scope.scope_name != 'mmdet': |
| warnings.warn('The current default scope ' |
| f'"{current_scope.scope_name}" is not "mmdet", ' |
| '`register_all_modules` will force the current' |
| 'default scope to be "mmdet". If this is not ' |
| 'expected, please set `init_default_scope=False`.') |
| |
| new_instance_name = f'mmdet-{datetime.datetime.now()}' |
| DefaultScope.get_instance(new_instance_name, scope_name='mmdet') |
|
|