| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from __future__ import annotations |
|
|
| import os |
| import time |
| from collections.abc import Mapping, MutableMapping |
| from typing import Any, cast |
|
|
| import torch |
| import torch.distributed as dist |
|
|
| from monai.apps.auto3dseg.data_analyzer import DataAnalyzer |
| from monai.apps.utils import get_logger |
| from monai.auto3dseg import SegSummarizer |
| from monai.bundle import BundleWorkflow, ConfigComponent, ConfigItem, ConfigParser, ConfigWorkflow |
| from monai.engines import SupervisedEvaluator, SupervisedTrainer, Trainer |
| from monai.fl.client import ClientAlgo, ClientAlgoStats |
| from monai.fl.utils.constants import ExtraItems, FiltersType, FlPhase, FlStatistics, ModelType, WeightType |
| from monai.fl.utils.exchange_object import ExchangeObject |
| from monai.networks.utils import copy_model_state, get_state_dict |
| from monai.utils import min_version, require_pkg |
| from monai.utils.enums import DataStatsKeys |
|
|
| logger = get_logger(__name__) |
|
|
|
|
| def convert_global_weights(global_weights: Mapping, local_var_dict: MutableMapping) -> tuple[MutableMapping, int]: |
| """Helper function to convert global weights to local weights format""" |
| |
| model_keys = global_weights.keys() |
| n_converted = 0 |
| for var_name in local_var_dict: |
| if var_name in model_keys: |
| weights = global_weights[var_name] |
| try: |
| |
| weights = torch.reshape(torch.as_tensor(weights), local_var_dict[var_name].shape) |
| |
| local_var_dict[var_name] = weights |
| n_converted += 1 |
| except Exception as e: |
| raise ValueError(f"Convert weight from {var_name} failed.") from e |
| return local_var_dict, n_converted |
|
|
|
|
| def compute_weight_diff(global_weights, local_var_dict): |
| if global_weights is None: |
| raise ValueError("Cannot compute weight differences if `global_weights` is None!") |
| if local_var_dict is None: |
| raise ValueError("Cannot compute weight differences if `local_var_dict` is None!") |
| |
| weight_diff = {} |
| n_diff = 0 |
| for name in global_weights: |
| if name not in local_var_dict: |
| continue |
| |
| weight_diff[name] = local_var_dict[name].cpu() - global_weights[name].cpu() |
| n_diff += 1 |
| if torch.any(torch.isnan(weight_diff[name])): |
| raise ValueError(f"Weights for {name} became NaN...") |
| if n_diff == 0: |
| raise RuntimeError("No weight differences computed!") |
| return weight_diff |
|
|
|
|
| def disable_ckpt_loaders(parser: ConfigParser) -> None: |
| if "validate#handlers" in parser: |
| for h in parser["validate#handlers"]: |
| if ConfigComponent.is_instantiable(h): |
| if "CheckpointLoader" in h["_target_"]: |
| h["_disabled_"] = True |
|
|
|
|
| class MonaiAlgoStats(ClientAlgoStats): |
| """ |
| Implementation of ``ClientAlgoStats`` to allow federated learning with MONAI bundle configurations. |
| |
| Args: |
| bundle_root: directory path of the bundle. |
| config_train_filename: bundle training config path relative to bundle_root. Can be a list of files; |
| defaults to "configs/train.json". only useful when `workflow` is None. |
| config_filters_filename: filter configuration file. Can be a list of files; defaults to `None`. |
| data_stats_transform_list: transforms to apply for the data stats result. |
| histogram_only: whether to only compute histograms. Defaults to False. |
| workflow: the bundle workflow to execute, usually it's training, evaluation or inference. |
| if None, will create an `ConfigWorkflow` internally based on `config_train_filename`. |
| """ |
|
|
| def __init__( |
| self, |
| bundle_root: str, |
| config_train_filename: str | list | None = "configs/train.json", |
| config_filters_filename: str | list | None = None, |
| data_stats_transform_list: list | None = None, |
| histogram_only: bool = False, |
| workflow: BundleWorkflow | None = None, |
| ): |
| self.logger = logger |
| self.bundle_root = bundle_root |
| self.config_train_filename = config_train_filename |
| self.config_filters_filename = config_filters_filename |
| self.train_data_key = "train" |
| self.eval_data_key = "eval" |
| self.data_stats_transform_list = data_stats_transform_list |
| self.histogram_only = histogram_only |
| self.workflow = None |
| if workflow is not None: |
| if not isinstance(workflow, BundleWorkflow): |
| raise ValueError("workflow must be a subclass of BundleWorkflow.") |
| if workflow.get_workflow_type() is None: |
| raise ValueError("workflow doesn't specify the type.") |
| self.workflow = workflow |
|
|
| self.client_name: str | None = None |
| self.app_root: str = "" |
| self.post_statistics_filters: Any = None |
| self.phase = FlPhase.IDLE |
| self.dataset_root: Any = None |
|
|
| def initialize(self, extra=None): |
| """ |
| Initialize routine to parse configuration files and extract main components such as trainer, evaluator, and filters. |
| |
| Args: |
| extra: Dict with additional information that should be provided by FL system, |
| i.e., `ExtraItems.CLIENT_NAME` and `ExtraItems.APP_ROOT`. |
| |
| """ |
| if extra is None: |
| extra = {} |
| self.client_name = extra.get(ExtraItems.CLIENT_NAME, "noname") |
| self.logger.info(f"Initializing {self.client_name} ...") |
|
|
| |
| self.app_root = extra.get(ExtraItems.APP_ROOT, "") |
| self.bundle_root = os.path.join(self.app_root, self.bundle_root) |
|
|
| if self.workflow is None: |
| config_train_files = self._add_config_files(self.config_train_filename) |
| self.workflow = ConfigWorkflow( |
| config_file=config_train_files, meta_file=None, logging_file=None, workflow_type="train" |
| ) |
| self.workflow.initialize() |
| self.workflow.bundle_root = self.bundle_root |
| |
| self.workflow.initialize() |
|
|
| config_filter_files = self._add_config_files(self.config_filters_filename) |
| filter_parser = ConfigParser() |
| if len(config_filter_files) > 0: |
| filter_parser.read_config(config_filter_files) |
| |
| self.post_statistics_filters = filter_parser.get_parsed_content( |
| FiltersType.POST_STATISTICS_FILTERS, default=ConfigItem(None, FiltersType.POST_STATISTICS_FILTERS) |
| ) |
| self.logger.info(f"Initialized {self.client_name}.") |
|
|
| def get_data_stats(self, extra: dict | None = None) -> ExchangeObject: |
| """ |
| Returns summary statistics about the local data. |
| |
| Args: |
| extra: Dict with additional information that can be provided by the FL system. |
| Both FlStatistics.HIST_BINS and FlStatistics.HIST_RANGE must be provided. |
| |
| Returns: |
| stats: ExchangeObject with summary statistics. |
| |
| """ |
| if extra is None: |
| raise ValueError("`extra` has to be set") |
|
|
| if self.workflow.dataset_dir: |
| self.phase = FlPhase.GET_DATA_STATS |
| self.logger.info(f"Computing statistics on {self.workflow.dataset_dir}") |
|
|
| if FlStatistics.HIST_BINS not in extra: |
| raise ValueError("FlStatistics.NUM_OF_BINS not specified in `extra`") |
| else: |
| hist_bins = extra[FlStatistics.HIST_BINS] |
| if FlStatistics.HIST_RANGE not in extra: |
| raise ValueError("FlStatistics.HIST_RANGE not specified in `extra`") |
| else: |
| hist_range = extra[FlStatistics.HIST_RANGE] |
|
|
| stats_dict = {} |
|
|
| |
| train_summary_stats, train_case_stats = self._get_data_key_stats( |
| data=self.workflow.train_dataset_data, |
| data_key=self.train_data_key, |
| hist_bins=hist_bins, |
| hist_range=hist_range, |
| output_path=os.path.join(self.app_root, "train_data_stats.yaml"), |
| ) |
| if train_case_stats: |
| |
| stats_dict.update({self.train_data_key: train_summary_stats}) |
|
|
| |
| eval_summary_stats = None |
| eval_case_stats = None |
| if self.workflow.val_dataset_data is not None: |
| eval_summary_stats, eval_case_stats = self._get_data_key_stats( |
| data=self.workflow.val_dataset_data, |
| data_key=self.eval_data_key, |
| hist_bins=hist_bins, |
| hist_range=hist_range, |
| output_path=os.path.join(self.app_root, "eval_data_stats.yaml"), |
| ) |
| else: |
| self.logger.warning("the datalist doesn't contain validation section.") |
| if eval_summary_stats: |
| |
| stats_dict.update({self.eval_data_key: eval_summary_stats}) |
|
|
| |
| if train_case_stats and eval_case_stats: |
| |
| total_summary_stats = self._compute_total_stats( |
| [train_case_stats, eval_case_stats], hist_bins, hist_range |
| ) |
| stats_dict.update({FlStatistics.TOTAL_DATA: total_summary_stats}) |
|
|
| |
| stats = ExchangeObject(statistics=stats_dict) |
| if self.post_statistics_filters is not None: |
| for _filter in self.post_statistics_filters: |
| stats = _filter(stats, extra) |
|
|
| return stats |
| else: |
| raise ValueError("data_root not set!") |
|
|
| def _get_data_key_stats(self, data, data_key, hist_bins, hist_range, output_path=None): |
| analyzer = DataAnalyzer( |
| datalist={data_key: data}, |
| dataroot=self.workflow.dataset_dir, |
| hist_bins=hist_bins, |
| hist_range=hist_range, |
| output_path=output_path, |
| histogram_only=self.histogram_only, |
| ) |
|
|
| self.logger.info(f"{self.client_name} compute data statistics on {data_key}...") |
| all_stats = analyzer.get_all_case_stats(transform_list=self.data_stats_transform_list, key=data_key) |
|
|
| case_stats = all_stats[DataStatsKeys.BY_CASE] |
|
|
| summary_stats = { |
| FlStatistics.DATA_STATS: all_stats[DataStatsKeys.SUMMARY], |
| FlStatistics.DATA_COUNT: len(data), |
| FlStatistics.FAIL_COUNT: len(data) - len(case_stats), |
| |
| } |
|
|
| return summary_stats, case_stats |
|
|
| @staticmethod |
| def _compute_total_stats(case_stats_lists, hist_bins, hist_range): |
| |
| total_case_stats = [] |
| for case_stats_list in case_stats_lists: |
| total_case_stats += case_stats_list |
|
|
| summarizer = SegSummarizer( |
| "image", "label", average=True, do_ccp=True, hist_bins=hist_bins, hist_range=hist_range |
| ) |
| total_summary_stats = summarizer.summarize(total_case_stats) |
|
|
| summary_stats = { |
| FlStatistics.DATA_STATS: total_summary_stats, |
| FlStatistics.DATA_COUNT: len(total_case_stats), |
| FlStatistics.FAIL_COUNT: 0, |
| } |
|
|
| return summary_stats |
|
|
| def _add_config_files(self, config_files): |
| files = [] |
| if config_files: |
| if isinstance(config_files, str): |
| files.append(os.path.join(self.bundle_root, config_files)) |
| elif isinstance(config_files, list): |
| for file in config_files: |
| if isinstance(file, str): |
| files.append(os.path.join(self.bundle_root, file)) |
| else: |
| raise ValueError(f"Expected config file to be of type str but got {type(file)}: {file}") |
| else: |
| raise ValueError( |
| f"Expected config files to be of type str or list but got {type(config_files)}: {config_files}" |
| ) |
| return files |
|
|
|
|
| @require_pkg(pkg_name="ignite", version="0.4.10", version_checker=min_version) |
| class MonaiAlgo(ClientAlgo, MonaiAlgoStats): |
| """ |
| Implementation of ``ClientAlgo`` to allow federated learning with MONAI bundle configurations. |
| |
| Args: |
| bundle_root: directory path of the bundle. |
| local_epochs: number of local epochs to execute during each round of local training; defaults to 1. |
| send_weight_diff: whether to send weight differences rather than full weights; defaults to `True`. |
| config_train_filename: bundle training config path relative to bundle_root. can be a list of files. |
| defaults to "configs/train.json". only useful when `train_workflow` is None. |
| train_kwargs: other args of the `ConfigWorkflow` of train, except for `config_file`, `meta_file`, |
| `logging_file`, `workflow_type`. only useful when `train_workflow` is None. |
| config_evaluate_filename: bundle evaluation config path relative to bundle_root. can be a list of files. |
| if "default", ["configs/train.json", "configs/evaluate.json"] will be used. |
| this arg is only useful when `eval_workflow` is None. |
| eval_kwargs: other args of the `ConfigWorkflow` of evaluation, except for `config_file`, `meta_file`, |
| `logging_file`, `workflow_type`. only useful when `eval_workflow` is None. |
| config_filters_filename: filter configuration file. Can be a list of files; defaults to `None`. |
| disable_ckpt_loading: do not use any CheckpointLoader if defined in train/evaluate configs; defaults to `True`. |
| best_model_filepath: location of best model checkpoint; defaults "models/model.pt" relative to `bundle_root`. |
| final_model_filepath: location of final model checkpoint; defaults "models/model_final.pt" relative to `bundle_root`. |
| save_dict_key: If a model checkpoint contains several state dicts, |
| the one defined by `save_dict_key` will be returned by `get_weights`; defaults to "model". |
| If all state dicts should be returned, set `save_dict_key` to None. |
| data_stats_transform_list: transforms to apply for the data stats result. |
| eval_workflow_name: the workflow name corresponding to the "config_evaluate_filename", default to "train" |
| as the default "config_evaluate_filename" overrides the train workflow config. |
| this arg is only useful when `eval_workflow` is None. |
| train_workflow: the bundle workflow to execute training, if None, will create a `ConfigWorkflow` internally |
| based on `config_train_filename` and `train_kwargs`. |
| eval_workflow: the bundle workflow to execute evaluation, if None, will create a `ConfigWorkflow` internally |
| based on `config_evaluate_filename`, `eval_kwargs`, `eval_workflow_name`. |
| |
| """ |
|
|
| def __init__( |
| self, |
| bundle_root: str, |
| local_epochs: int = 1, |
| send_weight_diff: bool = True, |
| config_train_filename: str | list | None = "configs/train.json", |
| train_kwargs: dict | None = None, |
| config_evaluate_filename: str | list | None = "default", |
| eval_kwargs: dict | None = None, |
| config_filters_filename: str | list | None = None, |
| disable_ckpt_loading: bool = True, |
| best_model_filepath: str | None = "models/model.pt", |
| final_model_filepath: str | None = "models/model_final.pt", |
| save_dict_key: str | None = "model", |
| data_stats_transform_list: list | None = None, |
| eval_workflow_name: str = "train", |
| train_workflow: BundleWorkflow | None = None, |
| eval_workflow: BundleWorkflow | None = None, |
| ): |
| self.logger = logger |
| self.bundle_root = bundle_root |
| self.local_epochs = local_epochs |
| self.send_weight_diff = send_weight_diff |
| self.config_train_filename = config_train_filename |
| self.train_kwargs = {} if train_kwargs is None else train_kwargs |
| if config_evaluate_filename == "default": |
| |
| config_evaluate_filename = ["configs/train.json", "configs/evaluate.json"] |
| self.config_evaluate_filename = config_evaluate_filename |
| self.eval_kwargs = {} if eval_kwargs is None else eval_kwargs |
| self.config_filters_filename = config_filters_filename |
| self.disable_ckpt_loading = disable_ckpt_loading |
| self.model_filepaths = {ModelType.BEST_MODEL: best_model_filepath, ModelType.FINAL_MODEL: final_model_filepath} |
| self.save_dict_key = save_dict_key |
| self.data_stats_transform_list = data_stats_transform_list |
| self.eval_workflow_name = eval_workflow_name |
| self.train_workflow = None |
| self.eval_workflow = None |
| if train_workflow is not None: |
| if not isinstance(train_workflow, BundleWorkflow) or train_workflow.get_workflow_type() != "train": |
| raise ValueError( |
| f"train workflow must be BundleWorkflow and set type in {BundleWorkflow.supported_train_type}." |
| ) |
| self.train_workflow = train_workflow |
| if eval_workflow is not None: |
| |
| if not isinstance(eval_workflow, BundleWorkflow) or eval_workflow.get_workflow_type() is None: |
| raise ValueError("train workflow must be BundleWorkflow and set type.") |
| self.eval_workflow = eval_workflow |
| self.stats_sender = None |
|
|
| self.app_root = "" |
| self.filter_parser: ConfigParser | None = None |
| self.trainer: SupervisedTrainer | None = None |
| self.evaluator: SupervisedEvaluator | None = None |
| self.pre_filters = None |
| self.post_weight_filters = None |
| self.post_evaluate_filters = None |
| self.iter_of_start_time = 0 |
| self.global_weights: Mapping | None = None |
|
|
| self.phase = FlPhase.IDLE |
| self.client_name = None |
| self.dataset_root = None |
|
|
| def initialize(self, extra=None): |
| """ |
| Initialize routine to parse configuration files and extract main components such as trainer, evaluator, and filters. |
| |
| Args: |
| extra: Dict with additional information that should be provided by FL system, |
| i.e., `ExtraItems.CLIENT_NAME` and `ExtraItems.APP_ROOT`. |
| |
| """ |
| self._set_cuda_device() |
| if extra is None: |
| extra = {} |
| self.client_name = extra.get(ExtraItems.CLIENT_NAME, "noname") |
| timestamp = time.strftime("%Y%m%d_%H%M%S") |
| self.logger.info(f"Initializing {self.client_name} ...") |
| |
| self.app_root = extra.get(ExtraItems.APP_ROOT, "") |
| self.bundle_root = os.path.join(self.app_root, self.bundle_root) |
|
|
| if self.train_workflow is None and self.config_train_filename is not None: |
| config_train_files = self._add_config_files(self.config_train_filename) |
| |
| |
| if "run_name" not in self.train_kwargs: |
| self.train_kwargs["run_name"] = f"{self.client_name}_{timestamp}" |
| self.train_workflow = ConfigWorkflow( |
| config_file=config_train_files, |
| meta_file=None, |
| logging_file=None, |
| workflow_type="train", |
| **self.train_kwargs, |
| ) |
| if self.train_workflow is not None: |
| self.train_workflow.initialize() |
| self.train_workflow.bundle_root = self.bundle_root |
| self.train_workflow.max_epochs = self.local_epochs |
| if self.disable_ckpt_loading and isinstance(self.train_workflow, ConfigWorkflow): |
| disable_ckpt_loaders(parser=self.train_workflow.parser) |
| |
| self.train_workflow.initialize() |
| self.trainer = self.train_workflow.trainer |
| if not isinstance(self.trainer, SupervisedTrainer): |
| raise ValueError(f"trainer must be SupervisedTrainer, but got: {type(self.trainer)}.") |
|
|
| if self.eval_workflow is None and self.config_evaluate_filename is not None: |
| config_eval_files = self._add_config_files(self.config_evaluate_filename) |
| |
| |
| if "run_name" not in self.eval_kwargs: |
| self.eval_kwargs["run_name"] = f"{self.client_name}_{timestamp}" |
| self.eval_workflow = ConfigWorkflow( |
| config_file=config_eval_files, |
| meta_file=None, |
| logging_file=None, |
| workflow_type=self.eval_workflow_name, |
| **self.eval_kwargs, |
| ) |
| if self.eval_workflow is not None: |
| self.eval_workflow.initialize() |
| self.eval_workflow.bundle_root = self.bundle_root |
| if self.disable_ckpt_loading and isinstance(self.eval_workflow, ConfigWorkflow): |
| disable_ckpt_loaders(parser=self.eval_workflow.parser) |
| |
| self.eval_workflow.initialize() |
| self.evaluator = self.eval_workflow.evaluator |
| if not isinstance(self.evaluator, SupervisedEvaluator): |
| raise ValueError(f"evaluator must be SupervisedEvaluator, but got: {type(self.evaluator)}.") |
|
|
| config_filter_files = self._add_config_files(self.config_filters_filename) |
| self.filter_parser = ConfigParser() |
| if len(config_filter_files) > 0: |
| self.filter_parser.read_config(config_filter_files) |
|
|
| |
| self.stats_sender = extra.get(ExtraItems.STATS_SENDER, self.stats_sender) |
| if self.stats_sender is not None: |
| self.stats_sender.attach(self.trainer) |
| self.stats_sender.attach(self.evaluator) |
|
|
| |
| self.pre_filters = self.filter_parser.get_parsed_content( |
| FiltersType.PRE_FILTERS, default=ConfigItem(None, FiltersType.PRE_FILTERS) |
| ) |
| self.post_weight_filters = self.filter_parser.get_parsed_content( |
| FiltersType.POST_WEIGHT_FILTERS, default=ConfigItem(None, FiltersType.POST_WEIGHT_FILTERS) |
| ) |
| self.post_evaluate_filters = self.filter_parser.get_parsed_content( |
| FiltersType.POST_EVALUATE_FILTERS, default=ConfigItem(None, FiltersType.POST_EVALUATE_FILTERS) |
| ) |
| self.post_statistics_filters = self.filter_parser.get_parsed_content( |
| FiltersType.POST_STATISTICS_FILTERS, default=ConfigItem(None, FiltersType.POST_STATISTICS_FILTERS) |
| ) |
| self.logger.info(f"Initialized {self.client_name}.") |
|
|
| def train(self, data: ExchangeObject, extra: dict | None = None) -> None: |
| """ |
| Train on client's local data. |
| |
| Args: |
| data: `ExchangeObject` containing the current global model weights. |
| extra: Dict with additional information that can be provided by the FL system. |
| |
| """ |
|
|
| self._set_cuda_device() |
| if extra is None: |
| extra = {} |
| if not isinstance(data, ExchangeObject): |
| raise ValueError(f"expected data to be ExchangeObject but received {type(data)}") |
|
|
| if self.trainer is None: |
| raise ValueError("self.trainer should not be None.") |
| if self.pre_filters is not None: |
| for _filter in self.pre_filters: |
| data = _filter(data, extra) |
| self.phase = FlPhase.TRAIN |
| self.logger.info(f"Load {self.client_name} weights...") |
| local_var_dict = get_state_dict(self.trainer.network) |
| self.global_weights, n_converted = convert_global_weights( |
| global_weights=cast(dict, data.weights), local_var_dict=local_var_dict |
| ) |
| self._check_converted(data.weights, local_var_dict, n_converted) |
|
|
| |
| self.trainer.state.max_epochs = self.trainer.state.epoch + self.local_epochs |
| |
| self.iter_of_start_time = self.trainer.state.iteration |
|
|
| _, updated_keys, _ = copy_model_state(src=cast(Mapping, self.global_weights), dst=self.trainer.network) |
| if len(updated_keys) == 0: |
| self.logger.warning("No weights loaded!") |
| self.logger.info(f"Start {self.client_name} training...") |
| self.trainer.run() |
|
|
| def get_weights(self, extra=None): |
| """ |
| Returns the current weights of the model. |
| |
| Args: |
| extra: Dict with additional information that can be provided by the FL system. |
| |
| Returns: |
| return_weights: `ExchangeObject` containing current weights (default) |
| or load requested model type from disk (`ModelType.BEST_MODEL` or `ModelType.FINAL_MODEL`). |
| |
| """ |
|
|
| self._set_cuda_device() |
| if extra is None: |
| extra = {} |
|
|
| |
| self.phase = FlPhase.GET_WEIGHTS |
|
|
| if ExtraItems.MODEL_TYPE in extra: |
| model_type = extra.get(ExtraItems.MODEL_TYPE) |
| if not isinstance(model_type, ModelType): |
| raise ValueError( |
| f"Expected requested model type to be of type `ModelType` but received {type(model_type)}" |
| ) |
| if model_type in self.model_filepaths: |
| model_path = os.path.join(self.bundle_root, cast(str, self.model_filepaths[model_type])) |
| if not os.path.isfile(model_path): |
| raise ValueError(f"No best model checkpoint exists at {model_path}") |
| weights = torch.load(model_path, map_location="cpu") |
| |
| if isinstance(weights, dict) and self.save_dict_key in weights: |
| weights = weights.get(self.save_dict_key) |
| weigh_type: WeightType | None = WeightType.WEIGHTS |
| stats: dict = {} |
| self.logger.info(f"Returning {model_type} checkpoint weights from {model_path}.") |
| else: |
| raise ValueError( |
| f"Requested model type {model_type} not specified in `model_filepaths`: {self.model_filepaths}" |
| ) |
| else: |
| if self.trainer: |
| weights = get_state_dict(self.trainer.network) |
| |
| for k in weights.keys(): |
| weights[k] = weights[k].cpu() |
| weigh_type = WeightType.WEIGHTS |
| stats = self.trainer.get_stats() |
| |
| stats[FlStatistics.NUM_EXECUTED_ITERATIONS] = self.trainer.state.iteration - self.iter_of_start_time |
| |
| if self.send_weight_diff: |
| weights = compute_weight_diff(global_weights=self.global_weights, local_var_dict=weights) |
| weigh_type = WeightType.WEIGHT_DIFF |
| self.logger.info("Returning current weight differences.") |
| else: |
| self.logger.info("Returning current weights.") |
| else: |
| weights = None |
| weigh_type = None |
| stats = dict() |
|
|
| if not isinstance(stats, dict): |
| raise ValueError(f"stats is not a dict, {stats}") |
| return_weights = ExchangeObject( |
| weights=weights, |
| optim=None, |
| weight_type=weigh_type, |
| statistics=stats, |
| ) |
|
|
| |
| if self.post_weight_filters is not None: |
| for _filter in self.post_weight_filters: |
| return_weights = _filter(return_weights, extra) |
|
|
| return return_weights |
|
|
| def evaluate(self, data: ExchangeObject, extra: dict | None = None) -> ExchangeObject: |
| """ |
| Evaluate on client's local data. |
| |
| Args: |
| data: `ExchangeObject` containing the current global model weights. |
| extra: Dict with additional information that can be provided by the FL system. |
| |
| Returns: |
| return_metrics: `ExchangeObject` containing evaluation metrics. |
| |
| """ |
|
|
| self._set_cuda_device() |
| if extra is None: |
| extra = {} |
| if not isinstance(data, ExchangeObject): |
| raise ValueError(f"expected data to be ExchangeObject but received {type(data)}") |
|
|
| if self.evaluator is None: |
| raise ValueError("self.evaluator should not be None.") |
| if self.pre_filters is not None: |
| for _filter in self.pre_filters: |
| data = _filter(data, extra) |
|
|
| self.phase = FlPhase.EVALUATE |
| self.logger.info(f"Load {self.client_name} weights...") |
| local_var_dict = get_state_dict(self.evaluator.network) |
| global_weights, n_converted = convert_global_weights( |
| global_weights=cast(dict, data.weights), local_var_dict=local_var_dict |
| ) |
| self._check_converted(data.weights, local_var_dict, n_converted) |
|
|
| _, updated_keys, _ = copy_model_state(src=global_weights, dst=self.evaluator.network) |
| if len(updated_keys) == 0: |
| self.logger.warning("No weights loaded!") |
| self.logger.info(f"Start {self.client_name} evaluating...") |
| if isinstance(self.trainer, Trainer): |
| self.evaluator.run(self.trainer.state.epoch + 1) |
| else: |
| self.evaluator.run() |
| return_metrics = ExchangeObject(metrics=self.evaluator.state.metrics) |
|
|
| if self.post_evaluate_filters is not None: |
| for _filter in self.post_evaluate_filters: |
| return_metrics = _filter(return_metrics, extra) |
| return return_metrics |
|
|
| def abort(self, extra=None): |
| """ |
| Abort the training or evaluation. |
| Args: |
| extra: Dict with additional information that can be provided by the FL system. |
| """ |
| self.logger.info(f"Aborting {self.client_name} during {self.phase} phase.") |
| if isinstance(self.trainer, Trainer): |
| self.logger.info(f"Aborting {self.client_name} trainer...") |
| self.trainer.interrupt() |
| if isinstance(self.evaluator, Trainer): |
| self.logger.info(f"Aborting {self.client_name} evaluator...") |
| self.evaluator.interrupt() |
|
|
| def finalize(self, extra: dict | None = None) -> None: |
| """ |
| Finalize the training or evaluation. |
| Args: |
| extra: Dict with additional information that can be provided by the FL system. |
| """ |
| self.logger.info(f"Terminating {self.client_name} during {self.phase} phase.") |
| if isinstance(self.trainer, Trainer): |
| self.logger.info(f"Terminating {self.client_name} trainer...") |
| self.trainer.terminate() |
| if isinstance(self.evaluator, Trainer): |
| self.logger.info(f"Terminating {self.client_name} evaluator...") |
| self.evaluator.terminate() |
| if self.train_workflow is not None: |
| self.train_workflow.finalize() |
| if self.eval_workflow is not None: |
| self.eval_workflow.finalize() |
|
|
| def _check_converted(self, global_weights, local_var_dict, n_converted): |
| if n_converted == 0: |
| raise RuntimeError( |
| f"No global weights converted! Received weight dict keys are {list(global_weights.keys())}" |
| ) |
| else: |
| self.logger.info( |
| f"Converted {n_converted} global variables to match {len(local_var_dict)} local variables." |
| ) |
|
|
| def _set_cuda_device(self): |
| if dist.is_initialized(): |
| self.rank = int(os.environ["LOCAL_RANK"]) |
| torch.cuda.set_device(self.rank) |
|
|