| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| """An estimator class for training with TensorFlow on Amazon SageMaker.""" |
| from __future__ import absolute_import |
|
|
| import logging |
| from typing import Optional, Union, Dict |
|
|
| from packaging import version |
|
|
| from sagemaker import image_uris, s3, utils |
| from sagemaker.deprecations import renamed_kwargs |
| from sagemaker.estimator import Framework, EstimatorBase |
| import sagemaker.fw_utils as fw |
| from sagemaker.tensorflow import defaults |
| from sagemaker.tensorflow.model import TensorFlowModel |
| from sagemaker.transformer import Transformer |
| from sagemaker.vpc_utils import VPC_CONFIG_DEFAULT |
| from sagemaker.workflow import is_pipeline_variable |
| from sagemaker.tensorflow.training_compiler.config import TrainingCompilerConfig |
| from sagemaker.workflow.entities import PipelineVariable |
|
|
| logger = logging.getLogger("sagemaker") |
|
|
|
|
| class TensorFlow(Framework): |
| """Handle end-to-end training and deployment of user-provided TensorFlow code.""" |
|
|
| _framework_name = "tensorflow" |
|
|
| _HIGHEST_LEGACY_MODE_ONLY_VERSION = version.Version("1.10.0") |
| _HIGHEST_PYTHON_2_VERSION = version.Version("2.1.1") |
|
|
| def __init__( |
| self, |
| py_version: Optional[str] = None, |
| framework_version: Optional[str] = None, |
| model_dir: Optional[Union[str, PipelineVariable]] = None, |
| image_uri: Optional[Union[str, PipelineVariable]] = None, |
| distribution: Optional[Dict[str, str]] = None, |
| compiler_config: Optional[TrainingCompilerConfig] = None, |
| **kwargs, |
| ): |
| """Initialize a ``TensorFlow`` estimator. |
| |
| Args: |
| py_version (str): Python version you want to use for executing your model training |
| code. Defaults to ``None``. Required unless ``image_uri`` is provided. |
| framework_version (str): TensorFlow version you want to use for executing your model |
| training code. Defaults to ``None``. Required unless ``image_uri`` is provided. |
| List of supported versions: |
| https://github.com/aws/sagemaker-python-sdk#tensorflow-sagemaker-estimators. |
| model_dir (str or PipelineVariable): S3 location where the checkpoint data and models |
| can be exported to during training (default: None). It will be passed in the |
| training script as one of the command line arguments. If not specified, |
| one is provided based on your training configuration: |
| |
| * *distributed training with SMDistributed or MPI with Horovod* - ``/opt/ml/model`` |
| * *single-machine training or distributed training without MPI* - \ |
| ``s3://{output_path}/model`` |
| * *Local Mode with local sources (file:// instead of s3://)* - \ |
| ``/opt/ml/shared/model`` |
| |
| To disable having ``model_dir`` passed to your training script, |
| set ``model_dir=False``. |
| image_uri (str or PipelineVariable): If specified, the estimator will use this image |
| for training and hosting, instead of selecting the appropriate SageMaker official |
| image based on framework_version and py_version. |
| It can be an ECR url or dockerhub image and tag. |
| |
| Examples: |
| 123.dkr.ecr.us-west-2.amazonaws.com/my-custom-image:1.0 |
| custom-image:latest. |
| |
| If ``framework_version`` or ``py_version`` are ``None``, then |
| ``image_uri`` is required. If also ``None``, then a ``ValueError`` |
| will be raised. |
| distribution (dict): A dictionary with information on how to run distributed training |
| (default: None). Currently, the following are supported: |
| distributed training with parameter servers, SageMaker Distributed (SMD) Data |
| and Model Parallelism, and MPI. SMD Model Parallelism can only be used with MPI. |
| |
| **To enable the SageMaker distributed data parallelism:** |
| |
| .. code:: python |
| |
| { "smdistributed": { "dataparallel": { "enabled": True } } } |
| |
| .. seealso:: |
| |
| To learn more, see :ref:`sdp_api_docs_toc`. |
| |
| **To enable the SageMaker distributed model parallelism:** |
| |
| .. code:: python |
| |
| { |
| "smdistributed": { |
| "modelparallel": { |
| "enabled":True, |
| "parameters": { |
| "partitions": 2, |
| "microbatches": 4, |
| "placement_strategy": "spread", |
| "pipeline": "interleaved", |
| "optimize": "speed", |
| "ddp": True, |
| } |
| }, |
| "mpi": { |
| "enabled" : True, |
| "processes_per_host" : 8, |
| } |
| } |
| |
| .. note:: |
| |
| The SageMaker distributed model parallel library internally uses MPI. |
| In order to use model parallelism, MPI also must be enabled. |
| |
| .. seealso:: |
| |
| To learn more, see :ref:`smp_api_docs_toc`. |
| |
| .. seealso:: |
| |
| To find a complete list of parameters for SageMaker model parallelism, |
| see :ref:`sm-sdk-modelparallel-general`. |
| |
| **To enable MPI:** |
| |
| .. code:: python |
| |
| { |
| "mpi": { |
| "enabled": True |
| } |
| } |
| |
| To learn more, see `Training with Horovod |
| <https://sagemaker.readthedocs.io/en/stable/frameworks/tensorflow/using_tf.html#training-with-horovod>`_. |
| |
| **To enable parameter server:** |
| |
| .. code:: python |
| |
| { |
| "parameter_server": { |
| "enabled": True |
| } |
| } |
| |
| To learn more, see `Training with parameter servers |
| <https://sagemaker.readthedocs.io/en/stable/frameworks/tensorflow/using_tf.html#training-with-parameter-servers>`_. |
| compiler_config (:class:`~sagemaker.tensorflow.TrainingCompilerConfig`): |
| Configures SageMaker Training Compiler to accelerate training. |
| |
| **kwargs: Additional kwargs passed to the Framework constructor. |
| |
| .. tip:: |
| |
| You can find additional parameters for initializing this class at |
| :class:`~sagemaker.estimator.Framework` and |
| :class:`~sagemaker.estimator.EstimatorBase`. |
| """ |
| distribution = renamed_kwargs("distributions", "distribution", distribution, kwargs) |
| instance_type = renamed_kwargs( |
| "train_instance_type", "instance_type", kwargs.get("instance_type"), kwargs |
| ) |
| fw.validate_version_or_image_args(framework_version, py_version, image_uri) |
| if py_version == "py2": |
| logger.warning( |
| fw.python_deprecation_warning(self._framework_name, defaults.LATEST_PY2_VERSION) |
| ) |
| self.framework_version = framework_version |
| self.py_version = py_version |
| self.instance_type = instance_type |
|
|
| if "enable_sagemaker_metrics" not in kwargs: |
| |
| if framework_version and version.Version(framework_version) >= version.Version("1.15"): |
| kwargs["enable_sagemaker_metrics"] = True |
|
|
| super(TensorFlow, self).__init__(image_uri=image_uri, **kwargs) |
| if distribution is not None: |
| distribution = fw.validate_distribution( |
| distribution, |
| self.instance_groups, |
| self._framework_name, |
| framework_version, |
| py_version, |
| image_uri, |
| kwargs, |
| ) |
| self.model_dir = model_dir |
| self.distribution = distribution or {} |
|
|
| self._validate_args(py_version=py_version) |
| if compiler_config is not None: |
| if not isinstance(compiler_config, TrainingCompilerConfig): |
| error_string = ( |
| f"Expected instance of type {TrainingCompilerConfig}" |
| f"for argument compiler_config. " |
| f"Instead got {type(compiler_config)}" |
| ) |
| raise ValueError(error_string) |
| if compiler_config: |
| compiler_config.validate(self) |
| self.compiler_config = compiler_config |
|
|
| def _validate_args(self, py_version): |
| """Placeholder docstring""" |
|
|
| if py_version == "py2" and self._only_python_3_supported(): |
| msg = ( |
| "Python 2 containers are only available with {} and lower versions. " |
| "Please use a Python 3 container.".format(defaults.LATEST_PY2_VERSION) |
| ) |
| raise AttributeError(msg) |
|
|
| if self.image_uri is None and self._only_legacy_mode_supported(): |
| legacy_image_uri = image_uris.retrieve( |
| "tensorflow", |
| self.sagemaker_session.boto_region_name, |
| instance_type=self.instance_type, |
| version=self.framework_version, |
| py_version=self.py_version, |
| image_scope="training", |
| ) |
|
|
| msg = ( |
| "TF {} supports only legacy mode. Please supply the image URI directly with " |
| "'image_uri={}' and set 'model_dir=False'. If you are using any legacy parameters " |
| "(training_steps, evaluation_steps, checkpoint_path, requirements_file), " |
| "make sure to pass them directly as hyperparameters instead. For more, see " |
| "https://sagemaker.readthedocs.io/en/v2.0.0.rc0/frameworks/tensorflow/upgrade_from_legacy.html." |
| ).format(self.framework_version, legacy_image_uri) |
|
|
| raise ValueError(msg) |
|
|
| def _only_legacy_mode_supported(self): |
| """Placeholder docstring""" |
| return version.Version(self.framework_version) <= self._HIGHEST_LEGACY_MODE_ONLY_VERSION |
|
|
| def _only_python_3_supported(self): |
| """Placeholder docstring""" |
| if not self.framework_version: |
| return False |
| return version.Version(self.framework_version) > self._HIGHEST_PYTHON_2_VERSION |
|
|
| @classmethod |
| def _prepare_init_params_from_job_description(cls, job_details, model_channel_name=None): |
| """Convert the job description to init params that can be handled by the class constructor |
| |
| Args: |
| job_details: the returned job details from a describe_training_job API call. |
| |
| Returns: |
| dictionary: The transformed init_params |
| |
| """ |
| init_params = super(TensorFlow, cls)._prepare_init_params_from_job_description( |
| job_details, model_channel_name |
| ) |
|
|
| image_uri = init_params.pop("image_uri") |
| framework, py_version, tag, script_mode = fw.framework_name_from_image(image_uri) |
|
|
| if not framework: |
| |
| |
| init_params["image_uri"] = image_uri |
| return init_params |
|
|
| model_dir = init_params["hyperparameters"].pop("model_dir", None) |
| if model_dir: |
| init_params["model_dir"] = model_dir |
| elif script_mode is None: |
| init_params["model_dir"] = False |
|
|
| init_params["py_version"] = py_version |
|
|
| |
| |
| |
| |
| init_params["framework_version"] = ( |
| "1.4" if tag == "1.0" else fw.framework_version_from_tag(tag) |
| ) |
|
|
| |
| if not script_mode: |
| init_params["image_uri"] = image_uri |
|
|
| if framework != cls._framework_name: |
| raise ValueError( |
| "Training job: {} didn't use image for requested framework".format( |
| job_details["TrainingJobName"] |
| ) |
| ) |
|
|
| return init_params |
|
|
| def create_model( |
| self, |
| role=None, |
| vpc_config_override=VPC_CONFIG_DEFAULT, |
| entry_point=None, |
| source_dir=None, |
| dependencies=None, |
| **kwargs, |
| ): |
| """Creates ``TensorFlowModel`` object to be used for creating SageMaker model entities. |
| |
| This can be done by deploying it to a SageMaker endpoint, |
| or starting SageMaker Batch Transform jobs. |
| |
| Args: |
| role (str): The ``TensorFlowModel``, which is also used during transform jobs. |
| If not specified, the role from the Estimator is used. |
| vpc_config_override (dict[str, list[str]]): Optional override for VpcConfig set on the |
| model. Default: use subnets and security groups from this Estimator. |
| |
| * 'Subnets' (list[str]): List of subnet ids. |
| * 'SecurityGroupIds' (list[str]): List of security group ids. |
| |
| entry_point (str): Path (absolute or relative) to the local Python source file which |
| should be executed as the entry point to training. If ``source_dir`` is specified, |
| then ``entry_point`` must point to a file located at the root of ``source_dir``. |
| If not specified and ``endpoint_type`` is 'tensorflow-serving', |
| no entry point is used. If ``endpoint_type`` is also ``None``, |
| then the training entry point is used. |
| source_dir (str): Path (absolute or relative or an S3 URI) to a directory with any other |
| serving source code dependencies aside from the entry point file (default: None). |
| dependencies (list[str]): A list of paths to directories (absolute or relative) with |
| any additional libraries that will be exported to the container (default: None). |
| **kwargs: Additional kwargs passed to |
| :class:`~sagemaker.tensorflow.model.TensorFlowModel`. |
| |
| Returns: |
| sagemaker.tensorflow.model.TensorFlowModel: A ``TensorFlowModel`` object. |
| See :class:`~sagemaker.tensorflow.model.TensorFlowModel` for full details. |
| """ |
| kwargs["name"] = self._get_or_create_name(kwargs.get("name")) |
|
|
| if "image_uri" not in kwargs: |
| kwargs["image_uri"] = self.image_uri |
|
|
| if "enable_network_isolation" not in kwargs: |
| kwargs["enable_network_isolation"] = self.enable_network_isolation() |
|
|
| return TensorFlowModel( |
| model_data=self.model_data, |
| role=role or self.role, |
| container_log_level=self.container_log_level, |
| framework_version=self.framework_version, |
| sagemaker_session=self.sagemaker_session, |
| vpc_config=self.get_vpc_config(vpc_config_override), |
| entry_point=entry_point, |
| source_dir=source_dir, |
| dependencies=dependencies, |
| **kwargs, |
| ) |
|
|
| def hyperparameters(self): |
| """Return hyperparameters used by your custom TensorFlow code during model training.""" |
| hyperparameters = super(TensorFlow, self).hyperparameters() |
| additional_hyperparameters = self._distribution_configuration(self.distribution) |
|
|
| if self.model_dir is not False: |
| self.model_dir = self.model_dir or self._default_s3_path( |
| "model", mpi=additional_hyperparameters.get(self.LAUNCH_MPI_ENV_NAME, False) |
| ) |
| additional_hyperparameters["model_dir"] = self.model_dir |
|
|
| hyperparameters.update( |
| EstimatorBase._json_encode_hyperparameters(additional_hyperparameters) |
| ) |
|
|
| if self.compiler_config: |
| training_compiler_hyperparameters = self.compiler_config._to_hyperparameter_dict() |
| hyperparameters.update( |
| EstimatorBase._json_encode_hyperparameters(training_compiler_hyperparameters) |
| ) |
|
|
| return hyperparameters |
|
|
| def _default_s3_path(self, directory, mpi=False): |
| """Placeholder docstring""" |
| local_code = utils.get_config_value("local.local_code", self.sagemaker_session.config) |
| if self.sagemaker_session.local_mode and local_code: |
| return "/opt/ml/shared/{}".format(directory) |
| if mpi: |
| return "/opt/ml/model" |
| if self._current_job_name: |
| if is_pipeline_variable(self.output_path): |
| output_path = "s3://{}".format(self.sagemaker_session.default_bucket()) |
| return s3.s3_path_join(output_path, self._current_job_name, directory) |
| return s3.s3_path_join(self.output_path, self._current_job_name, directory) |
| return None |
|
|
| def _validate_and_set_debugger_configs(self): |
| """Disable Debugger Hook Config for ParameterServer (PS) as it is not supported in smdebug. |
| |
| Else, set default HookConfig |
| """ |
| super(TensorFlow, self)._validate_and_set_debugger_configs() |
| ps_enabled = "parameter_server" in self.distribution and self.distribution[ |
| "parameter_server" |
| ].get("enabled", False) |
| if ps_enabled: |
| if self.debugger_hook_config is not None or self.debugger_rule_configs is not None: |
| logger.info( |
| "Amazon SageMaker Debugger does not currently support " |
| "Parameter Server distribution" |
| ) |
| self.debugger_hook_config = None |
| self.debugger_rule_configs = None |
|
|
| def transformer( |
| self, |
| instance_count, |
| instance_type, |
| strategy=None, |
| assemble_with=None, |
| output_path=None, |
| output_kms_key=None, |
| accept=None, |
| env=None, |
| max_concurrent_transforms=None, |
| max_payload=None, |
| tags=None, |
| role=None, |
| volume_kms_key=None, |
| entry_point=None, |
| vpc_config_override=VPC_CONFIG_DEFAULT, |
| enable_network_isolation=None, |
| model_name=None, |
| ): |
| """Return a ``Transformer`` that uses a SageMaker Model based on the training job. |
| |
| It reuses the SageMaker Session and base job name used by the Estimator. |
| |
| Args: |
| instance_count (int): Number of EC2 instances to use. |
| instance_type (str): Type of EC2 instance to use, for example, 'ml.c4.xlarge'. |
| strategy (str): The strategy used to decide how to batch records in a single request |
| (default: None). Valid values: 'MultiRecord' and 'SingleRecord'. |
| assemble_with (str): How the output is assembled (default: None). Valid values: 'Line' |
| or 'None'. |
| output_path (str): S3 location for saving the transform result. If not specified, |
| results are stored to a default bucket. |
| output_kms_key (str): Optional. KMS key ID for encrypting the transform output |
| (default: None). |
| accept (str): The accept header passed by the client to |
| the inference endpoint. If it is supported by the endpoint, |
| it will be the format of the batch transform output. |
| env (dict): Environment variables to be set for use during the transform job |
| (default: None). |
| max_concurrent_transforms (int): The maximum number of HTTP requests to be made to |
| each individual transform container at one time. |
| max_payload (int): Maximum size of the payload in a single HTTP request to the |
| container in MB. |
| tags (list[dict]): List of tags for labeling a transform job. If none specified, then |
| the tags used for the training job are used for the transform job. |
| role (str): The IAM Role ARN for the ``TensorFlowModel``, which is also used |
| during transform jobs. If not specified, the role from the Estimator is used. |
| volume_kms_key (str): Optional. KMS key ID for encrypting the volume attached to the ML |
| compute instance (default: None). |
| entry_point (str): Path (absolute or relative) to the local Python source file which |
| should be executed as the entry point to training. If ``source_dir`` is specified, |
| then ``entry_point`` must point to a file located at the root of ``source_dir``. |
| If not specified and ``endpoint_type`` is 'tensorflow-serving', |
| no entry point is used. If ``endpoint_type`` is also ``None``, |
| then the training entry point is used. |
| vpc_config_override (dict[str, list[str]]): Optional override for |
| the VpcConfig set on the model. |
| Default: use subnets and security groups from this Estimator. |
| |
| * 'Subnets' (list[str]): List of subnet ids. |
| * 'SecurityGroupIds' (list[str]): List of security group ids. |
| |
| enable_network_isolation (bool): Specifies whether container will |
| run in network isolation mode. Network isolation mode restricts |
| the container access to outside networks (such as the internet). |
| The container does not make any inbound or outbound network |
| calls. If True, a channel named "code" will be created for any |
| user entry script for inference. Also known as Internet-free mode. |
| If not specified, this setting is taken from the estimator's |
| current configuration. |
| model_name (str): Name to use for creating an Amazon SageMaker |
| model. If not specified, the estimator generates a default job name |
| based on the training image name and current timestamp. |
| """ |
| role = role or self.role |
| model_name = self._get_or_create_name(model_name) |
|
|
| if self.latest_training_job is None: |
| logger.warning( |
| "No finished training job found associated with this estimator. Please make sure " |
| "this estimator is only used for building workflow config" |
| ) |
| return Transformer( |
| model_name, |
| instance_count, |
| instance_type, |
| strategy=strategy, |
| assemble_with=assemble_with, |
| output_path=output_path, |
| output_kms_key=output_kms_key, |
| accept=accept, |
| max_concurrent_transforms=max_concurrent_transforms, |
| max_payload=max_payload, |
| env=env or {}, |
| tags=tags, |
| base_transform_job_name=self.base_job_name, |
| volume_kms_key=volume_kms_key, |
| sagemaker_session=self.sagemaker_session, |
| ) |
|
|
| if enable_network_isolation is None: |
| enable_network_isolation = self.enable_network_isolation() |
|
|
| model = self.create_model( |
| role=role, |
| vpc_config_override=vpc_config_override, |
| entry_point=entry_point, |
| enable_network_isolation=enable_network_isolation, |
| name=model_name, |
| ) |
|
|
| return model.transformer( |
| instance_count, |
| instance_type, |
| strategy=strategy, |
| assemble_with=assemble_with, |
| output_path=output_path, |
| output_kms_key=output_kms_key, |
| accept=accept, |
| env=env, |
| max_concurrent_transforms=max_concurrent_transforms, |
| max_payload=max_payload, |
| tags=tags, |
| volume_kms_key=volume_kms_key, |
| ) |
|
|