| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| """Placeholder docstring""" |
| from __future__ import absolute_import |
|
|
| import logging |
| from typing import Optional, Union, List, Dict |
|
|
| import sagemaker |
| from sagemaker import image_uris, ModelMetrics |
| from sagemaker.deserializers import JSONDeserializer |
| from sagemaker.drift_check_baselines import DriftCheckBaselines |
| from sagemaker.fw_utils import ( |
| model_code_key_prefix, |
| validate_version_or_image_args, |
| ) |
| from sagemaker.metadata_properties import MetadataProperties |
| from sagemaker.model import FrameworkModel, MODEL_SERVER_WORKERS_PARAM_NAME |
| from sagemaker.predictor import Predictor |
| from sagemaker.serializers import JSONSerializer |
| from sagemaker.session import Session |
| from sagemaker.utils import to_string |
| from sagemaker.workflow import is_pipeline_variable |
| from sagemaker.workflow.entities import PipelineVariable |
|
|
| logger = logging.getLogger("sagemaker") |
|
|
|
|
| class HuggingFacePredictor(Predictor): |
| """A Predictor for inference against Hugging Face Endpoints. |
| |
| This is able to serialize Python lists, dictionaries, and numpy arrays to |
| multidimensional tensors for Hugging Face inference. |
| """ |
|
|
| def __init__( |
| self, |
| endpoint_name, |
| sagemaker_session=None, |
| serializer=JSONSerializer(), |
| deserializer=JSONDeserializer(), |
| ): |
| """Initialize an ``HuggingFacePredictor``. |
| |
| Args: |
| endpoint_name (str): The name of the endpoint to perform inference |
| on. |
| sagemaker_session (sagemaker.session.Session): Session object that |
| manages interactions with Amazon SageMaker APIs and any other |
| AWS services needed. If not specified, the estimator creates one |
| using the default AWS configuration chain. |
| serializer (sagemaker.serializers.BaseSerializer): Optional. Default |
| serializes input data to .npy format. Handles lists and numpy |
| arrays. |
| deserializer (sagemaker.deserializers.BaseDeserializer): Optional. |
| Default parses the response from .npy format to numpy array. |
| """ |
| super(HuggingFacePredictor, self).__init__( |
| endpoint_name, |
| sagemaker_session, |
| serializer=serializer, |
| deserializer=deserializer, |
| ) |
|
|
|
|
| def _validate_pt_tf_versions(pytorch_version, tensorflow_version, image_uri): |
| """Placeholder docstring""" |
|
|
| if image_uri is not None: |
| return |
|
|
| if tensorflow_version is not None and pytorch_version is not None: |
| raise ValueError( |
| "tensorflow_version and pytorch_version are both not None. " |
| "Specify only tensorflow_version or pytorch_version." |
| ) |
| if tensorflow_version is None and pytorch_version is None: |
| raise ValueError( |
| "tensorflow_version and pytorch_version are both None. " |
| "Specify either tensorflow_version or pytorch_version." |
| ) |
|
|
|
|
| def fetch_framework_and_framework_version(tensorflow_version, pytorch_version): |
| """Function to check the framework used in HuggingFace class""" |
|
|
| if tensorflow_version is not None: |
| return ("tensorflow", tensorflow_version) |
| return ("pytorch", pytorch_version) |
|
|
|
|
| class HuggingFaceModel(FrameworkModel): |
| """A Hugging Face SageMaker ``Model`` that can be deployed to a SageMaker ``Endpoint``.""" |
|
|
| _framework_name = "huggingface" |
|
|
| def __init__( |
| self, |
| role: str, |
| model_data: Optional[Union[str, PipelineVariable]] = None, |
| entry_point: Optional[str] = None, |
| transformers_version: Optional[str] = None, |
| tensorflow_version: Optional[str] = None, |
| pytorch_version: Optional[str] = None, |
| py_version: Optional[str] = None, |
| image_uri: Optional[Union[str, PipelineVariable]] = None, |
| predictor_cls: callable = HuggingFacePredictor, |
| model_server_workers: Optional[Union[int, PipelineVariable]] = None, |
| **kwargs, |
| ): |
| """Initialize a HuggingFaceModel. |
| |
| Args: |
| model_data (str or PipelineVariable): The Amazon S3 location of a SageMaker |
| model data ``.tar.gz`` file. |
| role (str): An AWS IAM role specified with either the name or full ARN. The Amazon |
| SageMaker training jobs and APIs that create Amazon SageMaker |
| endpoints use this role to access training data and model |
| artifacts. After the endpoint is created, the inference code |
| might use the IAM role, if it needs to access an AWS resource. |
| entry_point (str): The absolute or relative path to the Python source |
| file that should be executed as the entry point to model |
| hosting. If ``source_dir`` is specified, then ``entry_point`` |
| must point to a file located at the root of ``source_dir``. |
| Defaults to None. |
| transformers_version (str): Transformers version you want to use for |
| executing your model training code. Defaults to None. Required |
| unless ``image_uri`` is provided. |
| tensorflow_version (str): TensorFlow version you want to use for |
| executing your inference code. Defaults to ``None``. Required unless |
| ``pytorch_version`` is provided. List of supported versions: |
| https://github.com/aws/sagemaker-python-sdk#huggingface-sagemaker-estimators. |
| pytorch_version (str): PyTorch version you want to use for |
| executing your inference code. Defaults to ``None``. Required unless |
| ``tensorflow_version`` is provided. List of supported versions: |
| https://github.com/aws/sagemaker-python-sdk#huggingface-sagemaker-estimators. |
| py_version (str): Python version you want to use for executing your |
| model training code. Defaults to ``None``. Required unless |
| ``image_uri`` is provided. |
| image_uri (str or PipelineVariable): A Docker image URI. Defaults to None. |
| If not specified, a default image for PyTorch will be used. If ``framework_version`` |
| or ``py_version`` are ``None``, then ``image_uri`` is required. If |
| also ``None``, then a ``ValueError`` will be raised. |
| predictor_cls (callable[str, sagemaker.session.Session]): A function |
| to call to create a predictor with an endpoint name and |
| SageMaker ``Session``. If specified, ``deploy()`` returns the |
| result of invoking this function on the created endpoint name. |
| model_server_workers (int or PipelineVariable): Optional. The number of |
| worker processes used by the inference server. If None, server will use one |
| worker per vCPU. |
| **kwargs: Keyword arguments passed to the superclass |
| :class:`~sagemaker.model.FrameworkModel` and, subsequently, its |
| superclass :class:`~sagemaker.model.Model`. |
| |
| .. tip:: |
| |
| You can find additional parameters for initializing this class at |
| :class:`~sagemaker.model.FrameworkModel` and |
| :class:`~sagemaker.model.Model`. |
| """ |
| validate_version_or_image_args(transformers_version, py_version, image_uri) |
| _validate_pt_tf_versions( |
| pytorch_version=pytorch_version, |
| tensorflow_version=tensorflow_version, |
| image_uri=image_uri, |
| ) |
| if py_version == "py2": |
| raise ValueError("py2 is not supported with HuggingFace images") |
| self.framework_version = transformers_version |
| self.pytorch_version = pytorch_version |
| self.tensorflow_version = tensorflow_version |
| self.py_version = py_version |
|
|
| super(HuggingFaceModel, self).__init__( |
| model_data, image_uri, role, entry_point, predictor_cls=predictor_cls, **kwargs |
| ) |
| self.sagemaker_session = self.sagemaker_session or Session() |
|
|
| self.model_server_workers = model_server_workers |
|
|
| |
| |
| |
| |
| def deploy( |
| self, |
| initial_instance_count=None, |
| instance_type=None, |
| serializer=None, |
| deserializer=None, |
| accelerator_type=None, |
| endpoint_name=None, |
| tags=None, |
| kms_key=None, |
| wait=True, |
| data_capture_config=None, |
| async_inference_config=None, |
| serverless_inference_config=None, |
| **kwargs, |
| ): |
| """Deploy this ``Model`` to an ``Endpoint`` and optionally return a ``Predictor``. |
| |
| Create a SageMaker ``Model`` and ``EndpointConfig``, and deploy an |
| ``Endpoint`` from this ``Model``. If ``self.predictor_cls`` is not None, |
| this method returns a the result of invoking ``self.predictor_cls`` on |
| the created endpoint name. |
| |
| The name of the created model is accessible in the ``name`` field of |
| this ``Model`` after deploy returns |
| |
| The name of the created endpoint is accessible in the |
| ``endpoint_name`` field of this ``Model`` after deploy returns. |
| |
| Args: |
| initial_instance_count (int): The initial number of instances to run |
| in the ``Endpoint`` created from this ``Model``. If not using |
| serverless inference, then it need to be a number larger or equals |
| to 1 (default: None) |
| instance_type (str): The EC2 instance type to deploy this Model to. |
| For example, 'ml.p2.xlarge', or 'local' for local mode. If not using |
| serverless inference, then it is required to deploy a model. |
| (default: None) |
| serializer (:class:`~sagemaker.serializers.BaseSerializer`): A |
| serializer object, used to encode data for an inference endpoint |
| (default: None). If ``serializer`` is not None, then |
| ``serializer`` will override the default serializer. The |
| default serializer is set by the ``predictor_cls``. |
| deserializer (:class:`~sagemaker.deserializers.BaseDeserializer`): A |
| deserializer object, used to decode data from an inference |
| endpoint (default: None). If ``deserializer`` is not None, then |
| ``deserializer`` will override the default deserializer. The |
| default deserializer is set by the ``predictor_cls``. |
| accelerator_type (str): Type of Elastic Inference accelerator to |
| deploy this model for model loading and inference, for example, |
| 'ml.eia1.medium'. If not specified, no Elastic Inference |
| accelerator will be attached to the endpoint. For more |
| information: |
| https://docs.aws.amazon.com/sagemaker/latest/dg/ei.html |
| endpoint_name (str): The name of the endpoint to create (default: |
| None). If not specified, a unique endpoint name will be created. |
| tags (List[dict[str, str]]): The list of tags to attach to this |
| specific endpoint. |
| kms_key (str): The ARN of the KMS key that is used to encrypt the |
| data on the storage volume attached to the instance hosting the |
| endpoint. |
| wait (bool): Whether the call should wait until the deployment of |
| this model completes (default: True). |
| data_capture_config (sagemaker.model_monitor.DataCaptureConfig): Specifies |
| configuration related to Endpoint data capture for use with |
| Amazon SageMaker Model Monitoring. Default: None. |
| async_inference_config (sagemaker.model_monitor.AsyncInferenceConfig): Specifies |
| configuration related to async endpoint. Use this configuration when trying |
| to create async endpoint and make async inference. If empty config object |
| passed through, will use default config to deploy async endpoint. Deploy a |
| real-time endpoint if it's None. (default: None) |
| serverless_inference_config (sagemaker.serverless.ServerlessInferenceConfig): |
| Specifies configuration related to serverless endpoint. Use this configuration |
| when trying to create serverless endpoint and make serverless inference. If |
| empty object passed through, will use pre-defined values in |
| ``ServerlessInferenceConfig`` class to deploy serverless endpoint. Deploy an |
| instance based endpoint if it's None. (default: None) |
| Raises: |
| ValueError: If arguments combination check failed in these circumstances: |
| - If no role is specified or |
| - If serverless inference config is not specified and instance type and instance |
| count are also not specified or |
| - If a wrong type of object is provided as serverless inference config or async |
| inference config |
| Returns: |
| callable[string, sagemaker.session.Session] or None: Invocation of |
| ``self.predictor_cls`` on the created endpoint name, if ``self.predictor_cls`` |
| is not None. Otherwise, return None. |
| """ |
|
|
| if not self.image_uri and instance_type is not None and instance_type.startswith("ml.inf"): |
| self.image_uri = self.serving_image_uri( |
| region_name=self.sagemaker_session.boto_session.region_name, |
| instance_type=instance_type, |
| ) |
|
|
| return super(HuggingFaceModel, self).deploy( |
| initial_instance_count, |
| instance_type, |
| serializer, |
| deserializer, |
| accelerator_type, |
| endpoint_name, |
| tags, |
| kms_key, |
| wait, |
| data_capture_config, |
| async_inference_config, |
| serverless_inference_config, |
| ) |
|
|
| def register( |
| self, |
| content_types: List[Union[str, PipelineVariable]], |
| response_types: List[Union[str, PipelineVariable]], |
| inference_instances: Optional[List[Union[str, PipelineVariable]]] = None, |
| transform_instances: Optional[List[Union[str, PipelineVariable]]] = None, |
| model_package_name: Optional[Union[str, PipelineVariable]] = None, |
| model_package_group_name: Optional[Union[str, PipelineVariable]] = None, |
| image_uri: Optional[Union[str, PipelineVariable]] = None, |
| model_metrics: Optional[ModelMetrics] = None, |
| metadata_properties: Optional[MetadataProperties] = None, |
| marketplace_cert: bool = False, |
| approval_status: Optional[Union[str, PipelineVariable]] = None, |
| description: Optional[str] = None, |
| drift_check_baselines: Optional[DriftCheckBaselines] = None, |
| customer_metadata_properties: Optional[Dict[str, Union[str, PipelineVariable]]] = None, |
| domain: Optional[Union[str, PipelineVariable]] = None, |
| sample_payload_url: Optional[Union[str, PipelineVariable]] = None, |
| task: Optional[Union[str, PipelineVariable]] = None, |
| framework: Optional[Union[str, PipelineVariable]] = None, |
| framework_version: Optional[Union[str, PipelineVariable]] = None, |
| nearest_model_name: Optional[Union[str, PipelineVariable]] = None, |
| data_input_configuration: Optional[Union[str, PipelineVariable]] = None, |
| ): |
| """Creates a model package for creating SageMaker models or listing on Marketplace. |
| |
| Args: |
| content_types (list[str] or list[PipelineVariable]): The supported MIME types |
| for the input data. |
| response_types (list[str] or list[PipelineVariable]): The supported MIME types |
| for the output data. |
| inference_instances (list[str] or list[PipelineVariable]): A list of the instance |
| types that are used to generate inferences in real-time (default: None). |
| transform_instances (list[str] or list[PipelineVariable]): A list of the instance types |
| on which a transformation job can be run or on which an endpoint can be deployed |
| (default: None). |
| model_package_name (str or PipelineVariable): Model Package name, exclusive to |
| `model_package_group_name`, using `model_package_name` makes the Model Package |
| un-versioned. Defaults to ``None``. |
| model_package_group_name (str or PipelineVariable): Model Package Group name, |
| exclusive to `model_package_name`, using `model_package_group_name` makes the |
| Model Package versioned. Defaults to ``None``. |
| image_uri (str or PipelineVariable): Inference image URI for the container. Model class' |
| self.image will be used if it is None. Defaults to ``None``. |
| model_metrics (ModelMetrics): ModelMetrics object. Defaults to ``None``. |
| metadata_properties (MetadataProperties): MetadataProperties object. |
| Defaults to ``None``. |
| marketplace_cert (bool): A boolean value indicating if the Model Package is certified |
| for AWS Marketplace. Defaults to ``False``. |
| approval_status (str or PipelineVariable): Model Approval Status, values can be |
| "Approved", "Rejected", or "PendingManualApproval". Defaults to |
| ``PendingManualApproval``. |
| description (str): Model Package description. Defaults to ``None``. |
| drift_check_baselines (DriftCheckBaselines): DriftCheckBaselines object (default: None). |
| customer_metadata_properties (dict[str, str] or dict[str, PipelineVariable]): |
| A dictionary of key-value paired metadata properties (default: None). |
| domain (str or PipelineVariable): Domain values can be "COMPUTER_VISION", |
| "NATURAL_LANGUAGE_PROCESSING", "MACHINE_LEARNING" (default: None). |
| sample_payload_url (str or PipelineVariable): The S3 path where the sample payload |
| is stored (default: None). |
| task (str or PipelineVariable): Task values which are supported by Inference Recommender |
| are "FILL_MASK", "IMAGE_CLASSIFICATION", "OBJECT_DETECTION", "TEXT_GENERATION", |
| "IMAGE_SEGMENTATION", "CLASSIFICATION", "REGRESSION", "OTHER" (default: None). |
| framework (str or PipelineVariable): Machine learning framework of the model package |
| container image (default: None). |
| framework_version (str or PipelineVariable): Framework version of the Model Package |
| Container Image (default: None). |
| nearest_model_name (str or PipelineVariable): Name of a pre-trained machine learning |
| benchmarked by Amazon SageMaker Inference Recommender (default: None). |
| data_input_configuration (str or PipelineVariable): Input object for the model |
| (default: None). |
| |
| Returns: |
| A `sagemaker.model.ModelPackage` instance. |
| """ |
| instance_type = inference_instances[0] if inference_instances else None |
| self._init_sagemaker_session_if_does_not_exist(instance_type) |
|
|
| if image_uri: |
| self.image_uri = image_uri |
| if not self.image_uri: |
| self.image_uri = self.serving_image_uri( |
| region_name=self.sagemaker_session.boto_session.region_name, |
| instance_type=instance_type, |
| ) |
| if not is_pipeline_variable(framework): |
| framework = ( |
| framework |
| or fetch_framework_and_framework_version( |
| self.tensorflow_version, self.pytorch_version |
| )[0] |
| ).upper() |
| return super(HuggingFaceModel, self).register( |
| content_types, |
| response_types, |
| inference_instances, |
| transform_instances, |
| model_package_name, |
| model_package_group_name, |
| image_uri, |
| model_metrics, |
| metadata_properties, |
| marketplace_cert, |
| approval_status, |
| description, |
| drift_check_baselines=drift_check_baselines, |
| customer_metadata_properties=customer_metadata_properties, |
| domain=domain, |
| sample_payload_url=sample_payload_url, |
| task=task, |
| framework=framework, |
| framework_version=framework_version |
| or fetch_framework_and_framework_version(self.tensorflow_version, self.pytorch_version)[ |
| 1 |
| ], |
| nearest_model_name=nearest_model_name, |
| data_input_configuration=data_input_configuration, |
| ) |
|
|
| def prepare_container_def( |
| self, instance_type=None, accelerator_type=None, serverless_inference_config=None |
| ): |
| """A container definition with framework configuration set in model environment variables. |
| |
| Args: |
| instance_type (str): The EC2 instance type to deploy this Model to. |
| For example, 'ml.p2.xlarge'. |
| accelerator_type (str): The Elastic Inference accelerator type to |
| deploy to the instance for loading and making inferences to the |
| model. |
| serverless_inference_config (sagemaker.serverless.ServerlessInferenceConfig): |
| Specifies configuration related to serverless endpoint. Instance type is |
| not provided in serverless inference. So this is used to find image URIs. |
| |
| Returns: |
| dict[str, str]: A container definition object usable with the |
| CreateModel API. |
| """ |
| deploy_image = self.image_uri |
| if not deploy_image: |
| if instance_type is None and serverless_inference_config is None: |
| raise ValueError( |
| "Must supply either an instance type (for choosing CPU vs GPU) or an image URI." |
| ) |
|
|
| region_name = self.sagemaker_session.boto_session.region_name |
| deploy_image = self.serving_image_uri( |
| region_name, |
| instance_type, |
| accelerator_type=accelerator_type, |
| serverless_inference_config=serverless_inference_config, |
| ) |
|
|
| deploy_key_prefix = model_code_key_prefix(self.key_prefix, self.name, deploy_image) |
| self._upload_code(deploy_key_prefix, repack=True) |
| deploy_env = dict(self.env) |
| deploy_env.update(self._script_mode_env_vars()) |
|
|
| if self.model_server_workers: |
| deploy_env[MODEL_SERVER_WORKERS_PARAM_NAME.upper()] = to_string( |
| self.model_server_workers |
| ) |
| return sagemaker.container_def( |
| deploy_image, self.repacked_model_data or self.model_data, deploy_env |
| ) |
|
|
| def serving_image_uri( |
| self, |
| region_name, |
| instance_type=None, |
| accelerator_type=None, |
| serverless_inference_config=None, |
| ): |
| """Create a URI for the serving image. |
| |
| Args: |
| region_name (str): AWS region where the image is uploaded. |
| instance_type (str): SageMaker instance type. Used to determine device type |
| (cpu/gpu/family-specific optimized). |
| accelerator_type (str): The Elastic Inference accelerator type to |
| deploy to the instance for loading and making inferences to the |
| model. |
| serverless_inference_config (sagemaker.serverless.ServerlessInferenceConfig): |
| Specifies configuration related to serverless endpoint. Instance type is |
| not provided in serverless inference. So this is used used to determine device type. |
| |
| Returns: |
| str: The appropriate image URI based on the given parameters. |
| |
| """ |
| if self.tensorflow_version is not None: |
| base_framework_version = ( |
| f"tensorflow{self.tensorflow_version}" |
| ) |
| else: |
| base_framework_version = f"pytorch{self.pytorch_version}" |
| return image_uris.retrieve( |
| self._framework_name, |
| region_name, |
| version=self.framework_version, |
| py_version=self.py_version, |
| instance_type=instance_type, |
| accelerator_type=accelerator_type, |
| image_scope="inference", |
| base_framework_version=base_framework_version, |
| serverless_inference_config=serverless_inference_config, |
| ) |
|
|