| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| """This module is the entry to run spark processing script. |
| |
| This module contains code related to Spark Processors, which are used |
| for Processing jobs. These jobs let customers perform data pre-processing, |
| post-processing, feature engineering, data validation, and model evaluation |
| on SageMaker using Spark and PySpark. |
| """ |
| from __future__ import absolute_import |
|
|
| import json |
| import logging |
| import os.path |
| import shutil |
| import subprocess |
| import tempfile |
| import time |
| import urllib.request |
| from enum import Enum |
| from io import BytesIO |
| from urllib.parse import urlparse |
|
|
| from typing import Union, List, Dict, Optional |
|
|
| from sagemaker import image_uris |
| from sagemaker.local.image import _ecr_login_if_needed, _pull_image |
| from sagemaker.processing import ProcessingInput, ProcessingOutput, ScriptProcessor |
| from sagemaker.s3 import S3Uploader |
| from sagemaker.session import Session |
| from sagemaker.network import NetworkConfig |
| from sagemaker.spark import defaults |
|
|
| from sagemaker.workflow import is_pipeline_variable |
| from sagemaker.workflow.entities import PipelineVariable |
| from sagemaker.workflow.functions import Join |
|
|
| logger = logging.getLogger(__name__) |
|
|
|
|
| class _SparkProcessorBase(ScriptProcessor): |
| """Handles Amazon SageMaker processing tasks for jobs using Spark. |
| |
| Base class for either PySpark or SparkJars. |
| """ |
|
|
| _default_command = "smspark-submit" |
| _conf_container_base_path = "/opt/ml/processing/input/" |
| _conf_container_input_name = "conf" |
| _conf_file_name = "configuration.json" |
| _valid_configuration_keys = ["Classification", "Properties", "Configurations"] |
| _valid_configuration_classifications = [ |
| "core-site", |
| "hadoop-env", |
| "hadoop-log4j", |
| "hive-env", |
| "hive-log4j", |
| "hive-exec-log4j", |
| "hive-site", |
| "spark-defaults", |
| "spark-env", |
| "spark-log4j", |
| "spark-hive-site", |
| "spark-metrics", |
| "yarn-env", |
| "yarn-site", |
| "export", |
| ] |
|
|
| _submit_jars_input_channel_name = "jars" |
| _submit_files_input_channel_name = "files" |
| _submit_py_files_input_channel_name = "py-files" |
| _submit_deps_error_message = ( |
| "Please specify a list of one or more S3 URIs, " |
| "local file paths, and/or local directory paths" |
| ) |
|
|
| |
| _history_server_port = "15050" |
| _history_server_url_suffix = f"/proxy/{_history_server_port}" |
| _spark_event_log_default_local_path = "/opt/ml/processing/spark-events/" |
|
|
| def __init__( |
| self, |
| role, |
| instance_type, |
| instance_count, |
| framework_version=None, |
| py_version=None, |
| container_version=None, |
| image_uri=None, |
| volume_size_in_gb=30, |
| volume_kms_key=None, |
| output_kms_key=None, |
| max_runtime_in_seconds=None, |
| base_job_name=None, |
| sagemaker_session=None, |
| env=None, |
| tags=None, |
| network_config=None, |
| ): |
| """Initialize a ``_SparkProcessorBase`` instance. |
| |
| The _SparkProcessorBase handles Amazon SageMaker processing tasks for |
| jobs using SageMaker Spark. |
| |
| Args: |
| framework_version (str): The version of SageMaker PySpark. |
| py_version (str): The version of python. |
| container_version (str): The version of spark container. |
| role (str): An AWS IAM role name or ARN. The Amazon SageMaker training jobs |
| and APIs that create Amazon SageMaker endpoints use this role |
| to access training data and model artifacts. After the endpoint |
| is created, the inference code might use the IAM role, if it |
| needs to access an AWS resource. |
| instance_type (str): Type of EC2 instance to use for |
| processing, for example, 'ml.c4.xlarge'. |
| instance_count (int): The number of instances to run |
| the Processing job with. Defaults to 1. |
| volume_size_in_gb (int): Size in GB of the EBS volume to |
| use for storing data during processing (default: 30). |
| volume_kms_key (str): A KMS key for the processing |
| volume. |
| output_kms_key (str): The KMS key id for all ProcessingOutputs. |
| max_runtime_in_seconds (int): Timeout in seconds. |
| After this amount of time Amazon SageMaker terminates the job |
| regardless of its current status. |
| base_job_name (str): Prefix for processing name. If not specified, |
| the processor generates a default job name, based on the |
| training image name and current timestamp. |
| sagemaker_session (sagemaker.session.Session): Session object which |
| manages interactions with Amazon |
| SageMaker APIs and any other AWS services needed. If not specified, |
| the processor creates one using the default AWS configuration chain. |
| env (dict): Environment variables to be passed to the processing job. |
| tags ([dict]): List of tags to be passed to the processing job. |
| network_config (sagemaker.network.NetworkConfig): A NetworkConfig |
| object that configures network isolation, encryption of |
| inter-container traffic, security group IDs, and subnets. |
| """ |
| self.history_server = None |
| self._spark_event_logs_s3_uri = None |
|
|
| session = sagemaker_session or Session() |
| region = session.boto_region_name |
|
|
| self.image_uri = self._retrieve_image_uri( |
| image_uri, framework_version, py_version, container_version, region, instance_type |
| ) |
|
|
| env = env or {} |
| command = [_SparkProcessorBase._default_command] |
|
|
| super(_SparkProcessorBase, self).__init__( |
| role=role, |
| image_uri=self.image_uri, |
| instance_count=instance_count, |
| instance_type=instance_type, |
| command=command, |
| volume_size_in_gb=volume_size_in_gb, |
| volume_kms_key=volume_kms_key, |
| output_kms_key=output_kms_key, |
| max_runtime_in_seconds=max_runtime_in_seconds, |
| base_job_name=base_job_name, |
| sagemaker_session=session, |
| env=env, |
| tags=tags, |
| network_config=network_config, |
| ) |
|
|
| def get_run_args( |
| self, |
| code, |
| inputs=None, |
| outputs=None, |
| arguments=None, |
| ): |
| """Returns a RunArgs object. |
| |
| For processors (:class:`~sagemaker.spark.processing.PySparkProcessor`, |
| :class:`~sagemaker.spark.processing.SparkJar`) that have special |
| run() arguments, this object contains the normalized arguments for passing to |
| :class:`~sagemaker.workflow.steps.ProcessingStep`. |
| |
| Args: |
| code (str): This can be an S3 URI or a local path to a file with the framework |
| script to run. |
| inputs (list[:class:`~sagemaker.processing.ProcessingInput`]): Input files for |
| the processing job. These must be provided as |
| :class:`~sagemaker.processing.ProcessingInput` objects (default: None). |
| outputs (list[:class:`~sagemaker.processing.ProcessingOutput`]): Outputs for |
| the processing job. These can be specified as either path strings or |
| :class:`~sagemaker.processing.ProcessingOutput` objects (default: None). |
| arguments (list[str]): A list of string arguments to be passed to a |
| processing job (default: None). |
| """ |
| return super().get_run_args( |
| code=code, |
| inputs=inputs, |
| outputs=outputs, |
| arguments=arguments, |
| ) |
|
|
| def run( |
| self, |
| submit_app, |
| inputs=None, |
| outputs=None, |
| arguments=None, |
| wait=True, |
| logs=True, |
| job_name=None, |
| experiment_config=None, |
| kms_key=None, |
| ): |
| """Runs a processing job. |
| |
| Args: |
| submit_app (str): .py or .jar file to submit to Spark as the primary application |
| inputs (list[:class:`~sagemaker.processing.ProcessingInput`]): Input files for |
| the processing job. These must be provided as |
| :class:`~sagemaker.processing.ProcessingInput` objects (default: None). |
| outputs (list[:class:`~sagemaker.processing.ProcessingOutput`]): Outputs for |
| the processing job. These can be specified as either path strings or |
| :class:`~sagemaker.processing.ProcessingOutput` objects (default: None). |
| arguments (list[str]): A list of string arguments to be passed to a |
| processing job (default: None). |
| wait (bool): Whether the call should wait until the job completes (default: True). |
| logs (bool): Whether to show the logs produced by the job. |
| Only meaningful when wait is True (default: True). |
| job_name (str): Processing job name. If not specified, the processor generates |
| a default job name, based on the base job name and current timestamp. |
| experiment_config (dict[str, str]): Experiment management configuration. |
| Optionally, the dict can contain three keys: |
| 'ExperimentName', 'TrialName', and 'TrialComponentDisplayName'. |
| The behavior of setting these keys is as follows: |
| * If `ExperimentName` is supplied but `TrialName` is not a Trial will be |
| automatically created and the job's Trial Component associated with the Trial. |
| * If `TrialName` is supplied and the Trial already exists the job's Trial Component |
| will be associated with the Trial. |
| * If both `ExperimentName` and `TrialName` are not supplied the trial component |
| will be unassociated. |
| * `TrialComponentDisplayName` is used for display in Studio. |
| kms_key (str): The ARN of the KMS key that is used to encrypt the |
| user code file (default: None). |
| """ |
| self._current_job_name = self._generate_current_job_name(job_name=job_name) |
|
|
| if is_pipeline_variable(submit_app): |
| raise ValueError( |
| "submit_app argument has to be a valid S3 URI or local file path " |
| + "rather than a pipeline variable" |
| ) |
|
|
| return super().run( |
| submit_app, |
| inputs, |
| outputs, |
| arguments, |
| wait, |
| logs, |
| job_name, |
| experiment_config, |
| kms_key, |
| ) |
|
|
| def _extend_processing_args(self, inputs, outputs, **kwargs): |
| """Extends processing job args such as inputs.""" |
|
|
| if kwargs.get("spark_event_logs_s3_uri"): |
| spark_event_logs_s3_uri = kwargs.get("spark_event_logs_s3_uri") |
| self._validate_s3_uri(spark_event_logs_s3_uri) |
|
|
| self._spark_event_logs_s3_uri = spark_event_logs_s3_uri |
| self.command.extend( |
| [ |
| "--local-spark-event-logs-dir", |
| _SparkProcessorBase._spark_event_log_default_local_path, |
| ] |
| ) |
|
|
| output = ProcessingOutput( |
| source=_SparkProcessorBase._spark_event_log_default_local_path, |
| destination=spark_event_logs_s3_uri, |
| s3_upload_mode="Continuous", |
| ) |
|
|
| outputs = outputs or [] |
| outputs.append(output) |
|
|
| if kwargs.get("configuration"): |
| configuration = kwargs.get("configuration") |
| self._validate_configuration(configuration) |
| inputs = inputs or [] |
| inputs.append(self._stage_configuration(configuration)) |
|
|
| return inputs, outputs |
|
|
| def start_history_server(self, spark_event_logs_s3_uri=None): |
| """Starts a Spark history server. |
| |
| Args: |
| spark_event_logs_s3_uri (str): S3 URI where Spark events are stored. |
| """ |
| if _ecr_login_if_needed(self.sagemaker_session.boto_session, self.image_uri): |
| logger.info("Pulling spark history server image...") |
| _pull_image(self.image_uri) |
| history_server_env_variables = self._prepare_history_server_env_variables( |
| spark_event_logs_s3_uri |
| ) |
| self.history_server = _HistoryServer( |
| history_server_env_variables, self.image_uri, self._get_network_config() |
| ) |
| self.history_server.run() |
| self._check_history_server() |
|
|
| def terminate_history_server(self): |
| """Terminates the Spark history server.""" |
| if self.history_server: |
| logger.info("History server is running, terminating history server") |
| self.history_server.down() |
| self.history_server = None |
|
|
| def _retrieve_image_uri( |
| self, image_uri, framework_version, py_version, container_version, region, instance_type |
| ): |
| """Builds an image URI.""" |
| if not image_uri: |
| if (py_version is None) != (container_version is None): |
| raise ValueError( |
| "Both or neither of py_version and container_version should be set" |
| ) |
|
|
| if container_version: |
| container_version = f"v{container_version}" |
|
|
| return image_uris.retrieve( |
| defaults.SPARK_NAME, |
| region, |
| version=framework_version, |
| instance_type=instance_type, |
| py_version=py_version, |
| container_version=container_version, |
| ) |
|
|
| return image_uri |
|
|
| def _validate_configuration(self, configuration): |
| """Validates the user-provided Hadoop/Spark/Hive configuration. |
| |
| This ensures that the list or dictionary the user provides will serialize to |
| JSON matching the schema of EMR's application configuration: |
| |
| https://docs.aws.amazon.com/emr/latest/ReleaseGuide/emr-configure-apps.html |
| """ |
| emr_configure_apps_url = ( |
| "https://docs.aws.amazon.com/emr/latest/ReleaseGuide/emr-configure-apps.html" |
| ) |
| if isinstance(configuration, dict): |
| keys = configuration.keys() |
| if "Classification" not in keys or "Properties" not in keys: |
| raise ValueError( |
| f"Missing one or more required keys in configuration dictionary " |
| f"{configuration} Please see {emr_configure_apps_url} for more information" |
| ) |
|
|
| for key in keys: |
| if key not in self._valid_configuration_keys: |
| raise ValueError( |
| f"Invalid key: {key}. Must be one of {self._valid_configuration_keys}. " |
| f"Please see {emr_configure_apps_url} for more information." |
| ) |
| if key == "Classification": |
| if configuration[key] not in self._valid_configuration_classifications: |
| raise ValueError( |
| f"Invalid classification: {key}. Must be one of " |
| f"{self._valid_configuration_classifications}" |
| ) |
|
|
| if isinstance(configuration, list): |
| for item in configuration: |
| self._validate_configuration(item) |
|
|
| def _stage_configuration(self, configuration): |
| """Serializes and uploads the user-provided EMR application configuration to S3. |
| |
| This method prepares an input channel. |
| |
| Args: |
| configuration (Dict): the configuration dict for the EMR application configuration. |
| """ |
|
|
| serialized_configuration = BytesIO(json.dumps(configuration).encode("utf-8")) |
| s3_uri = ( |
| f"s3://{self.sagemaker_session.default_bucket()}/{self._current_job_name}/" |
| f"input/{self._conf_container_input_name}/{self._conf_file_name}" |
| ) |
|
|
| S3Uploader.upload_string_as_file_body( |
| body=serialized_configuration, |
| desired_s3_uri=s3_uri, |
| sagemaker_session=self.sagemaker_session, |
| ) |
|
|
| conf_input = ProcessingInput( |
| source=s3_uri, |
| destination=f"{self._conf_container_base_path}{self._conf_container_input_name}", |
| input_name=_SparkProcessorBase._conf_container_input_name, |
| ) |
| return conf_input |
|
|
| def _stage_submit_deps(self, submit_deps, input_channel_name): |
| """Prepares a list of paths to jars, py-files, or files dependencies. |
| |
| This prepared list of paths is provided as `spark-submit` options. |
| The submit_deps list may include a combination of S3 URIs and local paths. |
| Any S3 URIs are appended to the `spark-submit` option value without modification. |
| Any local file paths are copied to a temp directory, uploaded to a default S3 URI, |
| and included as a ProcessingInput channel to provide as local files to the SageMaker |
| Spark container. |
| |
| :param submit_deps (list[str]): List of one or more dependency paths to include. |
| :param input_channel_name (str): The `spark-submit` option name associated with |
| the input channel. |
| :return (Optional[ProcessingInput], str): Tuple of (left) optional ProcessingInput |
| for the input channel, and (right) comma-delimited value for |
| `spark-submit` option. |
| """ |
| if not submit_deps: |
| raise ValueError( |
| f"submit_deps value may not be empty. {self._submit_deps_error_message}" |
| ) |
| if not input_channel_name: |
| raise ValueError("input_channel_name value may not be empty.") |
|
|
| input_channel_s3_uri = ( |
| f"s3://{self.sagemaker_session.default_bucket()}" |
| f"/{self._current_job_name}/input/{input_channel_name}" |
| ) |
|
|
| use_input_channel = False |
| spark_opt_s3_uris = [] |
| spark_opt_s3_uris_has_pipeline_var = False |
|
|
| with tempfile.TemporaryDirectory() as tmpdir: |
| for dep_path in submit_deps: |
| if is_pipeline_variable(dep_path): |
| spark_opt_s3_uris.append(dep_path) |
| spark_opt_s3_uris_has_pipeline_var = True |
| continue |
| dep_url = urlparse(dep_path) |
| |
| if dep_url.scheme in ["s3", "s3a"]: |
| spark_opt_s3_uris.append(dep_path) |
| |
| elif not dep_url.scheme or dep_url.scheme == "file": |
| if not os.path.isfile(dep_path): |
| raise ValueError( |
| f"submit_deps path {dep_path} is not a valid local file. " |
| f"{self._submit_deps_error_message}" |
| ) |
| logger.info( |
| "Copying dependency from local path %s to tmpdir %s", dep_path, tmpdir |
| ) |
| shutil.copy(dep_path, tmpdir) |
| else: |
| raise ValueError( |
| f"submit_deps path {dep_path} references unsupported filesystem " |
| f"scheme: {dep_url.scheme} {self._submit_deps_error_message}" |
| ) |
|
|
| |
| if os.listdir(tmpdir): |
| logger.info( |
| "Uploading dependencies from tmpdir %s to S3 %s", tmpdir, input_channel_s3_uri |
| ) |
| S3Uploader.upload( |
| local_path=tmpdir, |
| desired_s3_uri=input_channel_s3_uri, |
| sagemaker_session=self.sagemaker_session, |
| ) |
| use_input_channel = True |
|
|
| |
| |
| |
| if use_input_channel: |
| input_channel = ProcessingInput( |
| source=input_channel_s3_uri, |
| destination=f"{self._conf_container_base_path}{input_channel_name}", |
| input_name=input_channel_name, |
| ) |
| spark_opt = ( |
| Join(on=",", values=spark_opt_s3_uris + [input_channel.destination]) |
| if spark_opt_s3_uris_has_pipeline_var |
| else ",".join(spark_opt_s3_uris + [input_channel.destination]) |
| ) |
| |
| else: |
| input_channel = None |
| spark_opt = ( |
| Join(on=",", values=spark_opt_s3_uris) |
| if spark_opt_s3_uris_has_pipeline_var |
| else ",".join(spark_opt_s3_uris) |
| ) |
|
|
| return input_channel, spark_opt |
|
|
| def _get_network_config(self): |
| """Runs container with different network config based on different env.""" |
| if self._is_notebook_instance(): |
| return "--network host" |
|
|
| return f"-p 80:80 -p {self._history_server_port}:{self._history_server_port}" |
|
|
| def _prepare_history_server_env_variables(self, spark_event_logs_s3_uri): |
| """Gets all parameters required to run history server.""" |
| |
| history_server_env_variables = {} |
|
|
| if spark_event_logs_s3_uri: |
| history_server_env_variables[ |
| _HistoryServer.arg_event_logs_s3_uri |
| ] = spark_event_logs_s3_uri |
| |
| elif self._spark_event_logs_s3_uri is not None: |
| history_server_env_variables[ |
| _HistoryServer.arg_event_logs_s3_uri |
| ] = self._spark_event_logs_s3_uri |
| else: |
| raise ValueError( |
| "SPARK_EVENT_LOGS_S3_URI not present. You can specify spark_event_logs_s3_uri " |
| "either in run() or start_history_server()" |
| ) |
|
|
| history_server_env_variables.update(self._config_aws_credentials()) |
| region = self.sagemaker_session.boto_region_name |
| history_server_env_variables["AWS_REGION"] = region |
|
|
| if self._is_notebook_instance(): |
| history_server_env_variables[ |
| _HistoryServer.arg_remote_domain_name |
| ] = self._get_notebook_instance_domain() |
|
|
| return history_server_env_variables |
|
|
| def _is_notebook_instance(self): |
| """Determine whether it is a notebook instance.""" |
| return os.path.isfile("/opt/ml/metadata/resource-metadata.json") |
|
|
| def _get_notebook_instance_domain(self): |
| """Get the instance's domain.""" |
| region = self.sagemaker_session.boto_region_name |
| with open("/opt/ml/metadata/resource-metadata.json") as file: |
| data = json.load(file) |
| notebook_name = data["ResourceName"] |
|
|
| return f"https://{notebook_name}.notebook.{region}.sagemaker.aws" |
|
|
| def _check_history_server(self, ping_timeout=40): |
| """Print message indicating the status of history server. |
| |
| Pings port 15050 to check whether the history server is up. |
| Times out after `ping_timeout`. |
| |
| Args: |
| ping_timeout (int): Timeout in seconds (defaults to 40). |
| """ |
| |
| timeout = time.time() + ping_timeout |
|
|
| while True: |
| if self._is_history_server_started(): |
| if self._is_notebook_instance(): |
| logger.info( |
| "History server is up on %s%s", |
| self._get_notebook_instance_domain(), |
| self._history_server_url_suffix, |
| ) |
| else: |
| logger.info( |
| "History server is up on http://0.0.0.0%s", self._history_server_url_suffix |
| ) |
| break |
| if time.time() > timeout: |
| logger.error( |
| "History server failed to start. Please run 'docker logs history_server' " |
| "to see logs" |
| ) |
| break |
|
|
| time.sleep(1) |
|
|
| def _is_history_server_started(self): |
| """Check if history server is started.""" |
| try: |
| response = urllib.request.urlopen(f"http://localhost:{self._history_server_port}") |
| return response.status == 200 |
| except Exception: |
| return False |
|
|
| |
| def _validate_s3_uri(self, spark_output_s3_path): |
| """Validate whether the URI uses an S3 scheme. |
| |
| In the future, this validation will perform deeper S3 validation. |
| |
| Args: |
| spark_output_s3_path (str): The URI of the Spark output S3 Path. |
| """ |
| if is_pipeline_variable(spark_output_s3_path): |
| return |
|
|
| if urlparse(spark_output_s3_path).scheme != "s3": |
| raise ValueError( |
| f"Invalid s3 path: {spark_output_s3_path}. Please enter something like " |
| "s3://bucket-name/folder-name" |
| ) |
|
|
| def _config_aws_credentials(self): |
| """Configure AWS credentials.""" |
| try: |
| creds = self.sagemaker_session.boto_session.get_credentials() |
| access_key = creds.access_key |
| secret_key = creds.secret_key |
| token = creds.token |
|
|
| return { |
| "AWS_ACCESS_KEY_ID": str(access_key), |
| "AWS_SECRET_ACCESS_KEY": str(secret_key), |
| "AWS_SESSION_TOKEN": str(token), |
| } |
| except Exception as e: |
| logger.info("Could not get AWS credentials: %s", e) |
| return {} |
|
|
| def _handle_script_dependencies(self, inputs, submit_files, file_type): |
| """Handle script dependencies |
| |
| The method extends inputs and command based on input files and file_type |
| """ |
|
|
| if not submit_files: |
| return inputs |
|
|
| input_channel_name_dict = { |
| FileType.PYTHON: self._submit_py_files_input_channel_name, |
| FileType.JAR: self._submit_jars_input_channel_name, |
| FileType.FILE: self._submit_files_input_channel_name, |
| } |
|
|
| files_input, files_opt = self._stage_submit_deps( |
| submit_files, input_channel_name_dict[file_type] |
| ) |
|
|
| inputs = inputs or [] |
|
|
| if files_input: |
| inputs.append(files_input) |
|
|
| if files_opt: |
| self.command.extend([f"--{input_channel_name_dict[file_type]}", files_opt]) |
|
|
| return inputs |
|
|
|
|
| class PySparkProcessor(_SparkProcessorBase): |
| """Handles Amazon SageMaker processing tasks for jobs using PySpark.""" |
|
|
| def __init__( |
| self, |
| role: str, |
| instance_type: Union[int, PipelineVariable], |
| instance_count: Union[str, PipelineVariable], |
| framework_version: Optional[str] = None, |
| py_version: Optional[str] = None, |
| container_version: Optional[str] = None, |
| image_uri: Optional[Union[str, PipelineVariable]] = None, |
| volume_size_in_gb: Union[int, PipelineVariable] = 30, |
| volume_kms_key: Optional[Union[str, PipelineVariable]] = None, |
| output_kms_key: Optional[Union[str, PipelineVariable]] = None, |
| max_runtime_in_seconds: Optional[Union[int, PipelineVariable]] = None, |
| base_job_name: Optional[str] = None, |
| sagemaker_session: Optional[Session] = None, |
| env: Optional[Dict[str, Union[str, PipelineVariable]]] = None, |
| tags: Optional[List[Dict[str, Union[str, PipelineVariable]]]] = None, |
| network_config: Optional[NetworkConfig] = None, |
| ): |
| """Initialize an ``PySparkProcessor`` instance. |
| |
| The PySparkProcessor handles Amazon SageMaker processing tasks for jobs |
| using SageMaker PySpark. |
| |
| Args: |
| framework_version (str): The version of SageMaker PySpark. |
| py_version (str): The version of python. |
| container_version (str): The version of spark container. |
| role (str): An AWS IAM role name or ARN. The Amazon SageMaker training jobs |
| and APIs that create Amazon SageMaker endpoints use this role |
| to access training data and model artifacts. After the endpoint |
| is created, the inference code might use the IAM role, if it |
| needs to access an AWS resource. |
| instance_type (str or PipelineVariable): Type of EC2 instance to use for |
| processing, for example, 'ml.c4.xlarge'. |
| instance_count (int or PipelineVariable): The number of instances to run |
| the Processing job with. Defaults to 1. |
| volume_size_in_gb (int or PipelineVariable): Size in GB of the EBS volume to |
| use for storing data during processing (default: 30). |
| volume_kms_key (str or PipelineVariable): A KMS key for the processing |
| volume. |
| output_kms_key (str or PipelineVariable): The KMS key id for all ProcessingOutputs. |
| max_runtime_in_seconds (int or PipelineVariable): Timeout in seconds. |
| After this amount of time Amazon SageMaker terminates the job |
| regardless of its current status. |
| base_job_name (str): Prefix for processing name. If not specified, |
| the processor generates a default job name, based on the |
| training image name and current timestamp. |
| sagemaker_session (sagemaker.session.Session): Session object which |
| manages interactions with Amazon SageMaker APIs and any other |
| AWS services needed. If not specified, the processor creates one |
| using the default AWS configuration chain. |
| env (dict[str, str] or dict[str, PipelineVariable]): Environment variables to |
| be passed to the processing job. |
| tags (list[dict[str, str] or list[dict[str, PipelineVariable]]): List of tags to |
| be passed to the processing job. |
| network_config (sagemaker.network.NetworkConfig): A NetworkConfig |
| object that configures network isolation, encryption of |
| inter-container traffic, security group IDs, and subnets. |
| """ |
|
|
| super(PySparkProcessor, self).__init__( |
| role=role, |
| instance_count=instance_count, |
| instance_type=instance_type, |
| framework_version=framework_version, |
| py_version=py_version, |
| container_version=container_version, |
| image_uri=image_uri, |
| volume_size_in_gb=volume_size_in_gb, |
| volume_kms_key=volume_kms_key, |
| output_kms_key=output_kms_key, |
| max_runtime_in_seconds=max_runtime_in_seconds, |
| base_job_name=base_job_name, |
| sagemaker_session=sagemaker_session, |
| env=env, |
| tags=tags, |
| network_config=network_config, |
| ) |
|
|
| def get_run_args( |
| self, |
| submit_app, |
| submit_py_files=None, |
| submit_jars=None, |
| submit_files=None, |
| inputs=None, |
| outputs=None, |
| arguments=None, |
| job_name=None, |
| configuration=None, |
| spark_event_logs_s3_uri=None, |
| ): |
| """Returns a RunArgs object. |
| |
| This object contains the normalized inputs, outputs and arguments |
| needed when using a ``PySparkProcessor`` in a |
| :class:`~sagemaker.workflow.steps.ProcessingStep`. |
| |
| Args: |
| submit_app (str): Path (local or S3) to Python file to submit to Spark |
| as the primary application. This is translated to the `code` |
| property on the returned `RunArgs` object. |
| submit_py_files (list[str]): List of paths (local or S3) to provide for |
| `spark-submit --py-files` option |
| submit_jars (list[str]): List of paths (local or S3) to provide for |
| `spark-submit --jars` option |
| submit_files (list[str]): List of paths (local or S3) to provide for |
| `spark-submit --files` option |
| inputs (list[:class:`~sagemaker.processing.ProcessingInput`]): Input files for |
| the processing job. These must be provided as |
| :class:`~sagemaker.processing.ProcessingInput` objects (default: None). |
| outputs (list[:class:`~sagemaker.processing.ProcessingOutput`]): Outputs for |
| the processing job. These can be specified as either path strings or |
| :class:`~sagemaker.processing.ProcessingOutput` objects (default: None). |
| arguments (list[str]): A list of string arguments to be passed to a |
| processing job (default: None). |
| job_name (str): Processing job name. If not specified, the processor generates |
| a default job name, based on the base job name and current timestamp. |
| configuration (list[dict] or dict): Configuration for Hadoop, Spark, or Hive. |
| List or dictionary of EMR-style classifications. |
| https://docs.aws.amazon.com/emr/latest/ReleaseGuide/emr-configure-apps.html |
| spark_event_logs_s3_uri (str): S3 path where spark application events will |
| be published to. |
| """ |
| self._current_job_name = self._generate_current_job_name(job_name=job_name) |
|
|
| if not submit_app: |
| raise ValueError("submit_app is required") |
|
|
| extended_inputs, extended_outputs = self._extend_processing_args( |
| inputs=inputs, |
| outputs=outputs, |
| submit_py_files=submit_py_files, |
| submit_jars=submit_jars, |
| submit_files=submit_files, |
| configuration=configuration, |
| spark_event_logs_s3_uri=spark_event_logs_s3_uri, |
| ) |
|
|
| return super().get_run_args( |
| code=submit_app, |
| inputs=extended_inputs, |
| outputs=extended_outputs, |
| arguments=arguments, |
| ) |
|
|
| def run( |
| self, |
| submit_app: str, |
| submit_py_files: Optional[List[Union[str, PipelineVariable]]] = None, |
| submit_jars: Optional[List[Union[str, PipelineVariable]]] = None, |
| submit_files: Optional[List[Union[str, PipelineVariable]]] = None, |
| inputs: Optional[List[ProcessingInput]] = None, |
| outputs: Optional[List[ProcessingOutput]] = None, |
| arguments: Optional[List[Union[str, PipelineVariable]]] = None, |
| wait: bool = True, |
| logs: bool = True, |
| job_name: Optional[str] = None, |
| experiment_config: Optional[Dict[str, str]] = None, |
| configuration: Optional[Union[List[Dict], Dict]] = None, |
| spark_event_logs_s3_uri: Optional[Union[str, PipelineVariable]] = None, |
| kms_key: Optional[str] = None, |
| ): |
| """Runs a processing job. |
| |
| Args: |
| submit_app (str): Path (local or S3) to Python file to submit to Spark |
| as the primary application |
| submit_py_files (list[str] or list[PipelineVariable]): List of paths (local or S3) |
| to provide for `spark-submit --py-files` option |
| submit_jars (list[str] or list[PipelineVariable]): List of paths (local or S3) |
| to provide for `spark-submit --jars` option |
| submit_files (list[str] or list[PipelineVariable]): List of paths (local or S3) |
| to provide for `spark-submit --files` option |
| inputs (list[:class:`~sagemaker.processing.ProcessingInput`]): Input files for |
| the processing job. These must be provided as |
| :class:`~sagemaker.processing.ProcessingInput` objects (default: None). |
| outputs (list[:class:`~sagemaker.processing.ProcessingOutput`]): Outputs for |
| the processing job. These can be specified as either path strings or |
| :class:`~sagemaker.processing.ProcessingOutput` objects (default: None). |
| arguments (list[str] or list[PipelineVariable]): A list of string arguments to |
| be passed to a processing job (default: None). |
| wait (bool): Whether the call should wait until the job completes (default: True). |
| logs (bool): Whether to show the logs produced by the job. |
| Only meaningful when wait is True (default: True). |
| job_name (str): Processing job name. If not specified, the processor generates |
| a default job name, based on the base job name and current timestamp. |
| experiment_config (dict[str, str]): Experiment management configuration. |
| Optionally, the dict can contain three keys: |
| 'ExperimentName', 'TrialName', and 'TrialComponentDisplayName'. |
| The behavior of setting these keys is as follows: |
| * If `ExperimentName` is supplied but `TrialName` is not a Trial will be |
| automatically created and the job's Trial Component associated with the Trial. |
| * If `TrialName` is supplied and the Trial already exists the job's Trial Component |
| will be associated with the Trial. |
| * If both `ExperimentName` and `TrialName` are not supplied the trial component |
| will be unassociated. |
| * `TrialComponentDisplayName` is used for display in Studio. |
| configuration (list[dict] or dict): Configuration for Hadoop, Spark, or Hive. |
| List or dictionary of EMR-style classifications. |
| https://docs.aws.amazon.com/emr/latest/ReleaseGuide/emr-configure-apps.html |
| spark_event_logs_s3_uri (str or PipelineVariable): S3 path where spark application |
| events will be published to. |
| kms_key (str): The ARN of the KMS key that is used to encrypt the |
| user code file (default: None). |
| """ |
| self._current_job_name = self._generate_current_job_name(job_name=job_name) |
|
|
| if not submit_app: |
| raise ValueError("submit_app is required") |
|
|
| extended_inputs, extended_outputs = self._extend_processing_args( |
| inputs=inputs, |
| outputs=outputs, |
| submit_py_files=submit_py_files, |
| submit_jars=submit_jars, |
| submit_files=submit_files, |
| configuration=configuration, |
| spark_event_logs_s3_uri=spark_event_logs_s3_uri, |
| ) |
|
|
| return super().run( |
| submit_app=submit_app, |
| inputs=extended_inputs, |
| outputs=extended_outputs, |
| arguments=arguments, |
| wait=wait, |
| logs=logs, |
| job_name=self._current_job_name, |
| experiment_config=experiment_config, |
| kms_key=kms_key, |
| ) |
|
|
| def _extend_processing_args(self, inputs, outputs, **kwargs): |
| """Extends inputs and outputs. |
| |
| Args: |
| inputs: Processing inputs. |
| outputs: Processing outputs. |
| kwargs: Additional keyword arguments passed to `super()`. |
| """ |
| self.command = [_SparkProcessorBase._default_command] |
| extended_inputs = self._handle_script_dependencies( |
| inputs, kwargs.get("submit_py_files"), FileType.PYTHON |
| ) |
| extended_inputs = self._handle_script_dependencies( |
| extended_inputs, kwargs.get("submit_jars"), FileType.JAR |
| ) |
| extended_inputs = self._handle_script_dependencies( |
| extended_inputs, kwargs.get("submit_files"), FileType.FILE |
| ) |
|
|
| return super()._extend_processing_args(extended_inputs, outputs, **kwargs) |
|
|
|
|
| class SparkJarProcessor(_SparkProcessorBase): |
| """Handles Amazon SageMaker processing tasks for jobs using Spark with Java or Scala Jars.""" |
|
|
| def __init__( |
| self, |
| role: str, |
| instance_type: Union[int, PipelineVariable], |
| instance_count: Union[str, PipelineVariable], |
| framework_version: Optional[str] = None, |
| py_version: Optional[str] = None, |
| container_version: Optional[str] = None, |
| image_uri: Optional[Union[str, PipelineVariable]] = None, |
| volume_size_in_gb: Union[int, PipelineVariable] = 30, |
| volume_kms_key: Optional[Union[str, PipelineVariable]] = None, |
| output_kms_key: Optional[Union[str, PipelineVariable]] = None, |
| max_runtime_in_seconds: Optional[Union[int, PipelineVariable]] = None, |
| base_job_name: Optional[str] = None, |
| sagemaker_session: Optional[Session] = None, |
| env: Optional[Dict[str, Union[str, PipelineVariable]]] = None, |
| tags: Optional[List[Dict[str, Union[str, PipelineVariable]]]] = None, |
| network_config: Optional[NetworkConfig] = None, |
| ): |
| """Initialize a ``SparkJarProcessor`` instance. |
| |
| The SparkProcessor handles Amazon SageMaker processing tasks for jobs |
| using SageMaker Spark. |
| |
| Args: |
| framework_version (str): The version of SageMaker PySpark. |
| py_version (str): The version of python. |
| container_version (str): The version of spark container. |
| role (str): An AWS IAM role name or ARN. The Amazon SageMaker training jobs |
| and APIs that create Amazon SageMaker endpoints use this role |
| to access training data and model artifacts. After the endpoint |
| is created, the inference code might use the IAM role, if it |
| needs to access an AWS resource. |
| instance_type (str or PipelineVariable): Type of EC2 instance to use for |
| processing, for example, 'ml.c4.xlarge'. |
| instance_count (int or PipelineVariable): The number of instances to run |
| the Processing job with. Defaults to 1. |
| volume_size_in_gb (int or PipelineVariable): Size in GB of the EBS volume to |
| use for storing data during processing (default: 30). |
| volume_kms_key (str or PipelineVariable): A KMS key for the processing |
| volume. |
| output_kms_key (str or PipelineVariable): The KMS key id for all ProcessingOutputs. |
| max_runtime_in_seconds (int or PipelineVariable): Timeout in seconds. |
| After this amount of time Amazon SageMaker terminates the job |
| regardless of its current status. |
| base_job_name (str): Prefix for processing name. If not specified, |
| the processor generates a default job name, based on the |
| training image name and current timestamp. |
| sagemaker_session (sagemaker.session.Session): Session object which |
| manages interactions with Amazon SageMaker APIs and any other |
| AWS services needed. If not specified, the processor creates one |
| using the default AWS configuration chain. |
| env (dict[str, str] or dict[str, PipelineVariable]): Environment variables to |
| be passed to the processing job. |
| tags (list[dict[str, str] or list[dict[str, PipelineVariable]]): List of tags to |
| be passed to the processing job. |
| network_config (sagemaker.network.NetworkConfig): A NetworkConfig |
| object that configures network isolation, encryption of |
| inter-container traffic, security group IDs, and subnets. |
| """ |
|
|
| super(SparkJarProcessor, self).__init__( |
| role=role, |
| instance_count=instance_count, |
| instance_type=instance_type, |
| framework_version=framework_version, |
| py_version=py_version, |
| container_version=container_version, |
| image_uri=image_uri, |
| volume_size_in_gb=volume_size_in_gb, |
| volume_kms_key=volume_kms_key, |
| output_kms_key=output_kms_key, |
| max_runtime_in_seconds=max_runtime_in_seconds, |
| base_job_name=base_job_name, |
| sagemaker_session=sagemaker_session, |
| env=env, |
| tags=tags, |
| network_config=network_config, |
| ) |
|
|
| def get_run_args( |
| self, |
| submit_app, |
| submit_class=None, |
| submit_jars=None, |
| submit_files=None, |
| inputs=None, |
| outputs=None, |
| arguments=None, |
| job_name=None, |
| configuration=None, |
| spark_event_logs_s3_uri=None, |
| ): |
| """Returns a RunArgs object. |
| |
| This object contains the normalized inputs, outputs and arguments |
| needed when using a ``SparkJarProcessor`` in a |
| :class:`~sagemaker.workflow.steps.ProcessingStep`. |
| |
| Args: |
| submit_app (str): Path (local or S3) to Python file to submit to Spark |
| as the primary application. This is translated to the `code` |
| property on the returned `RunArgs` object |
| submit_class (str): Java class reference to submit to Spark as the primary |
| application |
| submit_jars (list[str]): List of paths (local or S3) to provide for |
| `spark-submit --jars` option |
| submit_files (list[str]): List of paths (local or S3) to provide for |
| `spark-submit --files` option |
| inputs (list[:class:`~sagemaker.processing.ProcessingInput`]): Input files for |
| the processing job. These must be provided as |
| :class:`~sagemaker.processing.ProcessingInput` objects (default: None). |
| outputs (list[:class:`~sagemaker.processing.ProcessingOutput`]): Outputs for |
| the processing job. These can be specified as either path strings or |
| :class:`~sagemaker.processing.ProcessingOutput` objects (default: None). |
| arguments (list[str]): A list of string arguments to be passed to a |
| processing job (default: None). |
| job_name (str): Processing job name. If not specified, the processor generates |
| a default job name, based on the base job name and current timestamp. |
| configuration (list[dict] or dict): Configuration for Hadoop, Spark, or Hive. |
| List or dictionary of EMR-style classifications. |
| https://docs.aws.amazon.com/emr/latest/ReleaseGuide/emr-configure-apps.html |
| spark_event_logs_s3_uri (str): S3 path where spark application events will |
| be published to. |
| """ |
| self._current_job_name = self._generate_current_job_name(job_name=job_name) |
|
|
| if not submit_app: |
| raise ValueError("submit_app is required") |
|
|
| extended_inputs, extended_outputs = self._extend_processing_args( |
| inputs=inputs, |
| outputs=outputs, |
| submit_class=submit_class, |
| submit_jars=submit_jars, |
| submit_files=submit_files, |
| configuration=configuration, |
| spark_event_logs_s3_uri=spark_event_logs_s3_uri, |
| ) |
|
|
| return super().get_run_args( |
| code=submit_app, |
| inputs=extended_inputs, |
| outputs=extended_outputs, |
| arguments=arguments, |
| ) |
|
|
| def run( |
| self, |
| submit_app: str, |
| submit_class: Union[str, PipelineVariable], |
| submit_jars: Optional[List[Union[str, PipelineVariable]]] = None, |
| submit_files: Optional[List[Union[str, PipelineVariable]]] = None, |
| inputs: Optional[List[ProcessingInput]] = None, |
| outputs: Optional[List[ProcessingOutput]] = None, |
| arguments: Optional[List[Union[str, PipelineVariable]]] = None, |
| wait: bool = True, |
| logs: bool = True, |
| job_name: Optional[str] = None, |
| experiment_config: Optional[Dict[str, str]] = None, |
| configuration: Optional[Union[List[Dict], Dict]] = None, |
| spark_event_logs_s3_uri: Optional[Union[str, PipelineVariable]] = None, |
| kms_key: Optional[str] = None, |
| ): |
| """Runs a processing job. |
| |
| Args: |
| submit_app (str): Path (local or S3) to Jar file to submit to Spark as |
| the primary application |
| submit_class (str or PipelineVariable): Java class reference to submit to Spark |
| as the primary application |
| submit_jars (list[str] or list[PipelineVariable]): List of paths (local or S3) |
| to provide for `spark-submit --jars` option |
| submit_files (list[str] or list[PipelineVariable]): List of paths (local or S3) |
| to provide for `spark-submit --files` option |
| inputs (list[:class:`~sagemaker.processing.ProcessingInput`]): Input files for |
| the processing job. These must be provided as |
| :class:`~sagemaker.processing.ProcessingInput` objects (default: None). |
| outputs (list[:class:`~sagemaker.processing.ProcessingOutput`]): Outputs for |
| the processing job. These can be specified as either path strings or |
| :class:`~sagemaker.processing.ProcessingOutput` objects (default: None). |
| arguments (list[str] or list[PipelineVariable]): A list of string arguments to |
| be passed to a processing job (default: None). |
| wait (bool): Whether the call should wait until the job completes (default: True). |
| logs (bool): Whether to show the logs produced by the job. |
| Only meaningful when wait is True (default: True). |
| job_name (str): Processing job name. If not specified, the processor generates |
| a default job name, based on the base job name and current timestamp. |
| experiment_config (dict[str, str]): Experiment management configuration. |
| Optionally, the dict can contain three keys: |
| 'ExperimentName', 'TrialName', and 'TrialComponentDisplayName'. |
| The behavior of setting these keys is as follows: |
| * If `ExperimentName` is supplied but `TrialName` is not a Trial will be |
| automatically created and the job's Trial Component associated with the Trial. |
| * If `TrialName` is supplied and the Trial already exists the job's Trial Component |
| will be associated with the Trial. |
| * If both `ExperimentName` and `TrialName` are not supplied the trial component |
| will be unassociated. |
| * `TrialComponentDisplayName` is used for display in Studio. |
| configuration (list[dict] or dict): Configuration for Hadoop, Spark, or Hive. |
| List or dictionary of EMR-style classifications. |
| https://docs.aws.amazon.com/emr/latest/ReleaseGuide/emr-configure-apps.html |
| spark_event_logs_s3_uri (str or PipelineVariable): S3 path where spark application |
| events will be published to. |
| kms_key (str): The ARN of the KMS key that is used to encrypt the |
| user code file (default: None). |
| """ |
| self._current_job_name = self._generate_current_job_name(job_name=job_name) |
|
|
| if not submit_app: |
| raise ValueError("submit_app is required") |
|
|
| extended_inputs, extended_outputs = self._extend_processing_args( |
| inputs=inputs, |
| outputs=outputs, |
| submit_class=submit_class, |
| submit_jars=submit_jars, |
| submit_files=submit_files, |
| configuration=configuration, |
| spark_event_logs_s3_uri=spark_event_logs_s3_uri, |
| ) |
|
|
| return super().run( |
| submit_app=submit_app, |
| inputs=extended_inputs, |
| outputs=extended_outputs, |
| arguments=arguments, |
| wait=wait, |
| logs=logs, |
| job_name=self._current_job_name, |
| experiment_config=experiment_config, |
| kms_key=kms_key, |
| ) |
|
|
| def _extend_processing_args(self, inputs, outputs, **kwargs): |
| self.command = [_SparkProcessorBase._default_command] |
| if kwargs.get("submit_class"): |
| self.command.extend(["--class", kwargs.get("submit_class")]) |
| else: |
| raise ValueError("submit_class is required") |
|
|
| extended_inputs = self._handle_script_dependencies( |
| inputs, kwargs.get("submit_jars"), FileType.JAR |
| ) |
| extended_inputs = self._handle_script_dependencies( |
| extended_inputs, kwargs.get("submit_files"), FileType.FILE |
| ) |
|
|
| return super()._extend_processing_args(extended_inputs, outputs, **kwargs) |
|
|
|
|
| class _HistoryServer: |
| """History server class that is responsible for starting history server.""" |
|
|
| _container_name = "history_server" |
| _entry_point = "smspark-history-server" |
| arg_event_logs_s3_uri = "event_logs_s3_uri" |
| arg_remote_domain_name = "remote_domain_name" |
|
|
| _history_server_args_format_map = { |
| arg_event_logs_s3_uri: "--event-logs-s3-uri {} ", |
| arg_remote_domain_name: "--remote-domain-name {} ", |
| } |
|
|
| def __init__(self, cli_args, image_uri, network_config): |
| self.cli_args = cli_args |
| self.image_uri = image_uri |
| self.network_config = network_config |
| self.run_history_server_command = self._get_run_history_server_cmd() |
|
|
| def run(self): |
| """Runs the history server.""" |
| self.down() |
| logger.info("Starting history server...") |
| subprocess.Popen( |
| self.run_history_server_command.split(), stdout=subprocess.PIPE, stderr=subprocess.PIPE |
| ) |
|
|
| def down(self): |
| """Stops and removes the container.""" |
| subprocess.call(["docker", "stop", self._container_name]) |
| subprocess.call(["docker", "rm", self._container_name]) |
| logger.info("History server terminated") |
|
|
| |
| |
| |
| def _get_run_history_server_cmd(self): |
| """Gets the history server command.""" |
| env_options = "" |
| ser_cli_args = "" |
| for key, value in self.cli_args.items(): |
| if key in self._history_server_args_format_map: |
| ser_cli_args += self._history_server_args_format_map[key].format(value) |
| else: |
| env_options += f"--env {key}={value} " |
|
|
| cmd = ( |
| f"docker run {env_options.strip()} --name {self._container_name} " |
| f"{self.network_config} --entrypoint {self._entry_point} {self.image_uri} " |
| f"{ser_cli_args.strip()}" |
| ) |
|
|
| return cmd |
|
|
|
|
| class FileType(Enum): |
| """Enum of file type""" |
|
|
| JAR = 1 |
| PYTHON = 2 |
| FILE = 3 |
|
|