| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| from __future__ import absolute_import |
|
|
| import datetime |
| import io |
| import logging |
| import os |
|
|
| import pytest |
| import six |
| from botocore.exceptions import ClientError |
| from mock import ANY, MagicMock, Mock, patch, call, mock_open |
|
|
| import sagemaker |
| from sagemaker import TrainingInput, Session, get_execution_role |
| from sagemaker.async_inference import AsyncInferenceConfig |
| from sagemaker.session import ( |
| _tuning_job_status, |
| _transform_job_status, |
| _train_done, |
| NOTEBOOK_METADATA_FILE, |
| ) |
| from sagemaker.tuner import WarmStartConfig, WarmStartTypes |
| from sagemaker.inputs import BatchDataCaptureConfig |
|
|
| STATIC_HPs = {"feature_dim": "784"} |
|
|
| SAMPLE_PARAM_RANGES = [{"Name": "mini_batch_size", "MinValue": "10", "MaxValue": "100"}] |
|
|
| ENV_INPUT = {"env_key1": "env_val1", "env_key2": "env_val2", "env_key3": "env_val3"} |
|
|
| REGION = "us-west-2" |
| STS_ENDPOINT = "sts.us-west-2.amazonaws.com" |
|
|
|
|
| @pytest.fixture() |
| def boto_session(): |
| boto_mock = Mock(name="boto_session", region_name=REGION) |
|
|
| client_mock = Mock() |
| client_mock._client_config.user_agent = ( |
| "Boto3/1.9.69 Python/3.6.5 Linux/4.14.77-70.82.amzn1.x86_64 Botocore/1.12.69 Resource" |
| ) |
| boto_mock.client.return_value = client_mock |
| return boto_mock |
|
|
|
|
| @patch("boto3.DEFAULT_SESSION") |
| def test_default_session(boto3_default_session): |
| sess = Session() |
| assert sess.boto_session is boto3_default_session |
|
|
|
|
| @patch("boto3.Session") |
| def test_new_session_created(boto3_session): |
| sess = Session() |
| assert sess.boto_session is boto3_session.return_value |
|
|
|
|
| def test_process(boto_session): |
| session = Session(boto_session) |
|
|
| process_request_args = { |
| "inputs": [ |
| { |
| "InputName": "input-1", |
| "S3Input": { |
| "S3Uri": "mocked_s3_uri_from_upload_data", |
| "LocalPath": "/container/path/", |
| "S3DataType": "Archive", |
| "S3InputMode": "File", |
| "S3DataDistributionType": "FullyReplicated", |
| "S3CompressionType": "None", |
| }, |
| }, |
| { |
| "InputName": "my_dataset", |
| "S3Input": { |
| "S3Uri": "s3://path/to/my/dataset/census.csv", |
| "LocalPath": "/container/path/", |
| "S3DataType": "Archive", |
| "S3InputMode": "File", |
| "S3DataDistributionType": "FullyReplicated", |
| "S3CompressionType": "None", |
| }, |
| }, |
| { |
| "InputName": "source", |
| "S3Input": { |
| "S3Uri": "mocked_s3_uri_from_upload_data", |
| "LocalPath": "/code/source", |
| "S3DataType": "Archive", |
| "S3InputMode": "File", |
| "S3DataDistributionType": "FullyReplicated", |
| "S3CompressionType": "None", |
| }, |
| }, |
| ], |
| "output_config": { |
| "Outputs": [ |
| { |
| "OutputName": "output-1", |
| "S3Output": { |
| "S3Uri": "s3://mybucket/current_job_name/output", |
| "LocalPath": "/data/output", |
| "S3UploadMode": "Continuous", |
| }, |
| }, |
| { |
| "OutputName": "my_output", |
| "S3Output": { |
| "S3Uri": "s3://uri/", |
| "LocalPath": "/container/path/", |
| "S3UploadMode": "Continuous", |
| }, |
| }, |
| ], |
| "KmsKeyId": "arn:aws:kms:us-west-2:012345678901:key/kms-key", |
| }, |
| "job_name": "current_job_name", |
| "resources": { |
| "ClusterConfig": { |
| "InstanceType": "ml.m4.xlarge", |
| "InstanceCount": 1, |
| "VolumeSizeInGB": 100, |
| } |
| }, |
| "stopping_condition": {"MaxRuntimeInSeconds": 3600}, |
| "app_specification": { |
| "ImageUri": "520713654638.dkr.ecr.us-west-2.amazonaws.com/sagemaker-scikit-learn:0.20.0-cpu-py3", |
| "ContainerArguments": ["--drop-columns", "'SelfEmployed'"], |
| "ContainerEntrypoint": ["python3", "/code/source/sklearn_transformer.py"], |
| }, |
| "environment": {"my_env_variable": 20}, |
| "network_config": { |
| "EnableInterContainerTrafficEncryption": True, |
| "EnableNetworkIsolation": True, |
| "VpcConfig": { |
| "SecurityGroupIds": ["my_security_group_id"], |
| "Subnets": ["my_subnet_id"], |
| }, |
| }, |
| "role_arn": ROLE, |
| "tags": [{"Name": "my-tag", "Value": "my-tag-value"}], |
| "experiment_config": {"ExperimentName": "AnExperiment"}, |
| } |
| session.process(**process_request_args) |
|
|
| expected_request = { |
| "ProcessingJobName": "current_job_name", |
| "ProcessingResources": { |
| "ClusterConfig": { |
| "InstanceType": "ml.m4.xlarge", |
| "InstanceCount": 1, |
| "VolumeSizeInGB": 100, |
| } |
| }, |
| "AppSpecification": { |
| "ImageUri": "520713654638.dkr.ecr.us-west-2.amazonaws.com/sagemaker-scikit-learn:0.20.0-cpu-py3", |
| "ContainerArguments": ["--drop-columns", "'SelfEmployed'"], |
| "ContainerEntrypoint": ["python3", "/code/source/sklearn_transformer.py"], |
| }, |
| "RoleArn": ROLE, |
| "ProcessingInputs": [ |
| { |
| "InputName": "input-1", |
| "S3Input": { |
| "S3Uri": "mocked_s3_uri_from_upload_data", |
| "LocalPath": "/container/path/", |
| "S3DataType": "Archive", |
| "S3InputMode": "File", |
| "S3DataDistributionType": "FullyReplicated", |
| "S3CompressionType": "None", |
| }, |
| }, |
| { |
| "InputName": "my_dataset", |
| "S3Input": { |
| "S3Uri": "s3://path/to/my/dataset/census.csv", |
| "LocalPath": "/container/path/", |
| "S3DataType": "Archive", |
| "S3InputMode": "File", |
| "S3DataDistributionType": "FullyReplicated", |
| "S3CompressionType": "None", |
| }, |
| }, |
| { |
| "InputName": "source", |
| "S3Input": { |
| "S3Uri": "mocked_s3_uri_from_upload_data", |
| "LocalPath": "/code/source", |
| "S3DataType": "Archive", |
| "S3InputMode": "File", |
| "S3DataDistributionType": "FullyReplicated", |
| "S3CompressionType": "None", |
| }, |
| }, |
| ], |
| "ProcessingOutputConfig": { |
| "Outputs": [ |
| { |
| "OutputName": "output-1", |
| "S3Output": { |
| "S3Uri": "s3://mybucket/current_job_name/output", |
| "LocalPath": "/data/output", |
| "S3UploadMode": "Continuous", |
| }, |
| }, |
| { |
| "OutputName": "my_output", |
| "S3Output": { |
| "S3Uri": "s3://uri/", |
| "LocalPath": "/container/path/", |
| "S3UploadMode": "Continuous", |
| }, |
| }, |
| ], |
| "KmsKeyId": "arn:aws:kms:us-west-2:012345678901:key/kms-key", |
| }, |
| "Environment": {"my_env_variable": 20}, |
| "NetworkConfig": { |
| "EnableInterContainerTrafficEncryption": True, |
| "EnableNetworkIsolation": True, |
| "VpcConfig": { |
| "SecurityGroupIds": ["my_security_group_id"], |
| "Subnets": ["my_subnet_id"], |
| }, |
| }, |
| "StoppingCondition": {"MaxRuntimeInSeconds": 3600}, |
| "Tags": [{"Name": "my-tag", "Value": "my-tag-value"}], |
| "ExperimentConfig": {"ExperimentName": "AnExperiment"}, |
| } |
|
|
| session.sagemaker_client.create_processing_job.assert_called_with(**expected_request) |
|
|
|
|
| def mock_exists(filepath_to_mock, exists_result): |
| unmocked_exists = os.path.exists |
|
|
| def side_effect(filepath): |
| if filepath == filepath_to_mock: |
| return exists_result |
| else: |
| return unmocked_exists(filepath) |
|
|
| return Mock(side_effect=side_effect) |
|
|
|
|
| def test_get_execution_role(): |
| session = Mock() |
| session.get_caller_identity_arn.return_value = "arn:aws:iam::369233609183:role/SageMakerRole" |
|
|
| actual = get_execution_role(session) |
| assert actual == "arn:aws:iam::369233609183:role/SageMakerRole" |
|
|
|
|
| def test_get_execution_role_works_with_service_role(): |
| session = Mock() |
| session.get_caller_identity_arn.return_value = ( |
| "arn:aws:iam::369233609183:role/service-role/AmazonSageMaker-ExecutionRole-20171129T072388" |
| ) |
|
|
| actual = get_execution_role(session) |
| assert ( |
| actual |
| == "arn:aws:iam::369233609183:role/service-role/AmazonSageMaker-ExecutionRole-20171129T072388" |
| ) |
|
|
|
|
| def test_get_execution_role_throws_exception_if_arn_is_not_role(): |
| session = Mock() |
| session.get_caller_identity_arn.return_value = "arn:aws:iam::369233609183:user/marcos" |
|
|
| with pytest.raises(ValueError) as error: |
| get_execution_role(session) |
| assert "The current AWS identity is not a role" in str(error.value) |
|
|
|
|
| def test_get_execution_role_throws_exception_if_arn_is_not_role_with_role_in_name(): |
| session = Mock() |
| session.get_caller_identity_arn.return_value = "arn:aws:iam::369233609183:user/marcos-role" |
|
|
| with pytest.raises(ValueError) as error: |
| get_execution_role(session) |
| assert "The current AWS identity is not a role" in str(error.value) |
|
|
|
|
| @patch("six.moves.builtins.open", mock_open(read_data='{"ResourceName": "SageMakerInstance"}')) |
| @patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, True)) |
| def test_get_caller_identity_arn_from_describe_notebook_instance(boto_session): |
| sess = Session(boto_session) |
| expected_role = "arn:aws:iam::369233609183:role/service-role/SageMakerRole-20171129T072388" |
| sess.sagemaker_client.describe_notebook_instance.return_value = {"RoleArn": expected_role} |
|
|
| actual = sess.get_caller_identity_arn() |
|
|
| assert actual == expected_role |
| sess.sagemaker_client.describe_notebook_instance.assert_called_once_with( |
| NotebookInstanceName="SageMakerInstance" |
| ) |
|
|
|
|
| @patch( |
| "six.moves.builtins.open", |
| mock_open( |
| read_data='{"ResourceName": "SageMakerInstance", ' |
| '"DomainId": "d-kbnw5yk6tg8j", ' |
| '"UserProfileName": "default-1617915559064"}' |
| ), |
| ) |
| @patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, True)) |
| def test_get_caller_identity_arn_from_describe_user_profile(boto_session): |
| sess = Session(boto_session) |
| expected_role = "arn:aws:iam::369233609183:role/service-role/SageMakerRole-20171129T072388" |
| sess.sagemaker_client.describe_user_profile.return_value = { |
| "UserSettings": {"ExecutionRole": expected_role} |
| } |
|
|
| actual = sess.get_caller_identity_arn() |
|
|
| assert actual == expected_role |
| sess.sagemaker_client.describe_user_profile.assert_called_once_with( |
| DomainId="d-kbnw5yk6tg8j", |
| UserProfileName="default-1617915559064", |
| ) |
|
|
|
|
| @patch( |
| "six.moves.builtins.open", |
| mock_open( |
| read_data='{"ResourceName": "SageMakerInstance", ' |
| '"DomainId": "d-kbnw5yk6tg8j", ' |
| '"UserProfileName": "default-1617915559064"}' |
| ), |
| ) |
| @patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, True)) |
| def test_get_caller_identity_arn_from_describe_domain_if_no_user_settings(boto_session): |
| sess = Session(boto_session) |
| expected_role = "arn:aws:iam::369233609183:role/service-role/SageMakerRole-20171129T072388" |
| sess.sagemaker_client.describe_user_profile.return_value = {} |
| sess.sagemaker_client.describe_domain.return_value = { |
| "DefaultUserSettings": {"ExecutionRole": expected_role} |
| } |
|
|
| actual = sess.get_caller_identity_arn() |
|
|
| assert actual == expected_role |
| sess.sagemaker_client.describe_user_profile.assert_called_once_with( |
| DomainId="d-kbnw5yk6tg8j", |
| UserProfileName="default-1617915559064", |
| ) |
| sess.sagemaker_client.describe_domain.assert_called_once_with(DomainId="d-kbnw5yk6tg8j") |
|
|
|
|
| @patch( |
| "six.moves.builtins.open", |
| mock_open( |
| read_data='{"ResourceName": "SageMakerInstance", ' |
| '"DomainId": "d-kbnw5yk6tg8j", ' |
| '"UserProfileName": "default-1617915559064"}' |
| ), |
| ) |
| @patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, True)) |
| def test_fallback_to_domain_if_role_unavailable_in_user_settings(boto_session): |
| sess = Session(boto_session) |
| expected_role = "expected_role" |
| sess.sagemaker_client.describe_user_profile.return_value = { |
| "DomainId": "d-kbnw5yk6tg8j", |
| "UserSettings": { |
| "JupyterServerAppSettings": {}, |
| "KernelGatewayAppSettings": {}, |
| }, |
| } |
|
|
| sess.sagemaker_client.describe_domain.return_value = { |
| "DefaultUserSettings": {"ExecutionRole": expected_role} |
| } |
|
|
| actual = sess.get_caller_identity_arn() |
|
|
| assert actual == expected_role |
| sess.sagemaker_client.describe_user_profile.assert_called_once_with( |
| DomainId="d-kbnw5yk6tg8j", |
| UserProfileName="default-1617915559064", |
| ) |
| sess.sagemaker_client.describe_domain.assert_called_once_with(DomainId="d-kbnw5yk6tg8j") |
|
|
|
|
| @patch("six.moves.builtins.open", mock_open(read_data='{"ResourceName": "SageMakerInstance"}')) |
| @patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, True)) |
| @patch("sagemaker.session.sts_regional_endpoint", return_value=STS_ENDPOINT) |
| def test_get_caller_identity_arn_from_a_role_after_describe_notebook_exception( |
| sts_regional_endpoint, boto_session |
| ): |
| sess = Session(boto_session) |
| exception = ClientError( |
| {"Error": {"Code": "ValidationException", "Message": "RecordNotFound"}}, "Operation" |
| ) |
| sess.sagemaker_client.describe_notebook_instance.side_effect = exception |
|
|
| arn = ( |
| "arn:aws:sts::369233609183:assumed-role/SageMakerRole/6d009ef3-5306-49d5-8efc-78db644d8122" |
| ) |
| sess.boto_session.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = { |
| "Arn": arn |
| } |
|
|
| expected_role = "arn:aws:iam::369233609183:role/SageMakerRole" |
| sess.boto_session.client("iam").get_role.return_value = {"Role": {"Arn": expected_role}} |
|
|
| with patch("logging.Logger.debug") as mock_logger: |
| actual = sess.get_caller_identity_arn() |
| mock_logger.assert_called_once() |
|
|
| sess.sagemaker_client.describe_notebook_instance.assert_called_once_with( |
| NotebookInstanceName="SageMakerInstance" |
| ) |
| assert actual == expected_role |
|
|
|
|
| @patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, False)) |
| @patch("sagemaker.session.sts_regional_endpoint", return_value=STS_ENDPOINT) |
| def test_get_caller_identity_arn_from_a_user(sts_regional_endpoint, boto_session): |
| sess = Session(boto_session) |
| arn = "arn:aws:iam::369233609183:user/mia" |
| sess.boto_session.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = { |
| "Arn": arn |
| } |
| sess.boto_session.client("iam").get_role.return_value = {"Role": {"Arn": arn}} |
|
|
| actual = sess.get_caller_identity_arn() |
| assert actual == "arn:aws:iam::369233609183:user/mia" |
|
|
|
|
| @patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, False)) |
| @patch("sagemaker.session.sts_regional_endpoint", return_value=STS_ENDPOINT) |
| def test_get_caller_identity_arn_from_an_user_without_permissions( |
| sts_regional_endpoint, boto_session |
| ): |
| sess = Session(boto_session) |
| arn = "arn:aws:iam::369233609183:user/mia" |
| sess.boto_session.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = { |
| "Arn": arn |
| } |
| sess.boto_session.client("iam").get_role.side_effect = ClientError({}, {}) |
|
|
| with patch("logging.Logger.warning") as mock_logger: |
| actual = sess.get_caller_identity_arn() |
| assert actual == "arn:aws:iam::369233609183:user/mia" |
| mock_logger.assert_called_once() |
|
|
|
|
| @patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, False)) |
| @patch("sagemaker.session.sts_regional_endpoint", return_value=STS_ENDPOINT) |
| def test_get_caller_identity_arn_from_a_role(sts_regional_endpoint, boto_session): |
| sess = Session(boto_session) |
| arn = ( |
| "arn:aws:sts::369233609183:assumed-role/SageMakerRole/6d009ef3-5306-49d5-8efc-78db644d8122" |
| ) |
| sess.boto_session.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = { |
| "Arn": arn |
| } |
|
|
| expected_role = "arn:aws:iam::369233609183:role/SageMakerRole" |
| sess.boto_session.client("iam").get_role.return_value = {"Role": {"Arn": expected_role}} |
|
|
| actual = sess.get_caller_identity_arn() |
| assert actual == expected_role |
|
|
|
|
| @patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, False)) |
| @patch("sagemaker.session.sts_regional_endpoint", return_value=STS_ENDPOINT) |
| def test_get_caller_identity_arn_from_an_execution_role(sts_regional_endpoint, boto_session): |
| sess = Session(boto_session) |
| sts_arn = "arn:aws:sts::369233609183:assumed-role/AmazonSageMaker-ExecutionRole-20171129T072388/SageMaker" |
| sess.boto_session.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = { |
| "Arn": sts_arn |
| } |
| iam_arn = "arn:aws:iam::369233609183:role/AmazonSageMaker-ExecutionRole-20171129T072388" |
| sess.boto_session.client("iam").get_role.return_value = {"Role": {"Arn": iam_arn}} |
|
|
| actual = sess.get_caller_identity_arn() |
| assert actual == iam_arn |
|
|
|
|
| @patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, False)) |
| @patch("sagemaker.session.sts_regional_endpoint", return_value=STS_ENDPOINT) |
| def test_get_caller_identity_arn_from_a_sagemaker_execution_role_with_iam_client_error( |
| sts_regional_endpoint, boto_session |
| ): |
| sess = Session(boto_session) |
| arn = "arn:aws:sts::369233609183:assumed-role/AmazonSageMaker-ExecutionRole-20171129T072388/SageMaker" |
| sess.boto_session.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = { |
| "Arn": arn |
| } |
|
|
| sess.boto_session.client("iam").get_role.side_effect = ClientError({}, {}) |
|
|
| actual = sess.get_caller_identity_arn() |
| assert ( |
| actual |
| == "arn:aws:iam::369233609183:role/service-role/AmazonSageMaker-ExecutionRole-20171129T072388" |
| ) |
|
|
|
|
| @patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, False)) |
| @patch("sagemaker.session.sts_regional_endpoint", return_value=STS_ENDPOINT) |
| def test_get_caller_identity_arn_from_role_with_path(sts_regional_endpoint, boto_session): |
| sess = Session(boto_session) |
| arn_prefix = "arn:aws:iam::369233609183:role" |
| role_name = "name" |
| sess.boto_session.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = { |
| "Arn": "/".join([arn_prefix, role_name]) |
| } |
|
|
| role_path = "path" |
| role_with_path = "/".join([arn_prefix, role_path, role_name]) |
| sess.boto_session.client("iam").get_role.return_value = {"Role": {"Arn": role_with_path}} |
|
|
| actual = sess.get_caller_identity_arn() |
| assert actual == role_with_path |
|
|
|
|
| def test_delete_endpoint(boto_session): |
| sess = Session(boto_session) |
| sess.delete_endpoint("my_endpoint") |
|
|
| boto_session.client().delete_endpoint.assert_called_with(EndpointName="my_endpoint") |
|
|
|
|
| def test_delete_endpoint_config(boto_session): |
| sess = Session(boto_session) |
| sess.delete_endpoint_config("my_endpoint_config") |
|
|
| boto_session.client().delete_endpoint_config.assert_called_with( |
| EndpointConfigName="my_endpoint_config" |
| ) |
|
|
|
|
| def test_delete_model(boto_session): |
| sess = Session(boto_session) |
|
|
| model_name = "my_model" |
| sess.delete_model(model_name) |
|
|
| boto_session.client().delete_model.assert_called_with(ModelName=model_name) |
|
|
|
|
| def test_user_agent_injected(boto_session): |
| assert ( |
| "AWS-SageMaker-Python-SDK" not in boto_session.client("sagemaker")._client_config.user_agent |
| ) |
|
|
| sess = Session(boto_session) |
|
|
| assert "AWS-SageMaker-Python-SDK" in sess.sagemaker_client._client_config.user_agent |
| assert "AWS-SageMaker-Python-SDK" in sess.sagemaker_runtime_client._client_config.user_agent |
| assert "AWS-SageMaker-Notebook-Instance" not in sess.sagemaker_client._client_config.user_agent |
| assert ( |
| "AWS-SageMaker-Notebook-Instance" |
| not in sess.sagemaker_runtime_client._client_config.user_agent |
| ) |
|
|
|
|
| def test_user_agent_injected_with_nbi(boto_session): |
| assert ( |
| "AWS-SageMaker-Python-SDK" not in boto_session.client("sagemaker")._client_config.user_agent |
| ) |
|
|
| with patch("six.moves.builtins.open", mock_open(read_data="120.0-0")) as mo: |
| sess = Session(boto_session) |
|
|
| mo.assert_called_with("/etc/opt/ml/sagemaker-notebook-instance-version.txt") |
|
|
| assert "AWS-SageMaker-Python-SDK" in sess.sagemaker_client._client_config.user_agent |
| assert "AWS-SageMaker-Python-SDK" in sess.sagemaker_runtime_client._client_config.user_agent |
| assert "AWS-SageMaker-Notebook-Instance" in sess.sagemaker_client._client_config.user_agent |
| assert ( |
| "AWS-SageMaker-Notebook-Instance" in sess.sagemaker_runtime_client._client_config.user_agent |
| ) |
|
|
|
|
| def test_user_agent_injected_with_nbi_ioerror(boto_session): |
| assert ( |
| "AWS-SageMaker-Python-SDK" not in boto_session.client("sagemaker")._client_config.user_agent |
| ) |
|
|
| with patch("six.moves.builtins.open", MagicMock(side_effect=IOError("File not found"))) as mo: |
| sess = Session(boto_session) |
|
|
| mo.assert_called_with("/etc/opt/ml/sagemaker-notebook-instance-version.txt") |
|
|
| assert "AWS-SageMaker-Python-SDK" in sess.sagemaker_client._client_config.user_agent |
| assert "AWS-SageMaker-Python-SDK" in sess.sagemaker_runtime_client._client_config.user_agent |
| assert "AWS-SageMaker-Notebook-Instance" not in sess.sagemaker_client._client_config.user_agent |
| assert ( |
| "AWS-SageMaker-Notebook-Instance" |
| not in sess.sagemaker_runtime_client._client_config.user_agent |
| ) |
|
|
|
|
| def test_training_input_all_defaults(): |
| prefix = "pre" |
| actual = TrainingInput(s3_data=prefix) |
| expected = { |
| "DataSource": { |
| "S3DataSource": { |
| "S3DataDistributionType": "FullyReplicated", |
| "S3DataType": "S3Prefix", |
| "S3Uri": prefix, |
| } |
| } |
| } |
| assert actual.config == expected |
|
|
|
|
| def test_training_input_all_arguments(): |
| prefix = "pre" |
| distribution = "FullyReplicated" |
| compression = "Gzip" |
| content_type = "text/csv" |
| record_wrapping = "RecordIO" |
| s3_data_type = "Manifestfile" |
| input_mode = "Pipe" |
| result = TrainingInput( |
| s3_data=prefix, |
| distribution=distribution, |
| compression=compression, |
| input_mode=input_mode, |
| content_type=content_type, |
| record_wrapping=record_wrapping, |
| s3_data_type=s3_data_type, |
| ) |
| expected = { |
| "DataSource": { |
| "S3DataSource": { |
| "S3DataDistributionType": distribution, |
| "S3DataType": s3_data_type, |
| "S3Uri": prefix, |
| } |
| }, |
| "CompressionType": compression, |
| "ContentType": content_type, |
| "RecordWrapperType": record_wrapping, |
| "InputMode": input_mode, |
| } |
|
|
| assert result.config == expected |
|
|
|
|
| IMAGE = "myimage" |
| S3_INPUT_URI = "s3://mybucket/data" |
| S3_OUTPUT = "s3://sagemaker-123/output/jobname" |
| ROLE = "SageMakerRole" |
| EXPANDED_ROLE = "arn:aws:iam::111111111111:role/ExpandedRole" |
| INSTANCE_COUNT = 1 |
| INSTANCE_TYPE = "ml.c4.xlarge" |
| ACCELERATOR_TYPE = "ml.eia.medium" |
| MAX_SIZE = 30 |
| MAX_TIME = 3 * 60 * 60 |
| JOB_NAME = "jobname" |
| TAGS = [{"Name": "some-tag", "Value": "value-for-tag"}] |
| VPC_CONFIG = {"Subnets": ["foo"], "SecurityGroupIds": ["bar"]} |
| METRIC_DEFINITONS = [{"Name": "validation-rmse", "Regex": "validation-rmse=(\\d+)"}] |
| EXPERIMENT_CONFIG = { |
| "ExperimentName": "dummyExp", |
| "TrialName": "dummyT", |
| "TrialComponentDisplayName": "dummyTC", |
| } |
| MODEL_CLIENT_CONFIG = {"InvocationsMaxRetries": 2, "InvocationsTimeoutInSeconds": 60} |
|
|
| DEFAULT_EXPECTED_TRAIN_JOB_ARGS = { |
| "OutputDataConfig": {"S3OutputPath": S3_OUTPUT}, |
| "RoleArn": EXPANDED_ROLE, |
| "ResourceConfig": { |
| "InstanceCount": INSTANCE_COUNT, |
| "InstanceType": INSTANCE_TYPE, |
| "VolumeSizeInGB": MAX_SIZE, |
| }, |
| "InputDataConfig": [ |
| { |
| "DataSource": { |
| "S3DataSource": { |
| "S3DataDistributionType": "FullyReplicated", |
| "S3DataType": "S3Prefix", |
| "S3Uri": S3_INPUT_URI, |
| } |
| }, |
| "ChannelName": "training", |
| } |
| ], |
| "AlgorithmSpecification": {"TrainingInputMode": "File", "TrainingImage": IMAGE}, |
| "TrainingJobName": JOB_NAME, |
| "StoppingCondition": {"MaxRuntimeInSeconds": MAX_TIME}, |
| "VpcConfig": VPC_CONFIG, |
| "ExperimentConfig": EXPERIMENT_CONFIG, |
| } |
|
|
| COMPLETED_DESCRIBE_JOB_RESULT = dict(DEFAULT_EXPECTED_TRAIN_JOB_ARGS) |
| COMPLETED_DESCRIBE_JOB_RESULT.update( |
| {"TrainingJobArn": "arn:aws:sagemaker:us-west-2:336:training-job/" + JOB_NAME} |
| ) |
| COMPLETED_DESCRIBE_JOB_RESULT.update({"TrainingJobStatus": "Completed"}) |
| COMPLETED_DESCRIBE_JOB_RESULT.update( |
| {"ModelArtifacts": {"S3ModelArtifacts": S3_OUTPUT + "/model/model.tar.gz"}} |
| ) |
| |
| COMPLETED_DESCRIBE_JOB_RESULT.update( |
| {"TrainingStartTime": datetime.datetime(2018, 2, 17, 7, 15, 0, 103000)} |
| ) |
| COMPLETED_DESCRIBE_JOB_RESULT.update( |
| {"TrainingEndTime": datetime.datetime(2018, 2, 17, 7, 19, 34, 953000)} |
| ) |
|
|
| STOPPED_DESCRIBE_JOB_RESULT = dict(COMPLETED_DESCRIBE_JOB_RESULT) |
| STOPPED_DESCRIBE_JOB_RESULT.update({"TrainingJobStatus": "Stopped"}) |
|
|
| IN_PROGRESS_DESCRIBE_JOB_RESULT = dict(DEFAULT_EXPECTED_TRAIN_JOB_ARGS) |
| IN_PROGRESS_DESCRIBE_JOB_RESULT.update({"TrainingJobStatus": "InProgress"}) |
|
|
| COMPLETED_DESCRIBE_TRANSFORM_JOB_RESULT = { |
| "TransformJobStatus": "Completed", |
| "ModelName": "some-model", |
| "TransformJobName": JOB_NAME, |
| "TransformResources": {"InstanceCount": INSTANCE_COUNT, "InstanceType": INSTANCE_TYPE}, |
| "TransformEndTime": datetime.datetime(2018, 2, 17, 7, 19, 34, 953000), |
| "TransformStartTime": datetime.datetime(2018, 2, 17, 7, 15, 0, 103000), |
| "TransformOutput": {"AssembleWith": "None", "KmsKeyId": "", "S3OutputPath": S3_OUTPUT}, |
| "TransformInput": { |
| "CompressionType": "None", |
| "ContentType": "text/csv", |
| "DataSource": {"S3DataType": "S3Prefix", "S3Uri": S3_INPUT_URI}, |
| "SplitType": "Line", |
| }, |
| } |
|
|
| STOPPED_DESCRIBE_TRANSFORM_JOB_RESULT = dict(COMPLETED_DESCRIBE_TRANSFORM_JOB_RESULT) |
| STOPPED_DESCRIBE_TRANSFORM_JOB_RESULT.update({"TransformJobStatus": "Stopped"}) |
|
|
| IN_PROGRESS_DESCRIBE_TRANSFORM_JOB_RESULT = dict(COMPLETED_DESCRIBE_TRANSFORM_JOB_RESULT) |
| IN_PROGRESS_DESCRIBE_TRANSFORM_JOB_RESULT.update({"TransformJobStatus": "InProgress"}) |
|
|
| SERVERLESS_INFERENCE_CONFIG = { |
| "MemorySizeInMB": 2048, |
| "MaxConcurrency": 2, |
| } |
|
|
|
|
| @pytest.fixture() |
| def sagemaker_session(): |
| boto_mock = MagicMock(name="boto_session") |
| boto_mock.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = { |
| "Account": "123" |
| } |
| ims = sagemaker.Session(boto_session=boto_mock, sagemaker_client=MagicMock()) |
| ims.expand_role = Mock(return_value=EXPANDED_ROLE) |
| return ims |
|
|
|
|
| def test_train_pack_to_request(sagemaker_session): |
| in_config = [ |
| { |
| "ChannelName": "training", |
| "DataSource": { |
| "S3DataSource": { |
| "S3DataDistributionType": "FullyReplicated", |
| "S3DataType": "S3Prefix", |
| "S3Uri": S3_INPUT_URI, |
| } |
| }, |
| } |
| ] |
|
|
| out_config = {"S3OutputPath": S3_OUTPUT} |
|
|
| resource_config = { |
| "InstanceCount": INSTANCE_COUNT, |
| "InstanceType": INSTANCE_TYPE, |
| "VolumeSizeInGB": MAX_SIZE, |
| } |
|
|
| stop_cond = {"MaxRuntimeInSeconds": MAX_TIME} |
|
|
| sagemaker_session.train( |
| image_uri=IMAGE, |
| input_mode="File", |
| input_config=in_config, |
| role=EXPANDED_ROLE, |
| job_name=JOB_NAME, |
| output_config=out_config, |
| resource_config=resource_config, |
| hyperparameters=None, |
| stop_condition=stop_cond, |
| tags=None, |
| vpc_config=VPC_CONFIG, |
| metric_definitions=None, |
| experiment_config=EXPERIMENT_CONFIG, |
| enable_sagemaker_metrics=None, |
| ) |
|
|
| assert sagemaker_session.sagemaker_client.method_calls[0] == ( |
| "create_training_job", |
| (), |
| DEFAULT_EXPECTED_TRAIN_JOB_ARGS, |
| ) |
|
|
|
|
| SAMPLE_STOPPING_CONDITION = {"MaxRuntimeInSeconds": MAX_TIME} |
|
|
| RESOURCE_CONFIG = { |
| "InstanceCount": INSTANCE_COUNT, |
| "InstanceType": INSTANCE_TYPE, |
| "VolumeSizeInGB": MAX_SIZE, |
| } |
|
|
| SAMPLE_INPUT = [ |
| { |
| "DataSource": { |
| "S3DataSource": { |
| "S3DataDistributionType": "FullyReplicated", |
| "S3DataType": "S3Prefix", |
| "S3Uri": S3_INPUT_URI, |
| } |
| }, |
| "ChannelName": "training", |
| } |
| ] |
|
|
| SAMPLE_OUTPUT = {"S3OutputPath": S3_OUTPUT} |
|
|
| SAMPLE_OBJECTIVE = {"Type": "Maximize", "MetricName": "val-score"} |
| SAMPLE_OBJECTIVE_2 = {"Type": "Maximize", "MetricName": "value-score"} |
|
|
| SAMPLE_METRIC_DEF = [{"Name": "train:progress", "Regex": "regex-1"}] |
| SAMPLE_METRIC_DEF_2 = [{"Name": "value-score", "Regex": "regex-2"}] |
|
|
| STATIC_HPs = {"feature_dim": "784"} |
| STATIC_HPs_2 = {"gamma": "0.1"} |
|
|
| SAMPLE_PARAM_RANGES = [{"Name": "mini_batch_size", "MinValue": "10", "MaxValue": "100"}] |
| SAMPLE_PARAM_RANGES_2 = [{"Name": "kernel", "Values": ["rbf", "sigmoid"]}] |
|
|
| SAMPLE_TUNING_JOB_REQUEST = { |
| "HyperParameterTuningJobName": "dummy-tuning-1", |
| "HyperParameterTuningJobConfig": { |
| "Strategy": "Bayesian", |
| "HyperParameterTuningJobObjective": SAMPLE_OBJECTIVE, |
| "ResourceLimits": {"MaxNumberOfTrainingJobs": 100, "MaxParallelTrainingJobs": 5}, |
| "ParameterRanges": SAMPLE_PARAM_RANGES, |
| "TrainingJobEarlyStoppingType": "Off", |
| }, |
| "TrainingJobDefinition": { |
| "StaticHyperParameters": STATIC_HPs, |
| "AlgorithmSpecification": { |
| "TrainingImage": "dummy-image-1", |
| "TrainingInputMode": "File", |
| "MetricDefinitions": SAMPLE_METRIC_DEF, |
| }, |
| "RoleArn": EXPANDED_ROLE, |
| "InputDataConfig": SAMPLE_INPUT, |
| "OutputDataConfig": SAMPLE_OUTPUT, |
| "ResourceConfig": RESOURCE_CONFIG, |
| "StoppingCondition": SAMPLE_STOPPING_CONDITION, |
| }, |
| } |
|
|
| SAMPLE_MULTI_ALGO_TUNING_JOB_REQUEST = { |
| "HyperParameterTuningJobName": "dummy-tuning-1", |
| "HyperParameterTuningJobConfig": { |
| "Strategy": "Bayesian", |
| "ResourceLimits": {"MaxNumberOfTrainingJobs": 100, "MaxParallelTrainingJobs": 5}, |
| "TrainingJobEarlyStoppingType": "Off", |
| }, |
| "TrainingJobDefinitions": [ |
| { |
| "DefinitionName": "estimator_1", |
| "TuningObjective": SAMPLE_OBJECTIVE, |
| "HyperParameterRanges": SAMPLE_PARAM_RANGES, |
| "StaticHyperParameters": STATIC_HPs, |
| "AlgorithmSpecification": { |
| "TrainingImage": "dummy-image-1", |
| "TrainingInputMode": "File", |
| "MetricDefinitions": SAMPLE_METRIC_DEF, |
| }, |
| "RoleArn": EXPANDED_ROLE, |
| "InputDataConfig": SAMPLE_INPUT, |
| "OutputDataConfig": SAMPLE_OUTPUT, |
| "ResourceConfig": RESOURCE_CONFIG, |
| "StoppingCondition": SAMPLE_STOPPING_CONDITION, |
| }, |
| { |
| "DefinitionName": "estimator_2", |
| "TuningObjective": SAMPLE_OBJECTIVE_2, |
| "HyperParameterRanges": SAMPLE_PARAM_RANGES_2, |
| "StaticHyperParameters": STATIC_HPs_2, |
| "AlgorithmSpecification": { |
| "TrainingImage": "dummy-image-2", |
| "TrainingInputMode": "File", |
| "MetricDefinitions": SAMPLE_METRIC_DEF_2, |
| }, |
| "RoleArn": EXPANDED_ROLE, |
| "InputDataConfig": SAMPLE_INPUT, |
| "OutputDataConfig": SAMPLE_OUTPUT, |
| "ResourceConfig": RESOURCE_CONFIG, |
| "StoppingCondition": SAMPLE_STOPPING_CONDITION, |
| }, |
| ], |
| } |
|
|
|
|
| @pytest.mark.parametrize( |
| "warm_start_type, parents", |
| [("IdenticalDataAndAlgorithm", {"p1", "p2", "p3"}), ("TransferLearning", {"p1", "p2", "p3"})], |
| ) |
| def test_tune_warm_start(sagemaker_session, warm_start_type, parents): |
| def assert_create_tuning_job_request(**kwrags): |
| assert ( |
| kwrags["HyperParameterTuningJobConfig"] |
| == SAMPLE_TUNING_JOB_REQUEST["HyperParameterTuningJobConfig"] |
| ) |
| assert kwrags["HyperParameterTuningJobName"] == "dummy-tuning-1" |
| assert kwrags["TrainingJobDefinition"] == SAMPLE_TUNING_JOB_REQUEST["TrainingJobDefinition"] |
| assert kwrags["WarmStartConfig"] == { |
| "WarmStartType": warm_start_type, |
| "ParentHyperParameterTuningJobs": [ |
| {"HyperParameterTuningJobName": parent} for parent in parents |
| ], |
| } |
|
|
| sagemaker_session.sagemaker_client.create_hyper_parameter_tuning_job.side_effect = ( |
| assert_create_tuning_job_request |
| ) |
| sagemaker_session.tune( |
| job_name="dummy-tuning-1", |
| strategy="Bayesian", |
| objective_type="Maximize", |
| objective_metric_name="val-score", |
| max_jobs=100, |
| max_parallel_jobs=5, |
| parameter_ranges=SAMPLE_PARAM_RANGES, |
| static_hyperparameters=STATIC_HPs, |
| image_uri="dummy-image-1", |
| input_mode="File", |
| metric_definitions=SAMPLE_METRIC_DEF, |
| role=EXPANDED_ROLE, |
| input_config=SAMPLE_INPUT, |
| output_config=SAMPLE_OUTPUT, |
| resource_config=RESOURCE_CONFIG, |
| stop_condition=SAMPLE_STOPPING_CONDITION, |
| tags=None, |
| warm_start_config=WarmStartConfig( |
| warm_start_type=WarmStartTypes(warm_start_type), parents=parents |
| ).to_input_req(), |
| ) |
|
|
|
|
| def test_create_tuning_job_without_training_config_or_list(sagemaker_session): |
| with pytest.raises( |
| ValueError, match="Either training_config or training_config_list should be provided." |
| ): |
| sagemaker_session.create_tuning_job( |
| job_name="dummy-tuning-1", |
| tuning_config={ |
| "strategy": "Bayesian", |
| "objective_type": "Maximize", |
| "objective_metric_name": "val-score", |
| "max_jobs": 100, |
| "max_parallel_jobs": 5, |
| "parameter_ranges": SAMPLE_PARAM_RANGES, |
| }, |
| ) |
|
|
|
|
| def test_create_tuning_job_with_both_training_config_and_list(sagemaker_session): |
| with pytest.raises( |
| ValueError, match="Only one of training_config and training_config_list should be provided." |
| ): |
| sagemaker_session.create_tuning_job( |
| job_name="dummy-tuning-1", |
| tuning_config={ |
| "strategy": "Bayesian", |
| "objective_type": "Maximize", |
| "objective_metric_name": "val-score", |
| "max_jobs": 100, |
| "max_parallel_jobs": 5, |
| "parameter_ranges": SAMPLE_PARAM_RANGES, |
| }, |
| training_config={"static_hyperparameters": STATIC_HPs, "image_uri": "dummy-image-1"}, |
| training_config_list=[ |
| { |
| "static_hyperparameters": STATIC_HPs, |
| "image_uri": "dummy-image-1", |
| "estimator_name": "estimator_1", |
| }, |
| { |
| "static_hyperparameters": STATIC_HPs_2, |
| "image_uri": "dummy-image-2", |
| "estimator_name": "estimator_2", |
| }, |
| ], |
| ) |
|
|
|
|
| def test_create_tuning_job(sagemaker_session): |
| def assert_create_tuning_job_request(**kwrags): |
| assert ( |
| kwrags["HyperParameterTuningJobConfig"] |
| == SAMPLE_TUNING_JOB_REQUEST["HyperParameterTuningJobConfig"] |
| ) |
| assert kwrags["HyperParameterTuningJobName"] == "dummy-tuning-1" |
| assert kwrags["TrainingJobDefinition"] == SAMPLE_TUNING_JOB_REQUEST["TrainingJobDefinition"] |
| assert "TrainingJobDefinitions" not in kwrags |
| assert kwrags.get("WarmStartConfig", None) is None |
|
|
| sagemaker_session.sagemaker_client.create_hyper_parameter_tuning_job.side_effect = ( |
| assert_create_tuning_job_request |
| ) |
| sagemaker_session.create_tuning_job( |
| job_name="dummy-tuning-1", |
| tuning_config={ |
| "strategy": "Bayesian", |
| "objective_type": "Maximize", |
| "objective_metric_name": "val-score", |
| "max_jobs": 100, |
| "max_parallel_jobs": 5, |
| "parameter_ranges": SAMPLE_PARAM_RANGES, |
| }, |
| training_config={ |
| "static_hyperparameters": STATIC_HPs, |
| "image_uri": "dummy-image-1", |
| "input_mode": "File", |
| "metric_definitions": SAMPLE_METRIC_DEF, |
| "role": EXPANDED_ROLE, |
| "input_config": SAMPLE_INPUT, |
| "output_config": SAMPLE_OUTPUT, |
| "resource_config": RESOURCE_CONFIG, |
| "stop_condition": SAMPLE_STOPPING_CONDITION, |
| }, |
| tags=None, |
| warm_start_config=None, |
| ) |
|
|
|
|
| def test_create_tuning_job_multi_algo(sagemaker_session): |
| def assert_create_tuning_job_request(**kwrags): |
| expected_tuning_config = SAMPLE_MULTI_ALGO_TUNING_JOB_REQUEST[ |
| "HyperParameterTuningJobConfig" |
| ] |
| assert kwrags["HyperParameterTuningJobConfig"] == expected_tuning_config |
| assert kwrags["HyperParameterTuningJobName"] == "dummy-tuning-1" |
| assert "TrainingJobDefinition" not in kwrags |
| assert ( |
| kwrags["TrainingJobDefinitions"] |
| == SAMPLE_MULTI_ALGO_TUNING_JOB_REQUEST["TrainingJobDefinitions"] |
| ) |
| assert kwrags.get("WarmStartConfig", None) is None |
|
|
| sagemaker_session.sagemaker_client.create_hyper_parameter_tuning_job.side_effect = ( |
| assert_create_tuning_job_request |
| ) |
| sagemaker_session.create_tuning_job( |
| job_name="dummy-tuning-1", |
| tuning_config={"strategy": "Bayesian", "max_jobs": 100, "max_parallel_jobs": 5}, |
| training_config_list=[ |
| { |
| "static_hyperparameters": STATIC_HPs, |
| "image_uri": "dummy-image-1", |
| "input_mode": "File", |
| "metric_definitions": SAMPLE_METRIC_DEF, |
| "role": EXPANDED_ROLE, |
| "input_config": SAMPLE_INPUT, |
| "output_config": SAMPLE_OUTPUT, |
| "resource_config": RESOURCE_CONFIG, |
| "stop_condition": SAMPLE_STOPPING_CONDITION, |
| "estimator_name": "estimator_1", |
| "objective_type": "Maximize", |
| "objective_metric_name": "val-score", |
| "parameter_ranges": SAMPLE_PARAM_RANGES, |
| }, |
| { |
| "static_hyperparameters": STATIC_HPs_2, |
| "image_uri": "dummy-image-2", |
| "input_mode": "File", |
| "metric_definitions": SAMPLE_METRIC_DEF_2, |
| "role": EXPANDED_ROLE, |
| "input_config": SAMPLE_INPUT, |
| "output_config": SAMPLE_OUTPUT, |
| "resource_config": RESOURCE_CONFIG, |
| "stop_condition": SAMPLE_STOPPING_CONDITION, |
| "estimator_name": "estimator_2", |
| "objective_type": "Maximize", |
| "objective_metric_name": "value-score", |
| "parameter_ranges": SAMPLE_PARAM_RANGES_2, |
| }, |
| ], |
| tags=None, |
| warm_start_config=None, |
| ) |
|
|
|
|
| def test_tune(sagemaker_session): |
| def assert_create_tuning_job_request(**kwrags): |
| assert ( |
| kwrags["HyperParameterTuningJobConfig"] |
| == SAMPLE_TUNING_JOB_REQUEST["HyperParameterTuningJobConfig"] |
| ) |
| assert kwrags["HyperParameterTuningJobName"] == "dummy-tuning-1" |
| assert kwrags["TrainingJobDefinition"] == SAMPLE_TUNING_JOB_REQUEST["TrainingJobDefinition"] |
| assert kwrags.get("WarmStartConfig", None) is None |
|
|
| sagemaker_session.sagemaker_client.create_hyper_parameter_tuning_job.side_effect = ( |
| assert_create_tuning_job_request |
| ) |
| sagemaker_session.tune( |
| job_name="dummy-tuning-1", |
| strategy="Bayesian", |
| objective_type="Maximize", |
| objective_metric_name="val-score", |
| max_jobs=100, |
| max_parallel_jobs=5, |
| parameter_ranges=SAMPLE_PARAM_RANGES, |
| static_hyperparameters=STATIC_HPs, |
| image_uri="dummy-image-1", |
| input_mode="File", |
| metric_definitions=SAMPLE_METRIC_DEF, |
| role=EXPANDED_ROLE, |
| input_config=SAMPLE_INPUT, |
| output_config=SAMPLE_OUTPUT, |
| resource_config=RESOURCE_CONFIG, |
| stop_condition=SAMPLE_STOPPING_CONDITION, |
| tags=None, |
| warm_start_config=None, |
| ) |
|
|
|
|
| def test_tune_with_encryption_flag(sagemaker_session): |
| def assert_create_tuning_job_request(**kwrags): |
| assert ( |
| kwrags["HyperParameterTuningJobConfig"] |
| == SAMPLE_TUNING_JOB_REQUEST["HyperParameterTuningJobConfig"] |
| ) |
| assert kwrags["HyperParameterTuningJobName"] == "dummy-tuning-1" |
| assert kwrags["TrainingJobDefinition"]["EnableInterContainerTrafficEncryption"] is True |
| assert kwrags.get("WarmStartConfig", None) is None |
|
|
| sagemaker_session.sagemaker_client.create_hyper_parameter_tuning_job.side_effect = ( |
| assert_create_tuning_job_request |
| ) |
| sagemaker_session.tune( |
| job_name="dummy-tuning-1", |
| strategy="Bayesian", |
| objective_type="Maximize", |
| objective_metric_name="val-score", |
| max_jobs=100, |
| max_parallel_jobs=5, |
| parameter_ranges=SAMPLE_PARAM_RANGES, |
| static_hyperparameters=STATIC_HPs, |
| image_uri="dummy-image-1", |
| input_mode="File", |
| metric_definitions=SAMPLE_METRIC_DEF, |
| role=EXPANDED_ROLE, |
| input_config=SAMPLE_INPUT, |
| output_config=SAMPLE_OUTPUT, |
| resource_config=RESOURCE_CONFIG, |
| stop_condition=SAMPLE_STOPPING_CONDITION, |
| tags=None, |
| warm_start_config=None, |
| encrypt_inter_container_traffic=True, |
| ) |
|
|
|
|
| def test_tune_with_spot_and_checkpoints(sagemaker_session): |
| def assert_create_tuning_job_request(**kwargs): |
| assert ( |
| kwargs["HyperParameterTuningJobConfig"] |
| == SAMPLE_TUNING_JOB_REQUEST["HyperParameterTuningJobConfig"] |
| ) |
| assert kwargs["HyperParameterTuningJobName"] == "dummy-tuning-1" |
| assert kwargs["TrainingJobDefinition"]["EnableManagedSpotTraining"] is True |
| assert ( |
| kwargs["TrainingJobDefinition"]["CheckpointConfig"]["S3Uri"] |
| == "s3://mybucket/checkpoints/" |
| ) |
| assert ( |
| kwargs["TrainingJobDefinition"]["CheckpointConfig"]["LocalPath"] == "/tmp/checkpoints" |
| ) |
| assert kwargs.get("WarmStartConfig", None) is None |
|
|
| sagemaker_session.sagemaker_client.create_hyper_parameter_tuning_job.side_effect = ( |
| assert_create_tuning_job_request |
| ) |
| sagemaker_session.tune( |
| job_name="dummy-tuning-1", |
| strategy="Bayesian", |
| objective_type="Maximize", |
| objective_metric_name="val-score", |
| max_jobs=100, |
| max_parallel_jobs=5, |
| parameter_ranges=SAMPLE_PARAM_RANGES, |
| static_hyperparameters=STATIC_HPs, |
| image_uri="dummy-image-1", |
| input_mode="File", |
| metric_definitions=SAMPLE_METRIC_DEF, |
| role=EXPANDED_ROLE, |
| input_config=SAMPLE_INPUT, |
| output_config=SAMPLE_OUTPUT, |
| resource_config=RESOURCE_CONFIG, |
| stop_condition=SAMPLE_STOPPING_CONDITION, |
| tags=None, |
| warm_start_config=None, |
| use_spot_instances=True, |
| checkpoint_s3_uri="s3://mybucket/checkpoints/", |
| checkpoint_local_path="/tmp/checkpoints", |
| ) |
|
|
|
|
| def test_stop_tuning_job(sagemaker_session): |
| sms = sagemaker_session |
| sms.sagemaker_client.stop_hyper_parameter_tuning_job = Mock( |
| name="stop_hyper_parameter_tuning_job" |
| ) |
|
|
| sagemaker_session.stop_tuning_job(JOB_NAME) |
| sms.sagemaker_client.stop_hyper_parameter_tuning_job.assert_called_once_with( |
| HyperParameterTuningJobName=JOB_NAME |
| ) |
|
|
|
|
| def test_stop_tuning_job_client_error_already_stopped(sagemaker_session): |
| sms = sagemaker_session |
| exception = ClientError({"Error": {"Code": "ValidationException"}}, "Operation") |
| sms.sagemaker_client.stop_hyper_parameter_tuning_job = Mock( |
| name="stop_hyper_parameter_tuning_job", side_effect=exception |
| ) |
| sagemaker_session.stop_tuning_job(JOB_NAME) |
|
|
| sms.sagemaker_client.stop_hyper_parameter_tuning_job.assert_called_once_with( |
| HyperParameterTuningJobName=JOB_NAME |
| ) |
|
|
|
|
| def test_stop_tuning_job_client_error(sagemaker_session): |
| error_response = {"Error": {"Code": "MockException", "Message": "MockMessage"}} |
| operation = "Operation" |
| exception = ClientError(error_response, operation) |
|
|
| sms = sagemaker_session |
| sms.sagemaker_client.stop_hyper_parameter_tuning_job = Mock( |
| name="stop_hyper_parameter_tuning_job", side_effect=exception |
| ) |
|
|
| with pytest.raises(ClientError) as e: |
| sagemaker_session.stop_tuning_job(JOB_NAME) |
|
|
| sms.sagemaker_client.stop_hyper_parameter_tuning_job.assert_called_once_with( |
| HyperParameterTuningJobName=JOB_NAME |
| ) |
| assert ( |
| "An error occurred (MockException) when calling the Operation operation: MockMessage" |
| in str(e) |
| ) |
|
|
|
|
| def test_train_pack_to_request_with_optional_params(sagemaker_session): |
| in_config = [ |
| { |
| "ChannelName": "training", |
| "DataSource": { |
| "S3DataSource": { |
| "S3DataDistributionType": "FullyReplicated", |
| "S3DataType": "S3Prefix", |
| "S3Uri": S3_INPUT_URI, |
| } |
| }, |
| } |
| ] |
|
|
| out_config = {"S3OutputPath": S3_OUTPUT} |
|
|
| resource_config = { |
| "InstanceCount": INSTANCE_COUNT, |
| "InstanceType": INSTANCE_TYPE, |
| "VolumeSizeInGB": MAX_SIZE, |
| } |
|
|
| stop_cond = {"MaxRuntimeInSeconds": MAX_TIME} |
| RETRY_STRATEGY = {"MaximumRetryAttempts": 2} |
| hyperparameters = {"foo": "bar"} |
|
|
| sagemaker_session.train( |
| image_uri=IMAGE, |
| input_mode="File", |
| input_config=in_config, |
| role=EXPANDED_ROLE, |
| job_name=JOB_NAME, |
| output_config=out_config, |
| resource_config=resource_config, |
| vpc_config=VPC_CONFIG, |
| hyperparameters=hyperparameters, |
| stop_condition=stop_cond, |
| tags=TAGS, |
| metric_definitions=METRIC_DEFINITONS, |
| encrypt_inter_container_traffic=True, |
| use_spot_instances=True, |
| checkpoint_s3_uri="s3://mybucket/checkpoints/", |
| checkpoint_local_path="/tmp/checkpoints", |
| enable_sagemaker_metrics=True, |
| environment=ENV_INPUT, |
| retry_strategy=RETRY_STRATEGY, |
| ) |
|
|
| _, _, actual_train_args = sagemaker_session.sagemaker_client.method_calls[0] |
|
|
| assert actual_train_args["VpcConfig"] == VPC_CONFIG |
| assert actual_train_args["HyperParameters"] == hyperparameters |
| assert actual_train_args["Tags"] == TAGS |
| assert actual_train_args["AlgorithmSpecification"]["MetricDefinitions"] == METRIC_DEFINITONS |
| assert actual_train_args["AlgorithmSpecification"]["EnableSageMakerMetricsTimeSeries"] is True |
| assert actual_train_args["EnableInterContainerTrafficEncryption"] is True |
| assert actual_train_args["EnableManagedSpotTraining"] is True |
| assert actual_train_args["CheckpointConfig"]["S3Uri"] == "s3://mybucket/checkpoints/" |
| assert actual_train_args["CheckpointConfig"]["LocalPath"] == "/tmp/checkpoints" |
| assert actual_train_args["Environment"] == ENV_INPUT |
| assert actual_train_args["RetryStrategy"] == RETRY_STRATEGY |
|
|
|
|
| def test_transform_pack_to_request(sagemaker_session): |
| model_name = "my-model" |
|
|
| in_config = { |
| "CompressionType": "None", |
| "ContentType": "text/csv", |
| "SplitType": "None", |
| "DataSource": {"S3DataSource": {"S3DataType": "S3Prefix", "S3Uri": S3_INPUT_URI}}, |
| } |
|
|
| out_config = {"S3OutputPath": S3_OUTPUT} |
|
|
| resource_config = {"InstanceCount": INSTANCE_COUNT, "InstanceType": INSTANCE_TYPE} |
|
|
| data_processing = {"OutputFilter": "$", "InputFilter": "$", "JoinSource": "Input"} |
|
|
| expected_args = { |
| "TransformJobName": JOB_NAME, |
| "ModelName": model_name, |
| "TransformInput": in_config, |
| "TransformOutput": out_config, |
| "TransformResources": resource_config, |
| "DataProcessing": data_processing, |
| } |
|
|
| sagemaker_session.transform( |
| job_name=JOB_NAME, |
| model_name=model_name, |
| strategy=None, |
| max_concurrent_transforms=None, |
| max_payload=None, |
| env=None, |
| input_config=in_config, |
| output_config=out_config, |
| resource_config=resource_config, |
| experiment_config=None, |
| model_client_config=None, |
| tags=None, |
| data_processing=data_processing, |
| batch_data_capture_config=None, |
| ) |
|
|
| _, _, actual_args = sagemaker_session.sagemaker_client.method_calls[0] |
| assert actual_args == expected_args |
|
|
|
|
| def test_transform_pack_to_request_with_optional_params(sagemaker_session): |
| strategy = "strategy" |
| max_concurrent_transforms = 1 |
| max_payload = 0 |
| env = {"FOO": "BAR"} |
|
|
| batch_data_capture_config = BatchDataCaptureConfig( |
| destination_s3_uri="test_uri", |
| kms_key_id="", |
| generate_inference_id=False, |
| ) |
|
|
| sagemaker_session.transform( |
| job_name=JOB_NAME, |
| model_name="my-model", |
| strategy=strategy, |
| max_concurrent_transforms=max_concurrent_transforms, |
| env=env, |
| max_payload=max_payload, |
| input_config={}, |
| output_config={}, |
| resource_config={}, |
| experiment_config=EXPERIMENT_CONFIG, |
| model_client_config=MODEL_CLIENT_CONFIG, |
| tags=TAGS, |
| data_processing=None, |
| batch_data_capture_config=batch_data_capture_config, |
| ) |
|
|
| _, _, actual_args = sagemaker_session.sagemaker_client.method_calls[0] |
| assert actual_args["BatchStrategy"] == strategy |
| assert actual_args["MaxConcurrentTransforms"] == max_concurrent_transforms |
| assert actual_args["MaxPayloadInMB"] == max_payload |
| assert actual_args["Environment"] == env |
| assert actual_args["Tags"] == TAGS |
| assert actual_args["ExperimentConfig"] == EXPERIMENT_CONFIG |
| assert actual_args["ModelClientConfig"] == MODEL_CLIENT_CONFIG |
| assert actual_args["DataCaptureConfig"] == batch_data_capture_config._to_request_dict() |
|
|
|
|
| @patch("sys.stdout", new_callable=io.BytesIO if six.PY2 else io.StringIO) |
| def test_color_wrap(bio): |
| color_wrap = sagemaker.logs.ColorWrap() |
| color_wrap(0, "hi there") |
| assert bio.getvalue() == "hi there\n" |
|
|
|
|
| class MockBotoException(ClientError): |
| def __init__(self, code): |
| self.response = {"Error": {"Code": code}} |
|
|
|
|
| DEFAULT_LOG_STREAMS = {"logStreams": [{"logStreamName": JOB_NAME + "/xxxxxxxxx"}]} |
| LIFECYCLE_LOG_STREAMS = [ |
| MockBotoException("ResourceNotFoundException"), |
| DEFAULT_LOG_STREAMS, |
| DEFAULT_LOG_STREAMS, |
| DEFAULT_LOG_STREAMS, |
| DEFAULT_LOG_STREAMS, |
| DEFAULT_LOG_STREAMS, |
| DEFAULT_LOG_STREAMS, |
| ] |
|
|
| DEFAULT_LOG_EVENTS = [ |
| {"nextForwardToken": None, "events": [{"timestamp": 1, "message": "hi there #1"}]}, |
| {"nextForwardToken": None, "events": []}, |
| ] |
| STREAM_LOG_EVENTS = [ |
| {"nextForwardToken": None, "events": [{"timestamp": 1, "message": "hi there #1"}]}, |
| {"nextForwardToken": None, "events": []}, |
| { |
| "nextForwardToken": None, |
| "events": [ |
| {"timestamp": 1, "message": "hi there #1"}, |
| {"timestamp": 2, "message": "hi there #2"}, |
| ], |
| }, |
| {"nextForwardToken": None, "events": []}, |
| { |
| "nextForwardToken": None, |
| "events": [ |
| {"timestamp": 2, "message": "hi there #2"}, |
| {"timestamp": 2, "message": "hi there #2a"}, |
| {"timestamp": 3, "message": "hi there #3"}, |
| ], |
| }, |
| {"nextForwardToken": None, "events": []}, |
| ] |
|
|
|
|
| @pytest.fixture() |
| def sagemaker_session_complete(): |
| boto_mock = MagicMock(name="boto_session") |
| boto_mock.client("logs").describe_log_streams.return_value = DEFAULT_LOG_STREAMS |
| boto_mock.client("logs").get_log_events.side_effect = DEFAULT_LOG_EVENTS |
| ims = sagemaker.Session(boto_session=boto_mock, sagemaker_client=MagicMock()) |
| ims.sagemaker_client.describe_training_job.return_value = COMPLETED_DESCRIBE_JOB_RESULT |
| ims.sagemaker_client.describe_transform_job.return_value = ( |
| COMPLETED_DESCRIBE_TRANSFORM_JOB_RESULT |
| ) |
| return ims |
|
|
|
|
| @pytest.fixture() |
| def sagemaker_session_stopped(): |
| boto_mock = MagicMock(name="boto_session") |
| boto_mock.client("logs").describe_log_streams.return_value = DEFAULT_LOG_STREAMS |
| boto_mock.client("logs").get_log_events.side_effect = DEFAULT_LOG_EVENTS |
| ims = sagemaker.Session(boto_session=boto_mock, sagemaker_client=MagicMock()) |
| ims.sagemaker_client.describe_training_job.return_value = STOPPED_DESCRIBE_JOB_RESULT |
| ims.sagemaker_client.describe_transform_job.return_value = STOPPED_DESCRIBE_TRANSFORM_JOB_RESULT |
| return ims |
|
|
|
|
| @pytest.fixture() |
| def sagemaker_session_ready_lifecycle(): |
| boto_mock = MagicMock(name="boto_session") |
| boto_mock.client("logs").describe_log_streams.return_value = DEFAULT_LOG_STREAMS |
| boto_mock.client("logs").get_log_events.side_effect = STREAM_LOG_EVENTS |
| ims = sagemaker.Session(boto_session=boto_mock, sagemaker_client=MagicMock()) |
| ims.sagemaker_client.describe_training_job.side_effect = [ |
| IN_PROGRESS_DESCRIBE_JOB_RESULT, |
| IN_PROGRESS_DESCRIBE_JOB_RESULT, |
| COMPLETED_DESCRIBE_JOB_RESULT, |
| ] |
| ims.sagemaker_client.describe_transform_job.side_effect = [ |
| IN_PROGRESS_DESCRIBE_TRANSFORM_JOB_RESULT, |
| IN_PROGRESS_DESCRIBE_TRANSFORM_JOB_RESULT, |
| COMPLETED_DESCRIBE_TRANSFORM_JOB_RESULT, |
| ] |
| return ims |
|
|
|
|
| @pytest.fixture() |
| def sagemaker_session_full_lifecycle(): |
| boto_mock = MagicMock(name="boto_session") |
| boto_mock.client("logs").describe_log_streams.side_effect = LIFECYCLE_LOG_STREAMS |
| boto_mock.client("logs").get_log_events.side_effect = STREAM_LOG_EVENTS |
| ims = sagemaker.Session(boto_session=boto_mock, sagemaker_client=MagicMock()) |
| ims.sagemaker_client.describe_training_job.side_effect = [ |
| IN_PROGRESS_DESCRIBE_JOB_RESULT, |
| IN_PROGRESS_DESCRIBE_JOB_RESULT, |
| COMPLETED_DESCRIBE_JOB_RESULT, |
| ] |
| ims.sagemaker_client.describe_transform_job.side_effect = [ |
| IN_PROGRESS_DESCRIBE_TRANSFORM_JOB_RESULT, |
| IN_PROGRESS_DESCRIBE_TRANSFORM_JOB_RESULT, |
| COMPLETED_DESCRIBE_TRANSFORM_JOB_RESULT, |
| ] |
| return ims |
|
|
|
|
| @patch("sagemaker.logs.ColorWrap") |
| def test_logs_for_job_no_wait(cw, sagemaker_session_complete): |
| ims = sagemaker_session_complete |
| ims.logs_for_job(JOB_NAME) |
| ims.sagemaker_client.describe_training_job.assert_called_once_with(TrainingJobName=JOB_NAME) |
| cw().assert_called_with(0, "hi there #1") |
|
|
|
|
| @patch("sagemaker.logs.ColorWrap") |
| def test_logs_for_job_no_wait_stopped_job(cw, sagemaker_session_stopped): |
| ims = sagemaker_session_stopped |
| ims.logs_for_job(JOB_NAME) |
| ims.sagemaker_client.describe_training_job.assert_called_once_with(TrainingJobName=JOB_NAME) |
| cw().assert_called_with(0, "hi there #1") |
|
|
|
|
| @patch("sagemaker.logs.ColorWrap") |
| def test_logs_for_job_wait_on_completed(cw, sagemaker_session_complete): |
| ims = sagemaker_session_complete |
| ims.logs_for_job(JOB_NAME, wait=True, poll=0) |
| assert ims.sagemaker_client.describe_training_job.call_args_list == [ |
| call(TrainingJobName=JOB_NAME) |
| ] |
| cw().assert_called_with(0, "hi there #1") |
|
|
|
|
| @patch("sagemaker.logs.ColorWrap") |
| def test_logs_for_job_wait_on_stopped(cw, sagemaker_session_stopped): |
| ims = sagemaker_session_stopped |
| ims.logs_for_job(JOB_NAME, wait=True, poll=0) |
| assert ims.sagemaker_client.describe_training_job.call_args_list == [ |
| call(TrainingJobName=JOB_NAME) |
| ] |
| cw().assert_called_with(0, "hi there #1") |
|
|
|
|
| @patch("sagemaker.logs.ColorWrap") |
| def test_logs_for_job_no_wait_on_running(cw, sagemaker_session_ready_lifecycle): |
| ims = sagemaker_session_ready_lifecycle |
| ims.logs_for_job(JOB_NAME) |
| assert ims.sagemaker_client.describe_training_job.call_args_list == [ |
| call(TrainingJobName=JOB_NAME) |
| ] |
| cw().assert_called_with(0, "hi there #1") |
|
|
|
|
| @patch("sagemaker.logs.ColorWrap") |
| @patch("time.time", side_effect=[0, 30, 60, 90, 120, 150, 180]) |
| def test_logs_for_job_full_lifecycle(time, cw, sagemaker_session_full_lifecycle): |
| ims = sagemaker_session_full_lifecycle |
| ims.logs_for_job(JOB_NAME, wait=True, poll=0) |
| assert ( |
| ims.sagemaker_client.describe_training_job.call_args_list |
| == [call(TrainingJobName=JOB_NAME)] * 3 |
| ) |
| assert cw().call_args_list == [ |
| call(0, "hi there #1"), |
| call(0, "hi there #2"), |
| call(0, "hi there #2a"), |
| call(0, "hi there #3"), |
| ] |
|
|
|
|
| @patch("sagemaker.logs.ColorWrap") |
| def test_logs_for_transform_job_no_wait(cw, sagemaker_session_complete): |
| ims = sagemaker_session_complete |
| ims.logs_for_transform_job(JOB_NAME) |
| ims.sagemaker_client.describe_transform_job.assert_called_once_with(TransformJobName=JOB_NAME) |
| cw().assert_called_with(0, "hi there #1") |
|
|
|
|
| @patch("sagemaker.logs.ColorWrap") |
| def test_logs_for_transform_job_no_wait_stopped_job(cw, sagemaker_session_stopped): |
| ims = sagemaker_session_stopped |
| ims.logs_for_transform_job(JOB_NAME) |
| ims.sagemaker_client.describe_transform_job.assert_called_once_with(TransformJobName=JOB_NAME) |
| cw().assert_called_with(0, "hi there #1") |
|
|
|
|
| @patch("sagemaker.logs.ColorWrap") |
| def test_logs_for_transform_job_wait_on_completed(cw, sagemaker_session_complete): |
| ims = sagemaker_session_complete |
| ims.logs_for_transform_job(JOB_NAME, wait=True, poll=0) |
| assert ims.sagemaker_client.describe_transform_job.call_args_list == [ |
| call(TransformJobName=JOB_NAME) |
| ] |
| cw().assert_called_with(0, "hi there #1") |
|
|
|
|
| @patch("sagemaker.logs.ColorWrap") |
| def test_logs_for_transform_job_wait_on_stopped(cw, sagemaker_session_stopped): |
| ims = sagemaker_session_stopped |
| ims.logs_for_transform_job(JOB_NAME, wait=True, poll=0) |
| assert ims.sagemaker_client.describe_transform_job.call_args_list == [ |
| call(TransformJobName=JOB_NAME) |
| ] |
| cw().assert_called_with(0, "hi there #1") |
|
|
|
|
| @patch("sagemaker.logs.ColorWrap") |
| def test_logs_for_transform_job_no_wait_on_running(cw, sagemaker_session_ready_lifecycle): |
| ims = sagemaker_session_ready_lifecycle |
| ims.logs_for_transform_job(JOB_NAME) |
| assert ims.sagemaker_client.describe_transform_job.call_args_list == [ |
| call(TransformJobName=JOB_NAME) |
| ] |
| cw().assert_called_with(0, "hi there #1") |
|
|
|
|
| @patch("sagemaker.logs.ColorWrap") |
| @patch("time.time", side_effect=[0, 30, 60, 90, 120, 150, 180]) |
| def test_logs_for_transform_job_full_lifecycle(time, cw, sagemaker_session_full_lifecycle): |
| ims = sagemaker_session_full_lifecycle |
| ims.logs_for_transform_job(JOB_NAME, wait=True, poll=0) |
| assert ( |
| ims.sagemaker_client.describe_transform_job.call_args_list |
| == [call(TransformJobName=JOB_NAME)] * 3 |
| ) |
| assert cw().call_args_list == [ |
| call(0, "hi there #1"), |
| call(0, "hi there #2"), |
| call(0, "hi there #2a"), |
| call(0, "hi there #3"), |
| ] |
|
|
|
|
| MODEL_NAME = "some-model" |
| PRIMARY_CONTAINER = { |
| "Environment": {}, |
| "Image": IMAGE, |
| "ModelDataUrl": "s3://sagemaker-123/output/jobname/model/model.tar.gz", |
| } |
|
|
|
|
| @patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER) |
| def test_create_model(expand_container_def, sagemaker_session): |
| model = sagemaker_session.create_model(MODEL_NAME, ROLE, PRIMARY_CONTAINER) |
|
|
| assert model == MODEL_NAME |
| sagemaker_session.sagemaker_client.create_model.assert_called_with( |
| ExecutionRoleArn=EXPANDED_ROLE, ModelName=MODEL_NAME, PrimaryContainer=PRIMARY_CONTAINER |
| ) |
|
|
|
|
| @patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER) |
| def test_create_model_with_tags(expand_container_def, sagemaker_session): |
| tags = [{"Key": "TagtestKey", "Value": "TagtestValue"}] |
| model = sagemaker_session.create_model(MODEL_NAME, ROLE, PRIMARY_CONTAINER, tags=tags) |
|
|
| assert model == MODEL_NAME |
| tags = [{"Value": "TagtestValue", "Key": "TagtestKey"}] |
| sagemaker_session.sagemaker_client.create_model.assert_called_with( |
| ExecutionRoleArn=EXPANDED_ROLE, |
| ModelName=MODEL_NAME, |
| PrimaryContainer=PRIMARY_CONTAINER, |
| Tags=tags, |
| ) |
|
|
|
|
| @patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER) |
| def test_create_model_with_primary_container(expand_container_def, sagemaker_session): |
| model = sagemaker_session.create_model(MODEL_NAME, ROLE, container_defs=PRIMARY_CONTAINER) |
|
|
| assert model == MODEL_NAME |
| sagemaker_session.sagemaker_client.create_model.assert_called_with( |
| ExecutionRoleArn=EXPANDED_ROLE, ModelName=MODEL_NAME, PrimaryContainer=PRIMARY_CONTAINER |
| ) |
|
|
|
|
| @patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER) |
| def test_create_model_with_both(expand_container_def, sagemaker_session): |
| with pytest.raises(ValueError): |
| sagemaker_session.create_model( |
| MODEL_NAME, ROLE, container_defs=PRIMARY_CONTAINER, primary_container=PRIMARY_CONTAINER |
| ) |
|
|
|
|
| CONTAINERS = [ |
| { |
| "Environment": {"SAGEMAKER_DEFAULT_INVOCATIONS_ACCEPT": "application/json"}, |
| "Image": "mi-1", |
| "ModelDataUrl": "s3://bucket/model_1.tar.gz", |
| "Framework": "TENSORFLOW", |
| "FrameworkVersion": "2.9", |
| "NearestModelName": "resnet50", |
| "ModelInput": { |
| "DataInputConfig": '{"input_1":[1,224,224,3]}', |
| }, |
| }, |
| {"Environment": {}, "Image": "mi-2", "ModelDataUrl": "s3://bucket/model_2.tar.gz"}, |
| ] |
|
|
|
|
| @patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER) |
| def test_create_pipeline_model(expand_container_def, sagemaker_session): |
| model = sagemaker_session.create_model(MODEL_NAME, ROLE, container_defs=CONTAINERS) |
|
|
| assert model == MODEL_NAME |
| sagemaker_session.sagemaker_client.create_model.assert_called_with( |
| ExecutionRoleArn=EXPANDED_ROLE, ModelName=MODEL_NAME, Containers=CONTAINERS |
| ) |
|
|
|
|
| @patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER) |
| def test_create_model_vpc_config(expand_container_def, sagemaker_session): |
| model = sagemaker_session.create_model(MODEL_NAME, ROLE, PRIMARY_CONTAINER, VPC_CONFIG) |
|
|
| assert model == MODEL_NAME |
| sagemaker_session.sagemaker_client.create_model.assert_called_with( |
| ExecutionRoleArn=EXPANDED_ROLE, |
| ModelName=MODEL_NAME, |
| PrimaryContainer=PRIMARY_CONTAINER, |
| VpcConfig=VPC_CONFIG, |
| ) |
|
|
|
|
| @patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER) |
| def test_create_pipeline_model_vpc_config(expand_container_def, sagemaker_session): |
| model = sagemaker_session.create_model(MODEL_NAME, ROLE, CONTAINERS, VPC_CONFIG) |
|
|
| assert model == MODEL_NAME |
| sagemaker_session.sagemaker_client.create_model.assert_called_with( |
| ExecutionRoleArn=EXPANDED_ROLE, |
| ModelName=MODEL_NAME, |
| Containers=CONTAINERS, |
| VpcConfig=VPC_CONFIG, |
| ) |
|
|
|
|
| @patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER) |
| def test_create_model_already_exists(expand_container_def, sagemaker_session, caplog): |
| error_response = { |
| "Error": {"Code": "ValidationException", "Message": "Cannot create already existing model"} |
| } |
| exception = ClientError(error_response, "Operation") |
| sagemaker_session.sagemaker_client.create_model.side_effect = exception |
|
|
| model = sagemaker_session.create_model(MODEL_NAME, ROLE, PRIMARY_CONTAINER) |
| assert model == MODEL_NAME |
|
|
| expected_warning = ( |
| "sagemaker", |
| logging.WARNING, |
| "Using already existing model: {}".format(MODEL_NAME), |
| ) |
| assert expected_warning in caplog.record_tuples |
|
|
|
|
| @patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER) |
| def test_create_model_failure(expand_container_def, sagemaker_session): |
| error_message = "this is expected" |
| sagemaker_session.sagemaker_client.create_model.side_effect = RuntimeError(error_message) |
|
|
| with pytest.raises(RuntimeError) as e: |
| sagemaker_session.create_model(MODEL_NAME, ROLE, PRIMARY_CONTAINER) |
|
|
| assert error_message in str(e) |
|
|
|
|
| def test_create_model_from_job(sagemaker_session): |
| ims = sagemaker_session |
| ims.sagemaker_client.describe_training_job.return_value = COMPLETED_DESCRIBE_JOB_RESULT |
| ims.create_model_from_job(JOB_NAME) |
|
|
| assert ( |
| call(TrainingJobName=JOB_NAME) in ims.sagemaker_client.describe_training_job.call_args_list |
| ) |
| ims.sagemaker_client.create_model.assert_called_with( |
| ExecutionRoleArn=EXPANDED_ROLE, |
| ModelName=JOB_NAME, |
| PrimaryContainer=PRIMARY_CONTAINER, |
| VpcConfig=VPC_CONFIG, |
| ) |
|
|
|
|
| def test_create_model_from_job_with_tags(sagemaker_session): |
| ims = sagemaker_session |
| ims.sagemaker_client.describe_training_job.return_value = COMPLETED_DESCRIBE_JOB_RESULT |
| ims.create_model_from_job(JOB_NAME, tags=TAGS) |
|
|
| assert ( |
| call(TrainingJobName=JOB_NAME) in ims.sagemaker_client.describe_training_job.call_args_list |
| ) |
| ims.sagemaker_client.create_model.assert_called_with( |
| ExecutionRoleArn=EXPANDED_ROLE, |
| ModelName=JOB_NAME, |
| PrimaryContainer=PRIMARY_CONTAINER, |
| VpcConfig=VPC_CONFIG, |
| Tags=TAGS, |
| ) |
|
|
|
|
| def test_create_model_from_job_with_image(sagemaker_session): |
| ims = sagemaker_session |
| ims.sagemaker_client.describe_training_job.return_value = COMPLETED_DESCRIBE_JOB_RESULT |
| ims.create_model_from_job(JOB_NAME, image_uri="some-image") |
| [create_model_call] = ims.sagemaker_client.create_model.call_args_list |
| assert dict(create_model_call[1]["PrimaryContainer"])["Image"] == "some-image" |
|
|
|
|
| def test_create_model_from_job_with_container_def(sagemaker_session): |
| ims = sagemaker_session |
| ims.sagemaker_client.describe_training_job.return_value = COMPLETED_DESCRIBE_JOB_RESULT |
| ims.create_model_from_job( |
| JOB_NAME, |
| image_uri="some-image", |
| model_data_url="some-data", |
| env={"a": "b"}, |
| ) |
| [create_model_call] = ims.sagemaker_client.create_model.call_args_list |
| c_def = create_model_call[1]["PrimaryContainer"] |
| assert c_def["Image"] == "some-image" |
| assert c_def["ModelDataUrl"] == "some-data" |
| assert c_def["Environment"] == {"a": "b"} |
|
|
|
|
| def test_create_model_from_job_with_vpc_config_override(sagemaker_session): |
| vpc_config_override = {"Subnets": ["foo", "bar"], "SecurityGroupIds": ["baz"]} |
|
|
| ims = sagemaker_session |
| ims.sagemaker_client.describe_training_job.return_value = COMPLETED_DESCRIBE_JOB_RESULT |
| ims.create_model_from_job(JOB_NAME, vpc_config_override=vpc_config_override) |
| assert ims.sagemaker_client.create_model.call_args[1]["VpcConfig"] == vpc_config_override |
|
|
| ims.create_model_from_job(JOB_NAME, vpc_config_override=None) |
| assert "VpcConfig" not in ims.sagemaker_client.create_model.call_args[1] |
|
|
|
|
| def test_endpoint_from_production_variants(sagemaker_session): |
| ims = sagemaker_session |
| ims.sagemaker_client.describe_endpoint = Mock(return_value={"EndpointStatus": "InService"}) |
| pvs = [ |
| sagemaker.production_variant("A", "ml.p2.xlarge"), |
| sagemaker.production_variant("B", "p299.4096xlarge"), |
| ] |
| ex = ClientError( |
| {"Error": {"Code": "ValidationException", "Message": "Could not find your thing"}}, "b" |
| ) |
| ims.sagemaker_client.describe_endpoint_config = Mock(side_effect=ex) |
| sagemaker_session.endpoint_from_production_variants("some-endpoint", pvs) |
| sagemaker_session.sagemaker_client.create_endpoint.assert_called_with( |
| EndpointConfigName="some-endpoint", EndpointName="some-endpoint", Tags=[] |
| ) |
| sagemaker_session.sagemaker_client.create_endpoint_config.assert_called_with( |
| EndpointConfigName="some-endpoint", ProductionVariants=pvs |
| ) |
|
|
|
|
| def test_create_endpoint_config_with_tags(sagemaker_session): |
| tags = [{"Key": "TagtestKey", "Value": "TagtestValue"}] |
|
|
| sagemaker_session.create_endpoint_config("endpoint-test", "simple-model", 1, "local", tags=tags) |
|
|
| sagemaker_session.sagemaker_client.create_endpoint_config.assert_called_with( |
| EndpointConfigName="endpoint-test", ProductionVariants=ANY, Tags=tags |
| ) |
|
|
|
|
| def test_endpoint_from_production_variants_with_tags(sagemaker_session): |
| ims = sagemaker_session |
| ims.sagemaker_client.describe_endpoint = Mock(return_value={"EndpointStatus": "InService"}) |
| pvs = [ |
| sagemaker.production_variant("A", "ml.p2.xlarge"), |
| sagemaker.production_variant("B", "p299.4096xlarge"), |
| ] |
| ex = ClientError( |
| {"Error": {"Code": "ValidationException", "Message": "Could not find your thing"}}, "b" |
| ) |
| ims.sagemaker_client.describe_endpoint_config = Mock(side_effect=ex) |
| tags = [{"ModelName": "TestModel"}] |
| sagemaker_session.endpoint_from_production_variants("some-endpoint", pvs, tags) |
| sagemaker_session.sagemaker_client.create_endpoint.assert_called_with( |
| EndpointConfigName="some-endpoint", EndpointName="some-endpoint", Tags=tags |
| ) |
| sagemaker_session.sagemaker_client.create_endpoint_config.assert_called_with( |
| EndpointConfigName="some-endpoint", ProductionVariants=pvs, Tags=tags |
| ) |
|
|
|
|
| def test_endpoint_from_production_variants_with_accelerator_type(sagemaker_session): |
| ims = sagemaker_session |
| ims.sagemaker_client.describe_endpoint = Mock(return_value={"EndpointStatus": "InService"}) |
| pvs = [ |
| sagemaker.production_variant("A", "ml.p2.xlarge", accelerator_type=ACCELERATOR_TYPE), |
| sagemaker.production_variant("B", "p299.4096xlarge", accelerator_type=ACCELERATOR_TYPE), |
| ] |
| ex = ClientError( |
| {"Error": {"Code": "ValidationException", "Message": "Could not find your thing"}}, "b" |
| ) |
| ims.sagemaker_client.describe_endpoint_config = Mock(side_effect=ex) |
| tags = [{"ModelName": "TestModel"}] |
| sagemaker_session.endpoint_from_production_variants("some-endpoint", pvs, tags) |
| sagemaker_session.sagemaker_client.create_endpoint.assert_called_with( |
| EndpointConfigName="some-endpoint", EndpointName="some-endpoint", Tags=tags |
| ) |
| sagemaker_session.sagemaker_client.create_endpoint_config.assert_called_with( |
| EndpointConfigName="some-endpoint", ProductionVariants=pvs, Tags=tags |
| ) |
|
|
|
|
| def test_endpoint_from_production_variants_with_serverless_inference_config(sagemaker_session): |
| ims = sagemaker_session |
| ims.sagemaker_client.describe_endpoint = Mock(return_value={"EndpointStatus": "InService"}) |
| pvs = [ |
| sagemaker.production_variant( |
| "A", "ml.p2.xlarge", serverless_inference_config=SERVERLESS_INFERENCE_CONFIG |
| ), |
| sagemaker.production_variant( |
| "B", "p299.4096xlarge", serverless_inference_config=SERVERLESS_INFERENCE_CONFIG |
| ), |
| ] |
| ex = ClientError( |
| {"Error": {"Code": "ValidationException", "Message": "Could not find your thing"}}, "b" |
| ) |
| ims.sagemaker_client.describe_endpoint_config = Mock(side_effect=ex) |
| tags = [{"ModelName": "TestModel"}] |
| sagemaker_session.endpoint_from_production_variants("some-endpoint", pvs, tags) |
| sagemaker_session.sagemaker_client.create_endpoint.assert_called_with( |
| EndpointConfigName="some-endpoint", EndpointName="some-endpoint", Tags=tags |
| ) |
| sagemaker_session.sagemaker_client.create_endpoint_config.assert_called_with( |
| EndpointConfigName="some-endpoint", ProductionVariants=pvs, Tags=tags |
| ) |
|
|
|
|
| def test_endpoint_from_production_variants_with_async_config(sagemaker_session): |
| ims = sagemaker_session |
| ims.sagemaker_client.describe_endpoint = Mock(return_value={"EndpointStatus": "InService"}) |
| pvs = [ |
| sagemaker.production_variant("A", "ml.p2.xlarge"), |
| sagemaker.production_variant("B", "p299.4096xlarge"), |
| ] |
| ex = ClientError( |
| {"Error": {"Code": "ValidationException", "Message": "Could not find your thing"}}, "b" |
| ) |
| ims.sagemaker_client.describe_endpoint_config = Mock(side_effect=ex) |
| sagemaker_session.endpoint_from_production_variants( |
| "some-endpoint", |
| pvs, |
| async_inference_config_dict=AsyncInferenceConfig, |
| ) |
| sagemaker_session.sagemaker_client.create_endpoint.assert_called_with( |
| EndpointConfigName="some-endpoint", EndpointName="some-endpoint", Tags=[] |
| ) |
| sagemaker_session.sagemaker_client.create_endpoint_config.assert_called_with( |
| EndpointConfigName="some-endpoint", |
| ProductionVariants=pvs, |
| AsyncInferenceConfig=AsyncInferenceConfig, |
| ) |
|
|
|
|
| def test_update_endpoint_succeed(sagemaker_session): |
| sagemaker_session.sagemaker_client.describe_endpoint = Mock( |
| return_value={"EndpointStatus": "InService"} |
| ) |
| endpoint_name = "some-endpoint" |
| endpoint_config = "some-endpoint-config" |
| returned_endpoint_name = sagemaker_session.update_endpoint(endpoint_name, endpoint_config) |
| assert returned_endpoint_name == endpoint_name |
|
|
|
|
| def test_update_endpoint_no_wait(sagemaker_session): |
| sagemaker_session.sagemaker_client.describe_endpoint = Mock( |
| return_value={"EndpointStatus": "Updating"} |
| ) |
| endpoint_name = "some-endpoint" |
| endpoint_config = "some-endpoint-config" |
| returned_endpoint_name = sagemaker_session.update_endpoint( |
| endpoint_name, endpoint_config, wait=False |
| ) |
| assert returned_endpoint_name == endpoint_name |
|
|
|
|
| def test_update_endpoint_non_existing_endpoint(sagemaker_session): |
| error = ClientError( |
| {"Error": {"Code": "ValidationException", "Message": "Could not find entity"}}, "foo" |
| ) |
| expected_error_message = ( |
| "Endpoint with name 'non-existing-endpoint' does not exist; " |
| "please use an existing endpoint name" |
| ) |
| sagemaker_session.sagemaker_client.describe_endpoint = Mock(side_effect=error) |
| with pytest.raises(ValueError, match=expected_error_message): |
| sagemaker_session.update_endpoint("non-existing-endpoint", "non-existing-config") |
|
|
|
|
| def test_create_endpoint_config_from_existing(sagemaker_session): |
| pvs = [sagemaker.production_variant("A", "ml.m4.xlarge")] |
| tags = [{"Key": "aws:cloudformation:stackname", "Value": "this-tag-should-be-ignored"}] |
| existing_endpoint_arn = "arn:aws:sagemaker:us-west-2:123412341234:endpoint-config/foo" |
| kms_key = "kms" |
| sagemaker_session.sagemaker_client.describe_endpoint_config.return_value = { |
| "Tags": tags, |
| "ProductionVariants": pvs, |
| "EndpointConfigArn": existing_endpoint_arn, |
| "KmsKeyId": kms_key, |
| } |
| sagemaker_session.sagemaker_client.list_tags.return_value = {"Tags": tags} |
|
|
| existing_endpoint_name = "foo" |
| new_endpoint_name = "new-foo" |
| sagemaker_session.create_endpoint_config_from_existing( |
| existing_endpoint_name, new_endpoint_name |
| ) |
|
|
| sagemaker_session.sagemaker_client.describe_endpoint_config.assert_called_with( |
| EndpointConfigName=existing_endpoint_name |
| ) |
| sagemaker_session.sagemaker_client.list_tags.assert_called_with( |
| ResourceArn=existing_endpoint_arn, MaxResults=50 |
| ) |
| sagemaker_session.sagemaker_client.create_endpoint_config.assert_called_with( |
| EndpointConfigName=new_endpoint_name, ProductionVariants=pvs, KmsKeyId=kms_key |
| ) |
|
|
|
|
| @patch("time.sleep") |
| def test_wait_for_tuning_job(sleep, sagemaker_session): |
| hyperparameter_tuning_job_desc = {"HyperParameterTuningJobStatus": "Completed"} |
| sagemaker_session.sagemaker_client.describe_hyper_parameter_tuning_job = Mock( |
| name="describe_hyper_parameter_tuning_job", return_value=hyperparameter_tuning_job_desc |
| ) |
|
|
| result = sagemaker_session.wait_for_tuning_job(JOB_NAME) |
| assert result["HyperParameterTuningJobStatus"] == "Completed" |
|
|
|
|
| def test_tune_job_status(sagemaker_session): |
| hyperparameter_tuning_job_desc = {"HyperParameterTuningJobStatus": "Completed"} |
| sagemaker_session.sagemaker_client.describe_hyper_parameter_tuning_job = Mock( |
| name="describe_hyper_parameter_tuning_job", return_value=hyperparameter_tuning_job_desc |
| ) |
|
|
| result = _tuning_job_status(sagemaker_session.sagemaker_client, JOB_NAME) |
|
|
| assert result["HyperParameterTuningJobStatus"] == "Completed" |
|
|
|
|
| def test_tune_job_status_none(sagemaker_session): |
| hyperparameter_tuning_job_desc = {"HyperParameterTuningJobStatus": "InProgress"} |
| sagemaker_session.sagemaker_client.describe_hyper_parameter_tuning_job = Mock( |
| name="describe_hyper_parameter_tuning_job", return_value=hyperparameter_tuning_job_desc |
| ) |
|
|
| result = _tuning_job_status(sagemaker_session.sagemaker_client, JOB_NAME) |
|
|
| assert result is None |
|
|
|
|
| @patch("time.sleep") |
| def test_wait_for_transform_job_completed(sleep, sagemaker_session): |
| transform_job_desc = {"TransformJobStatus": "Completed"} |
| sagemaker_session.sagemaker_client.describe_transform_job = Mock( |
| name="describe_transform_job", return_value=transform_job_desc |
| ) |
|
|
| assert sagemaker_session.wait_for_transform_job(JOB_NAME)["TransformJobStatus"] == "Completed" |
|
|
|
|
| @patch("time.sleep") |
| def test_wait_for_transform_job_in_progress(sleep, sagemaker_session): |
| transform_job_desc_in_progress = {"TransformJobStatus": "InProgress"} |
| transform_job_desc_in_completed = {"TransformJobStatus": "Completed"} |
| sagemaker_session.sagemaker_client.describe_transform_job = Mock( |
| name="describe_transform_job", |
| side_effect=[transform_job_desc_in_progress, transform_job_desc_in_completed], |
| ) |
|
|
| assert ( |
| sagemaker_session.wait_for_transform_job(JOB_NAME, 1)["TransformJobStatus"] == "Completed" |
| ) |
| assert 2 == sagemaker_session.sagemaker_client.describe_transform_job.call_count |
|
|
|
|
| def test_transform_job_status(sagemaker_session): |
| transform_job_desc = {"TransformJobStatus": "Completed"} |
| sagemaker_session.sagemaker_client.describe_transform_job = Mock( |
| name="describe_transform_job", return_value=transform_job_desc |
| ) |
|
|
| result = _transform_job_status(sagemaker_session.sagemaker_client, JOB_NAME) |
| assert result["TransformJobStatus"] == "Completed" |
|
|
|
|
| def test_transform_job_status_none(sagemaker_session): |
| transform_job_desc = {"TransformJobStatus": "InProgress"} |
| sagemaker_session.sagemaker_client.describe_transform_job = Mock( |
| name="describe_transform_job", return_value=transform_job_desc |
| ) |
|
|
| result = _transform_job_status(sagemaker_session.sagemaker_client, JOB_NAME) |
| assert result is None |
|
|
|
|
| def test_train_done_completed(sagemaker_session): |
| training_job_desc = {"TrainingJobStatus": "Completed"} |
| sagemaker_session.sagemaker_client.describe_training_job = Mock( |
| name="describe_training_job", return_value=training_job_desc |
| ) |
|
|
| actual_job_desc, training_finished = _train_done( |
| sagemaker_session.sagemaker_client, JOB_NAME, None |
| ) |
|
|
| assert actual_job_desc["TrainingJobStatus"] == "Completed" |
| assert training_finished is True |
|
|
|
|
| def test_train_done_in_progress(sagemaker_session): |
| training_job_desc = {"TrainingJobStatus": "InProgress"} |
| sagemaker_session.sagemaker_client.describe_training_job = Mock( |
| name="describe_training_job", return_value=training_job_desc |
| ) |
|
|
| actual_job_desc, training_finished = _train_done( |
| sagemaker_session.sagemaker_client, JOB_NAME, None |
| ) |
|
|
| assert actual_job_desc["TrainingJobStatus"] == "InProgress" |
| assert training_finished is False |
|
|
|
|
| DEFAULT_EXPECTED_AUTO_ML_JOB_ARGS = { |
| "AutoMLJobName": JOB_NAME, |
| "InputDataConfig": [ |
| { |
| "DataSource": {"S3DataSource": {"S3DataType": "S3Prefix", "S3Uri": S3_INPUT_URI}}, |
| "TargetAttributeName": "y", |
| } |
| ], |
| "OutputDataConfig": {"S3OutputPath": S3_OUTPUT}, |
| "AutoMLJobConfig": { |
| "CompletionCriteria": { |
| "MaxCandidates": 10, |
| "MaxAutoMLJobRuntimeInSeconds": 36000, |
| "MaxRuntimePerTrainingJobInSeconds": 3600 * 2, |
| } |
| }, |
| "RoleArn": EXPANDED_ROLE, |
| "GenerateCandidateDefinitionsOnly": False, |
| } |
|
|
|
|
| COMPLETE_EXPECTED_AUTO_ML_JOB_ARGS = { |
| "AutoMLJobName": JOB_NAME, |
| "InputDataConfig": [ |
| { |
| "DataSource": {"S3DataSource": {"S3DataType": "S3Prefix", "S3Uri": S3_INPUT_URI}}, |
| "CompressionType": "Gzip", |
| "TargetAttributeName": "y", |
| } |
| ], |
| "OutputDataConfig": {"S3OutputPath": S3_OUTPUT}, |
| "ProblemType": "Regression", |
| "AutoMLJobObjective": {"Type": "type", "MetricName": "metric-name"}, |
| "AutoMLJobConfig": { |
| "CompletionCriteria": { |
| "MaxCandidates": 10, |
| "MaxAutoMLJobRuntimeInSeconds": 36000, |
| "MaxRuntimePerTrainingJobInSeconds": 3600 * 2, |
| }, |
| "SecurityConfig": { |
| "VolumeKmsKeyId": "volume-kms-key-id-string", |
| "EnableInterContainerTrafficEncryption": False, |
| "VpcConfig": {"SecurityGroupIds": ["security-group-id"], "Subnets": ["subnet"]}, |
| }, |
| }, |
| "RoleArn": EXPANDED_ROLE, |
| "GenerateCandidateDefinitionsOnly": True, |
| "Tags": ["tag"], |
| } |
|
|
| COMPLETE_EXPECTED_LIST_CANDIDATES_ARGS = { |
| "AutoMLJobName": JOB_NAME, |
| "StatusEquals": "Completed", |
| "SortOrder": "Descending", |
| "SortBy": "Status", |
| "MaxResults": 10, |
| } |
|
|
|
|
| def test_auto_ml_pack_to_request(sagemaker_session): |
| input_config = [ |
| { |
| "DataSource": {"S3DataSource": {"S3DataType": "S3Prefix", "S3Uri": S3_INPUT_URI}}, |
| "TargetAttributeName": "y", |
| } |
| ] |
|
|
| output_config = {"S3OutputPath": S3_OUTPUT} |
|
|
| auto_ml_job_config = { |
| "CompletionCriteria": { |
| "MaxCandidates": 10, |
| "MaxAutoMLJobRuntimeInSeconds": 36000, |
| "MaxRuntimePerTrainingJobInSeconds": 3600 * 2, |
| } |
| } |
|
|
| job_name = JOB_NAME |
| role = EXPANDED_ROLE |
|
|
| sagemaker_session.auto_ml(input_config, output_config, auto_ml_job_config, role, job_name) |
|
|
| assert sagemaker_session.sagemaker_client.method_calls[0] == ( |
| "create_auto_ml_job", |
| (), |
| DEFAULT_EXPECTED_AUTO_ML_JOB_ARGS, |
| ) |
|
|
|
|
| def test_auto_ml_pack_to_request_with_optional_args(sagemaker_session): |
| input_config = [ |
| { |
| "DataSource": {"S3DataSource": {"S3DataType": "S3Prefix", "S3Uri": S3_INPUT_URI}}, |
| "CompressionType": "Gzip", |
| "TargetAttributeName": "y", |
| } |
| ] |
|
|
| output_config = {"S3OutputPath": S3_OUTPUT} |
|
|
| auto_ml_job_config = { |
| "CompletionCriteria": { |
| "MaxCandidates": 10, |
| "MaxAutoMLJobRuntimeInSeconds": 36000, |
| "MaxRuntimePerTrainingJobInSeconds": 3600 * 2, |
| }, |
| "SecurityConfig": { |
| "VolumeKmsKeyId": "volume-kms-key-id-string", |
| "EnableInterContainerTrafficEncryption": False, |
| "VpcConfig": {"SecurityGroupIds": ["security-group-id"], "Subnets": ["subnet"]}, |
| }, |
| } |
|
|
| job_name = JOB_NAME |
| role = EXPANDED_ROLE |
|
|
| sagemaker_session.auto_ml( |
| input_config, |
| output_config, |
| auto_ml_job_config, |
| role, |
| job_name, |
| problem_type="Regression", |
| job_objective={"Type": "type", "MetricName": "metric-name"}, |
| generate_candidate_definitions_only=True, |
| tags=["tag"], |
| ) |
|
|
| assert sagemaker_session.sagemaker_client.method_calls[0] == ( |
| "create_auto_ml_job", |
| (), |
| COMPLETE_EXPECTED_AUTO_ML_JOB_ARGS, |
| ) |
|
|
|
|
| def test_list_candidates_for_auto_ml_job_default(sagemaker_session): |
| sagemaker_session.list_candidates(job_name=JOB_NAME) |
| sagemaker_session.sagemaker_client.list_candidates_for_auto_ml_job.assert_called_once() |
| sagemaker_session.sagemaker_client.list_candidates_for_auto_ml_job.assert_called_with( |
| AutoMLJobName=JOB_NAME |
| ) |
|
|
|
|
| def test_list_candidates_for_auto_ml_job_with_optional_args(sagemaker_session): |
| sagemaker_session.list_candidates( |
| job_name=JOB_NAME, |
| status_equals="Completed", |
| sort_order="Descending", |
| sort_by="Status", |
| max_results=10, |
| ) |
| sagemaker_session.sagemaker_client.list_candidates_for_auto_ml_job.assert_called_once() |
| sagemaker_session.sagemaker_client.list_candidates_for_auto_ml_job.assert_called_with( |
| **COMPLETE_EXPECTED_LIST_CANDIDATES_ARGS |
| ) |
|
|
|
|
| def test_describe_tuning_job(sagemaker_session): |
| job_name = "hyper-parameter-tuning" |
| sagemaker_session.describe_tuning_job(job_name=job_name) |
| sagemaker_session.sagemaker_client.describe_hyper_parameter_tuning_job.assert_called_with( |
| HyperParameterTuningJobName=job_name |
| ) |
|
|
|
|
| @pytest.fixture |
| def pipeline_empty_definition(): |
| return '{"Version": "2020-12-01", ' '"Metadata": {}, ' '"Parameters": [], ' '"Steps": []}' |
|
|
|
|
| @pytest.fixture |
| def pipeline_role_arn(): |
| return "my:pipeline:role:arn" |
|
|
|
|
| def test_describe_model(sagemaker_session): |
| model_name = "sagemaker-model-name" |
| sagemaker_session.describe_model(name=model_name) |
| sagemaker_session.sagemaker_client.describe_model.assert_called_with(ModelName=model_name) |
|
|
|
|
| def test_create_model_package_from_containers(sagemaker_session): |
| model_package_name = "sagemaker-model-package" |
| sagemaker_session.create_model_package_from_containers(model_package_name=model_package_name) |
| sagemaker_session.sagemaker_client.create_model_package.assert_called_once() |
|
|
|
|
| def test_create_model_package_from_containers_name_conflict(sagemaker_session): |
| model_package_name = "sagemaker-model-package" |
| model_package_group_name = "sagemaker-model-package-group" |
| with pytest.raises(ValueError) as error: |
| sagemaker_session.create_model_package_from_containers( |
| model_package_name=model_package_name, |
| model_package_group_name=model_package_group_name, |
| ) |
| assert ( |
| "model_package_name and model_package_group_name cannot be present at the same " |
| "time." == str(error) |
| ) |
|
|
|
|
| def test_create_model_package_from_containers_incomplete_args(sagemaker_session): |
| model_package_name = "sagemaker-model-package" |
| containers = ["dummy-container"] |
| with pytest.raises(ValueError) as error: |
| sagemaker_session.create_model_package_from_containers( |
| model_package_name=model_package_name, |
| containers=containers, |
| ) |
| assert ( |
| "content_types and response_types " |
| "must be provided if containers is present." == str(error) |
| ) |
|
|
|
|
| def test_create_model_package_from_containers_without_model_package_group_name(sagemaker_session): |
| model_package_name = "sagemaker-model-package" |
| containers = ["dummy-container"] |
| content_types = ["application/json"] |
| response_types = ["application/json"] |
| with pytest.raises(ValueError) as error: |
| sagemaker_session.create_model_package_from_containers( |
| model_package_name=model_package_name, |
| containers=containers, |
| content_types=content_types, |
| response_types=response_types, |
| ) |
| assert ( |
| "inference_inferences and transform_instances " |
| "must be provided if model_package_group_name is not present." == str(error) |
| ) |
|
|
|
|
| def test_create_model_package_from_containers_all_args(sagemaker_session): |
| model_package_name = "sagemaker-model-package" |
| containers = ["dummy-container"] |
| content_types = ["application/json"] |
| response_types = ["application/json"] |
| inference_instances = ["ml.m4.xlarge"] |
| transform_instances = ["ml.m4.xlarget"] |
| model_metrics = { |
| "Bias": { |
| "ContentType": "content-type", |
| "S3Uri": "s3://...", |
| } |
| } |
| drift_check_baselines = { |
| "Bias": { |
| "ConfigFile": { |
| "ContentType": "content-type", |
| "S3Uri": "s3://...", |
| } |
| } |
| } |
|
|
| metadata_properties = { |
| "CommitId": "test-commit-id", |
| "Repository": "test-repository", |
| "GeneratedBy": "sagemaker-python-sdk", |
| "ProjectId": "unit-test", |
| } |
| marketplace_cert = (True,) |
| approval_status = ("Approved",) |
| description = "description" |
| customer_metadata_properties = {"key1": "value1"} |
| domain = "COMPUTER_VISION" |
| task = "IMAGE_CLASSIFICATION" |
| sample_payload_url = "s3://test-bucket/model" |
| sagemaker_session.create_model_package_from_containers( |
| containers=containers, |
| content_types=content_types, |
| response_types=response_types, |
| inference_instances=inference_instances, |
| transform_instances=transform_instances, |
| model_package_name=model_package_name, |
| model_metrics=model_metrics, |
| metadata_properties=metadata_properties, |
| marketplace_cert=marketplace_cert, |
| approval_status=approval_status, |
| description=description, |
| drift_check_baselines=drift_check_baselines, |
| customer_metadata_properties=customer_metadata_properties, |
| domain=domain, |
| sample_payload_url=sample_payload_url, |
| task=task, |
| ) |
| expected_args = { |
| "ModelPackageName": model_package_name, |
| "InferenceSpecification": { |
| "Containers": containers, |
| "SupportedContentTypes": content_types, |
| "SupportedResponseMIMETypes": response_types, |
| "SupportedRealtimeInferenceInstanceTypes": inference_instances, |
| "SupportedTransformInstanceTypes": transform_instances, |
| }, |
| "ModelPackageDescription": description, |
| "ModelMetrics": model_metrics, |
| "MetadataProperties": metadata_properties, |
| "CertifyForMarketplace": marketplace_cert, |
| "ModelApprovalStatus": approval_status, |
| "DriftCheckBaselines": drift_check_baselines, |
| "CustomerMetadataProperties": customer_metadata_properties, |
| "Domain": domain, |
| "SamplePayloadUrl": sample_payload_url, |
| "Task": task, |
| } |
| sagemaker_session.sagemaker_client.create_model_package.assert_called_with(**expected_args) |
|
|
|
|
| def test_create_model_package_from_containers_without_instance_types(sagemaker_session): |
| model_package_group_name = "sagemaker-model-package-group-name-1.0" |
| containers = ["dummy-container"] |
| content_types = ["application/json"] |
| response_types = ["application/json"] |
| model_metrics = { |
| "Bias": { |
| "ContentType": "content-type", |
| "S3Uri": "s3://...", |
| } |
| } |
| drift_check_baselines = { |
| "Bias": { |
| "ConfigFile": { |
| "ContentType": "content-type", |
| "S3Uri": "s3://...", |
| } |
| } |
| } |
|
|
| metadata_properties = { |
| "CommitId": "test-commit-id", |
| "Repository": "test-repository", |
| "GeneratedBy": "sagemaker-python-sdk", |
| "ProjectId": "unit-test", |
| } |
| marketplace_cert = (True,) |
| approval_status = ("Approved",) |
| description = "description" |
| customer_metadata_properties = {"key1": "value1"} |
| sagemaker_session.create_model_package_from_containers( |
| containers=containers, |
| content_types=content_types, |
| response_types=response_types, |
| model_package_group_name=model_package_group_name, |
| model_metrics=model_metrics, |
| metadata_properties=metadata_properties, |
| marketplace_cert=marketplace_cert, |
| approval_status=approval_status, |
| description=description, |
| drift_check_baselines=drift_check_baselines, |
| customer_metadata_properties=customer_metadata_properties, |
| ) |
| expected_args = { |
| "ModelPackageGroupName": model_package_group_name, |
| "InferenceSpecification": { |
| "Containers": containers, |
| "SupportedContentTypes": content_types, |
| "SupportedResponseMIMETypes": response_types, |
| }, |
| "ModelPackageDescription": description, |
| "ModelMetrics": model_metrics, |
| "MetadataProperties": metadata_properties, |
| "CertifyForMarketplace": marketplace_cert, |
| "ModelApprovalStatus": approval_status, |
| "DriftCheckBaselines": drift_check_baselines, |
| "CustomerMetadataProperties": customer_metadata_properties, |
| } |
| sagemaker_session.sagemaker_client.create_model_package.assert_called_with(**expected_args) |
|
|
|
|
| def test_create_model_package_from_containers_with_one_instance_types(sagemaker_session): |
| model_package_group_name = "sagemaker-model-package-group-name-1.0" |
| containers = ["dummy-container"] |
| content_types = ["application/json"] |
| response_types = ["application/json"] |
| transform_instances = ["ml.m5.xlarge"] |
| model_metrics = { |
| "Bias": { |
| "ContentType": "content-type", |
| "S3Uri": "s3://...", |
| } |
| } |
| drift_check_baselines = { |
| "Bias": { |
| "ConfigFile": { |
| "ContentType": "content-type", |
| "S3Uri": "s3://...", |
| } |
| } |
| } |
|
|
| metadata_properties = { |
| "CommitId": "test-commit-id", |
| "Repository": "test-repository", |
| "GeneratedBy": "sagemaker-python-sdk", |
| "ProjectId": "unit-test", |
| } |
| marketplace_cert = (True,) |
| approval_status = ("Approved",) |
| description = "description" |
| customer_metadata_properties = {"key1": "value1"} |
| sagemaker_session.create_model_package_from_containers( |
| containers=containers, |
| content_types=content_types, |
| response_types=response_types, |
| transform_instances=transform_instances, |
| model_package_group_name=model_package_group_name, |
| model_metrics=model_metrics, |
| metadata_properties=metadata_properties, |
| marketplace_cert=marketplace_cert, |
| approval_status=approval_status, |
| description=description, |
| drift_check_baselines=drift_check_baselines, |
| customer_metadata_properties=customer_metadata_properties, |
| ) |
| expected_args = { |
| "ModelPackageGroupName": model_package_group_name, |
| "InferenceSpecification": { |
| "Containers": containers, |
| "SupportedContentTypes": content_types, |
| "SupportedResponseMIMETypes": response_types, |
| "SupportedTransformInstanceTypes": transform_instances, |
| }, |
| "ModelPackageDescription": description, |
| "ModelMetrics": model_metrics, |
| "MetadataProperties": metadata_properties, |
| "CertifyForMarketplace": marketplace_cert, |
| "ModelApprovalStatus": approval_status, |
| "DriftCheckBaselines": drift_check_baselines, |
| "CustomerMetadataProperties": customer_metadata_properties, |
| } |
| sagemaker_session.sagemaker_client.create_model_package.assert_called_with(**expected_args) |
|
|
|
|
| @pytest.fixture |
| def feature_group_dummy_definitions(): |
| return [{"FeatureName": "feature1", "FeatureType": "String"}] |
|
|
|
|
| def test_feature_group_create(sagemaker_session, feature_group_dummy_definitions): |
| sagemaker_session.create_feature_group( |
| feature_group_name="MyFeatureGroup", |
| record_identifier_name="feature1", |
| event_time_feature_name="feature2", |
| feature_definitions=feature_group_dummy_definitions, |
| role_arn="dummy_role", |
| ) |
| assert sagemaker_session.sagemaker_client.create_feature_group.called_with( |
| FeatureGroupName="MyFeatureGroup", |
| RecordIdentifierFeatureName="feature1", |
| EventTimeFeatureName="feature2", |
| FeatureDefinitions=feature_group_dummy_definitions, |
| RoleArn="dummy_role", |
| ) |
|
|
|
|
| def test_feature_group_delete(sagemaker_session): |
| sagemaker_session.delete_feature_group(feature_group_name="MyFeatureGroup") |
| assert sagemaker_session.sagemaker_client.delete_feature_group.called_with( |
| FeatureGroupName="MyFeatureGroup", |
| ) |
|
|
|
|
| def test_feature_group_describe(sagemaker_session): |
| sagemaker_session.describe_feature_group(feature_group_name="MyFeatureGroup") |
| assert sagemaker_session.sagemaker_client.describe_feature_group.called_with( |
| FeatureGroupName="MyFeatureGroup", |
| ) |
|
|
|
|
| def test_feature_group_update(sagemaker_session, feature_group_dummy_definitions): |
| sagemaker_session.update_feature_group( |
| feature_group_name="MyFeatureGroup", |
| feature_additions=feature_group_dummy_definitions, |
| ) |
| assert sagemaker_session.sagemaker_client.update_feature_group.called_with( |
| FeatureGroupName="MyFeatureGroup", |
| FeatureAdditions=feature_group_dummy_definitions, |
| ) |
|
|
|
|
| def test_feature_metadata_update(sagemaker_session): |
| parameter_additions = [ |
| { |
| "key": "TestKey", |
| "value": "TestValue", |
| } |
| ] |
| parameter_removals = ["TestKey"] |
|
|
| sagemaker_session.update_feature_metadata( |
| feature_group_name="TestFeatureGroup", |
| feature_name="TestFeature", |
| description="TestDescription", |
| parameter_additions=parameter_additions, |
| parameter_removals=parameter_removals, |
| ) |
| assert sagemaker_session.sagemaker_client.update_feature_group.called_with( |
| feature_group_name="TestFeatureGroup", |
| FeatureName="TestFeature", |
| Description="TestDescription", |
| ParameterAdditions=parameter_additions, |
| ParameterRemovals=parameter_removals, |
| ) |
| sagemaker_session.update_feature_metadata( |
| feature_group_name="TestFeatureGroup", |
| feature_name="TestFeature", |
| ) |
| assert sagemaker_session.sagemaker_client.update_feature_group.called_with( |
| feature_group_name="TestFeatureGroup", |
| FeatureName="TestFeature", |
| ) |
|
|
|
|
| def test_feature_metadata_describe(sagemaker_session): |
| sagemaker_session.describe_feature_metadata( |
| feature_group_name="MyFeatureGroup", feature_name="TestFeature" |
| ) |
| assert sagemaker_session.sagemaker_client.describe_feature_metadata.called_with( |
| FeatureGroupName="MyFeatureGroup", FeatureName="TestFeature" |
| ) |
|
|
|
|
| def test_start_query_execution(sagemaker_session): |
| athena_mock = Mock() |
| sagemaker_session.boto_session.client( |
| "athena", region_name=sagemaker_session.boto_region_name |
| ).return_value = athena_mock |
| sagemaker_session.start_query_execution( |
| catalog="catalog", |
| database="database", |
| query_string="query", |
| output_location="s3://results", |
| ) |
| assert athena_mock.start_query_execution.called_once_with( |
| QueryString="query", |
| QueryExecutionContext={"Catalog": "catalog", "Database": "database"}, |
| OutputLocation="s3://results", |
| ) |
|
|
|
|
| def test_get_query_execution(sagemaker_session): |
| athena_mock = Mock() |
| sagemaker_session.boto_session.client( |
| "athena", region_name=sagemaker_session.boto_region_name |
| ).return_value = athena_mock |
| sagemaker_session.get_query_execution(query_execution_id="query_id") |
| assert athena_mock.get_query_execution.called_with(QueryExecutionId="query_id") |
|
|
|
|
| def test_download_athena_query_result(sagemaker_session): |
| sagemaker_session.s3_client = Mock() |
| sagemaker_session.download_athena_query_result( |
| bucket="bucket", |
| prefix="prefix", |
| query_execution_id="query_id", |
| filename="filename", |
| ) |
| assert sagemaker_session.s3_client.download_file.called_with( |
| Bucket="bucket", |
| Key="prefix/query_id.csv", |
| Filename="filename", |
| ) |
|
|
|
|
| @patch("sagemaker.session.Session.get_query_execution") |
| def test_wait_for_athena_query(query_execution, sagemaker_session): |
| query_execution.return_value = {"QueryExecution": {"Status": {"State": "SUCCEEDED"}}} |
| sagemaker_session.wait_for_athena_query(query_execution_id="query_id") |
| assert query_execution.called_with(query_execution_id="query_id") |
|
|