| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| from __future__ import absolute_import |
|
|
| import random |
| import string |
|
|
| from botocore.credentials import Credentials |
|
|
| import base64 |
| import logging |
| import json |
| import os |
| import subprocess |
| import tarfile |
|
|
| import pytest |
| import yaml |
| from mock import patch, Mock, MagicMock |
|
|
| import sagemaker |
| from sagemaker.local.image import _SageMakerContainer, _aws_credentials |
|
|
| REGION = "us-west-2" |
| BUCKET_NAME = "mybucket" |
| EXPANDED_ROLE = "arn:aws:iam::111111111111:role/ExpandedRole" |
| TRAINING_JOB_NAME = "my-job" |
| INPUT_DATA_CONFIG = [ |
| { |
| "ChannelName": "a", |
| "DataUri": "file:///tmp/source1", |
| "DataSource": { |
| "FileDataSource": { |
| "FileDataDistributionType": "FullyReplicated", |
| "FileUri": "file:///tmp/source1", |
| } |
| }, |
| }, |
| { |
| "ChannelName": "b", |
| "DataUri": "s3://my-own-bucket/prefix", |
| "DataSource": { |
| "S3DataSource": { |
| "S3DataDistributionType": "FullyReplicated", |
| "S3DataType": "S3Prefix", |
| "S3Uri": "s3://my-own-bucket/prefix", |
| } |
| }, |
| }, |
| ] |
|
|
| OUTPUT_DATA_CONFIG = {"S3OutputPath": ""} |
|
|
| HYPERPARAMETERS = { |
| "a": 1, |
| "b": json.dumps("bee"), |
| "sagemaker_submit_directory": json.dumps("s3://my_bucket/code"), |
| } |
|
|
| LOCAL_CODE_HYPERPARAMETERS = { |
| "a": 1, |
| "b": 2, |
| "sagemaker_submit_directory": json.dumps("file:///tmp/code"), |
| } |
|
|
| ENVIRONMENT = {"MYVAR": "HELLO_WORLD"} |
|
|
|
|
| @pytest.fixture() |
| def sagemaker_session(): |
| boto_mock = MagicMock(name="boto_session", region_name=REGION) |
| boto_mock.client("sts").get_caller_identity.return_value = {"Account": "123"} |
| boto_mock.resource("s3").Bucket(BUCKET_NAME).objects.filter.return_value = [] |
|
|
| sms = sagemaker.Session(boto_session=boto_mock, sagemaker_client=MagicMock()) |
|
|
| sms.default_bucket = Mock(name="default_bucket", return_value=BUCKET_NAME) |
| sms.expand_role = Mock(return_value=EXPANDED_ROLE) |
|
|
| return sms |
|
|
|
|
| def test_sagemaker_container_hosts_should_have_lowercase_names(): |
| random.seed(a=42) |
|
|
| def assert_all_lowercase(hosts): |
| for host in hosts: |
| assert host.lower() == host |
|
|
| sagemaker_container = _SageMakerContainer("local", 2, "my-image", sagemaker_session=Mock()) |
| assert_all_lowercase(sagemaker_container.hosts) |
|
|
| sagemaker_container = _SageMakerContainer("local", 10, "my-image", sagemaker_session=Mock()) |
| assert_all_lowercase(sagemaker_container.hosts) |
|
|
| sagemaker_container = _SageMakerContainer("local", 1, "my-image", sagemaker_session=Mock()) |
| assert_all_lowercase(sagemaker_container.hosts) |
|
|
|
|
| @patch("sagemaker.local.local_session.LocalSession") |
| def test_write_config_file(LocalSession, tmpdir): |
| sagemaker_container = _SageMakerContainer("local", 2, "my-image") |
| sagemaker_container.container_root = str(tmpdir.mkdir("container-root")) |
| host = "algo-1" |
|
|
| sagemaker.local.image._create_config_file_directories(sagemaker_container.container_root, host) |
|
|
| container_root = sagemaker_container.container_root |
| config_file_root = os.path.join(container_root, host, "input", "config") |
|
|
| hyperparameters_file = os.path.join(config_file_root, "hyperparameters.json") |
| resource_config_file = os.path.join(config_file_root, "resourceconfig.json") |
| input_data_config_file = os.path.join(config_file_root, "inputdataconfig.json") |
|
|
| |
| sagemaker_container.write_config_files(host, HYPERPARAMETERS, INPUT_DATA_CONFIG) |
|
|
| assert os.path.exists(hyperparameters_file) |
| assert os.path.exists(resource_config_file) |
| assert os.path.exists(input_data_config_file) |
|
|
| hyperparameters_data = json.load(open(hyperparameters_file)) |
| resource_config_data = json.load(open(resource_config_file)) |
| input_data_config_data = json.load(open(input_data_config_file)) |
|
|
| |
| for k, v in HYPERPARAMETERS.items(): |
| assert k in hyperparameters_data |
| assert hyperparameters_data[k] == v |
|
|
| |
| assert resource_config_data["current_host"] == host |
| assert resource_config_data["hosts"] == sagemaker_container.hosts |
|
|
| |
| for channel in INPUT_DATA_CONFIG: |
| assert channel["ChannelName"] in input_data_config_data |
|
|
|
|
| @patch("sagemaker.local.local_session.LocalSession") |
| def test_write_config_files_input_content_type(LocalSession, tmpdir): |
| sagemaker_container = _SageMakerContainer("local", 1, "my-image") |
| sagemaker_container.container_root = str(tmpdir.mkdir("container-root")) |
| host = "algo-1" |
|
|
| sagemaker.local.image._create_config_file_directories(sagemaker_container.container_root, host) |
|
|
| container_root = sagemaker_container.container_root |
| config_file_root = os.path.join(container_root, host, "input", "config") |
|
|
| input_data_config_file = os.path.join(config_file_root, "inputdataconfig.json") |
|
|
| |
| input_data_config = [ |
| { |
| "ChannelName": "channel_a", |
| "DataUri": "file:///tmp/source1", |
| "ContentType": "text/csv", |
| "DataSource": { |
| "FileDataSource": { |
| "FileDataDistributionType": "FullyReplicated", |
| "FileUri": "file:///tmp/source1", |
| } |
| }, |
| }, |
| { |
| "ChannelName": "channel_b", |
| "DataUri": "s3://my-own-bucket/prefix", |
| "DataSource": { |
| "S3DataSource": { |
| "S3DataDistributionType": "FullyReplicated", |
| "S3DataType": "S3Prefix", |
| "S3Uri": "s3://my-own-bucket/prefix", |
| } |
| }, |
| }, |
| ] |
| sagemaker_container.write_config_files(host, HYPERPARAMETERS, input_data_config) |
|
|
| assert os.path.exists(input_data_config_file) |
| parsed_input_config = json.load(open(input_data_config_file)) |
| |
| for channel in input_data_config: |
| assert channel["ChannelName"] in parsed_input_config |
|
|
| |
| assert "ContentType" in parsed_input_config["channel_a"] |
| assert parsed_input_config["channel_a"]["ContentType"] == "text/csv" |
|
|
| |
| assert "ContentType" not in parsed_input_config["channel_b"] |
|
|
|
|
| @patch("sagemaker.local.local_session.LocalSession") |
| def test_retrieve_artifacts(LocalSession, tmpdir): |
| sagemaker_container = _SageMakerContainer("local", 2, "my-image") |
| sagemaker_container.hosts = ["algo-1", "algo-2"] |
| sagemaker_container.container_root = str(tmpdir.mkdir("container-root")) |
|
|
| volume1 = os.path.join(sagemaker_container.container_root, "algo-1") |
| volume2 = os.path.join(sagemaker_container.container_root, "algo-2") |
| os.mkdir(volume1) |
| os.mkdir(volume2) |
|
|
| compose_data = { |
| "services": { |
| "algo-1": { |
| "volumes": [ |
| "%s:/opt/ml/model" % os.path.join(volume1, "model"), |
| "%s:/opt/ml/output" % os.path.join(volume1, "output"), |
| ] |
| }, |
| "algo-2": { |
| "volumes": [ |
| "%s:/opt/ml/model" % os.path.join(volume2, "model"), |
| "%s:/opt/ml/output" % os.path.join(volume2, "output"), |
| ] |
| }, |
| } |
| } |
|
|
| dirs = [ |
| ("model", volume1), |
| ("model/data", volume1), |
| ("model", volume2), |
| ("model/data", volume2), |
| ("model/tmp", volume2), |
| ("output", volume1), |
| ("output/data", volume1), |
| ("output", volume2), |
| ("output/data", volume2), |
| ("output/log", volume2), |
| ] |
|
|
| files = [ |
| ("model/data/model.json", volume1), |
| ("model/data/variables.csv", volume1), |
| ("model/data/model.json", volume2), |
| ("model/data/variables2.csv", volume2), |
| ("model/tmp/something-else.json", volume2), |
| ("output/data/loss.json", volume1), |
| ("output/data/accuracy.json", volume1), |
| ("output/data/loss.json", volume2), |
| ("output/data/accuracy2.json", volume2), |
| ("output/log/warnings.txt", volume2), |
| ] |
|
|
| expected_model = [ |
| "data", |
| "data/model.json", |
| "data/variables.csv", |
| "data/variables2.csv", |
| "tmp/something-else.json", |
| ] |
| expected_output = [ |
| "data", |
| "log", |
| "data/loss.json", |
| "data/accuracy.json", |
| "data/accuracy2.json", |
| "log/warnings.txt", |
| ] |
|
|
| for d, volume in dirs: |
| os.mkdir(os.path.join(volume, d)) |
|
|
| |
| for f, volume in files: |
| open(os.path.join(volume, f), "a").close() |
|
|
| output_path = str(tmpdir.mkdir("exported_files")) |
| output_data_config = {"S3OutputPath": "file://%s" % output_path} |
|
|
| model_artifacts = sagemaker_container.retrieve_artifacts( |
| compose_data, output_data_config, sagemaker_session |
| ).replace("file://", "") |
| artifacts = os.path.dirname(model_artifacts) |
|
|
| |
| assert set(os.listdir(artifacts)) == {"model.tar.gz", "output.tar.gz"} |
|
|
| |
| tar = tarfile.open(os.path.join(output_path, "model.tar.gz")) |
| model_tar_files = [m.name for m in tar.getmembers()] |
| for f in expected_model: |
| assert f in model_tar_files |
|
|
| tar = tarfile.open(os.path.join(output_path, "output.tar.gz")) |
| output_tar_files = [m.name for m in tar.getmembers()] |
| for f in expected_output: |
| assert f in output_tar_files |
|
|
|
|
| def test_stream_output(): |
| |
| with pytest.raises(RuntimeError): |
| p = subprocess.Popen( |
| ["ls", "/some/unknown/path"], stdout=subprocess.PIPE, stderr=subprocess.PIPE |
| ) |
| sagemaker.local.image._stream_output(p) |
|
|
| p = subprocess.Popen(["echo", "hello"], stdout=subprocess.PIPE, stderr=subprocess.PIPE) |
| exit_code = sagemaker.local.image._stream_output(p) |
| assert exit_code == 0 |
|
|
|
|
| def test_check_output(): |
| with pytest.raises(Exception): |
| sagemaker.local.image._check_output(["ls", "/some/unknown/path"]) |
|
|
| msg = "hello!" |
|
|
| output = sagemaker.local.image._check_output(["echo", msg]).strip() |
| assert output == msg |
|
|
| output = sagemaker.local.image._check_output("echo %s" % msg).strip() |
| assert output == msg |
|
|
|
|
| @patch("sagemaker.local.local_session.LocalSession", Mock()) |
| @patch("sagemaker.local.image._stream_output", Mock()) |
| @patch("sagemaker.local.image._SageMakerContainer._cleanup") |
| @patch("sagemaker.local.image._SageMakerContainer.retrieve_artifacts") |
| @patch("sagemaker.local.data.get_data_source_instance") |
| @patch("subprocess.Popen") |
| def test_train( |
| popen, get_data_source_instance, retrieve_artifacts, cleanup, tmpdir, sagemaker_session, caplog |
| ): |
| data_source = Mock() |
| data_source.get_root_dir.return_value = "foo" |
| get_data_source_instance.return_value = data_source |
|
|
| caplog.set_level(logging.INFO) |
|
|
| directories = [str(tmpdir.mkdir("container-root")), str(tmpdir.mkdir("data"))] |
| with patch( |
| "sagemaker.local.image._SageMakerContainer._create_tmp_folder", side_effect=directories |
| ): |
|
|
| instance_count = 2 |
| image = "my-image" |
| sagemaker_container = _SageMakerContainer( |
| "local", instance_count, image, sagemaker_session=sagemaker_session |
| ) |
| sagemaker_container.train( |
| INPUT_DATA_CONFIG, OUTPUT_DATA_CONFIG, HYPERPARAMETERS, ENVIRONMENT, TRAINING_JOB_NAME |
| ) |
|
|
| docker_compose_file = os.path.join( |
| sagemaker_container.container_root, "docker-compose.yaml" |
| ) |
| call_args = popen.call_args[0][0] |
| assert call_args is not None |
|
|
| expected = [ |
| "docker-compose", |
| "-f", |
| docker_compose_file, |
| "up", |
| "--build", |
| "--abort-on-container-exit", |
| ] |
| for i, v in enumerate(expected): |
| assert call_args[i] == v |
|
|
| with open(docker_compose_file, "r") as f: |
| config = yaml.load(f, Loader=yaml.SafeLoader) |
| assert len(config["services"]) == instance_count |
| for h in sagemaker_container.hosts: |
| assert config["services"][h]["image"] == image |
| assert config["services"][h]["command"] == "train" |
| |
| |
| assert ( |
| "TRAINING_JOB_NAME={}".format(TRAINING_JOB_NAME) |
| in config["services"][h]["environment"] |
| ) |
|
|
| |
| assert os.path.exists(os.path.join(sagemaker_container.container_root, "output")) |
| assert os.path.exists(os.path.join(sagemaker_container.container_root, "output/data")) |
|
|
| retrieve_artifacts.assert_called_once() |
| cleanup.assert_called_once() |
| assert "[Masked]" in caplog.text |
|
|
|
|
| @patch("sagemaker.local.local_session.LocalSession", Mock()) |
| @patch("sagemaker.local.image._stream_output", Mock()) |
| @patch("sagemaker.local.image._SageMakerContainer._cleanup", Mock()) |
| @patch("sagemaker.local.data.get_data_source_instance") |
| def test_train_with_hyperparameters_without_job_name( |
| get_data_source_instance, tmpdir, sagemaker_session |
| ): |
| data_source = Mock() |
| data_source.get_root_dir.return_value = "foo" |
| get_data_source_instance.return_value = data_source |
|
|
| directories = [str(tmpdir.mkdir("container-root")), str(tmpdir.mkdir("data"))] |
| with patch( |
| "sagemaker.local.image._SageMakerContainer._create_tmp_folder", side_effect=directories |
| ): |
| instance_count = 2 |
| image = "my-image" |
| sagemaker_container = _SageMakerContainer( |
| "local", instance_count, image, sagemaker_session=sagemaker_session |
| ) |
| sagemaker_container.train( |
| INPUT_DATA_CONFIG, OUTPUT_DATA_CONFIG, HYPERPARAMETERS, ENVIRONMENT, TRAINING_JOB_NAME |
| ) |
|
|
| docker_compose_file = os.path.join( |
| sagemaker_container.container_root, "docker-compose.yaml" |
| ) |
|
|
| with open(docker_compose_file, "r") as f: |
| config = yaml.load(f, Loader=yaml.SafeLoader) |
| for h in sagemaker_container.hosts: |
| assert ( |
| "TRAINING_JOB_NAME={}".format(TRAINING_JOB_NAME) |
| in config["services"][h]["environment"] |
| ) |
|
|
|
|
| @patch("sagemaker.local.local_session.LocalSession", Mock()) |
| @patch("sagemaker.local.image._stream_output", side_effect=RuntimeError("this is expected")) |
| @patch("sagemaker.local.image._SageMakerContainer._cleanup") |
| @patch("sagemaker.local.image._SageMakerContainer.retrieve_artifacts") |
| @patch("sagemaker.local.data.get_data_source_instance") |
| @patch("subprocess.Popen", Mock()) |
| def test_train_error( |
| get_data_source_instance, retrieve_artifacts, cleanup, _stream_output, tmpdir, sagemaker_session |
| ): |
| data_source = Mock() |
| data_source.get_root_dir.return_value = "foo" |
| get_data_source_instance.return_value = data_source |
|
|
| directories = [str(tmpdir.mkdir("container-root")), str(tmpdir.mkdir("data"))] |
| with patch( |
| "sagemaker.local.image._SageMakerContainer._create_tmp_folder", side_effect=directories |
| ): |
| instance_count = 2 |
| image = "my-image" |
| sagemaker_container = _SageMakerContainer( |
| "local", instance_count, image, sagemaker_session=sagemaker_session |
| ) |
|
|
| with pytest.raises(RuntimeError) as e: |
| sagemaker_container.train( |
| INPUT_DATA_CONFIG, |
| OUTPUT_DATA_CONFIG, |
| HYPERPARAMETERS, |
| ENVIRONMENT, |
| TRAINING_JOB_NAME, |
| ) |
|
|
| assert "this is expected" in str(e) |
|
|
| retrieve_artifacts.assert_called_once() |
| cleanup.assert_called_once() |
|
|
|
|
| @patch("sagemaker.local.local_session.LocalSession", Mock()) |
| @patch("sagemaker.local.image._stream_output", Mock()) |
| @patch("sagemaker.local.image._SageMakerContainer._cleanup", Mock()) |
| @patch("sagemaker.local.data.get_data_source_instance") |
| @patch("subprocess.Popen", Mock()) |
| def test_train_local_code(get_data_source_instance, tmpdir, sagemaker_session): |
| data_source = Mock() |
| data_source.get_root_dir.return_value = "foo" |
| get_data_source_instance.return_value = data_source |
|
|
| directories = [str(tmpdir.mkdir("container-root")), str(tmpdir.mkdir("data"))] |
| with patch( |
| "sagemaker.local.image._SageMakerContainer._create_tmp_folder", side_effect=directories |
| ): |
| instance_count = 2 |
| image = "my-image" |
| sagemaker_container = _SageMakerContainer( |
| "local", instance_count, image, sagemaker_session=sagemaker_session |
| ) |
|
|
| sagemaker_container.train( |
| INPUT_DATA_CONFIG, |
| OUTPUT_DATA_CONFIG, |
| LOCAL_CODE_HYPERPARAMETERS, |
| ENVIRONMENT, |
| TRAINING_JOB_NAME, |
| ) |
|
|
| docker_compose_file = os.path.join( |
| sagemaker_container.container_root, "docker-compose.yaml" |
| ) |
| shared_folder_path = os.path.join(sagemaker_container.container_root, "shared") |
|
|
| with open(docker_compose_file, "r") as f: |
| config = yaml.load(f, Loader=yaml.SafeLoader) |
| assert len(config["services"]) == instance_count |
|
|
| for h in sagemaker_container.hosts: |
| assert config["services"][h]["image"] == image |
| assert config["services"][h]["command"] == "train" |
| volumes = config["services"][h]["volumes"] |
| assert "%s:/opt/ml/code" % "/tmp/code" in volumes |
| assert "%s:/opt/ml/shared" % shared_folder_path in volumes |
|
|
| config_file_root = os.path.join( |
| sagemaker_container.container_root, h, "input", "config" |
| ) |
| hyperparameters_file = os.path.join(config_file_root, "hyperparameters.json") |
| hyperparameters_data = json.load(open(hyperparameters_file)) |
| assert hyperparameters_data["sagemaker_submit_directory"] == json.dumps("/opt/ml/code") |
|
|
|
|
| @patch("sagemaker.local.local_session.LocalSession", Mock()) |
| @patch("sagemaker.local.image._stream_output", Mock()) |
| @patch("sagemaker.local.image._SageMakerContainer._cleanup", Mock()) |
| @patch("sagemaker.local.data.get_data_source_instance") |
| @patch("subprocess.Popen", Mock()) |
| def test_train_local_intermediate_output(get_data_source_instance, tmpdir, sagemaker_session): |
| data_source = Mock() |
| data_source.get_root_dir.return_value = "foo" |
| get_data_source_instance.return_value = data_source |
|
|
| directories = [str(tmpdir.mkdir("container-root")), str(tmpdir.mkdir("data"))] |
| with patch( |
| "sagemaker.local.image._SageMakerContainer._create_tmp_folder", side_effect=directories |
| ): |
| instance_count = 2 |
| image = "my-image" |
| sagemaker_container = _SageMakerContainer( |
| "local", instance_count, image, sagemaker_session=sagemaker_session |
| ) |
|
|
| output_path = str(tmpdir.mkdir("customer_intermediate_output")) |
| output_data_config = {"S3OutputPath": "file://%s" % output_path} |
| hyperparameters = {"sagemaker_s3_output": output_path} |
|
|
| sagemaker_container.train( |
| INPUT_DATA_CONFIG, output_data_config, hyperparameters, ENVIRONMENT, TRAINING_JOB_NAME |
| ) |
|
|
| docker_compose_file = os.path.join( |
| sagemaker_container.container_root, "docker-compose.yaml" |
| ) |
| intermediate_folder_path = os.path.join(output_path, "output/intermediate") |
|
|
| with open(docker_compose_file, "r") as f: |
| config = yaml.load(f, Loader=yaml.SafeLoader) |
| assert len(config["services"]) == instance_count |
| for h in sagemaker_container.hosts: |
| assert config["services"][h]["image"] == image |
| assert config["services"][h]["command"] == "train" |
| volumes = config["services"][h]["volumes"] |
| assert "%s:/opt/ml/output/intermediate" % intermediate_folder_path in volumes |
|
|
|
|
| def test_container_has_gpu_support(tmpdir, sagemaker_session): |
| instance_count = 1 |
| image = "my-image" |
| sagemaker_container = _SageMakerContainer( |
| "local_gpu", instance_count, image, sagemaker_session=sagemaker_session |
| ) |
|
|
| docker_host = sagemaker_container._create_docker_host("host-1", {}, set(), "train", []) |
| assert "deploy" in docker_host |
| assert docker_host["deploy"] == { |
| "resources": {"reservations": {"devices": [{"capabilities": ["gpu"]}]}} |
| } |
|
|
|
|
| def test_container_does_not_enable_nvidia_docker_for_cpu_containers(sagemaker_session): |
| instance_count = 1 |
| image = "my-image" |
| sagemaker_container = _SageMakerContainer( |
| "local", instance_count, image, sagemaker_session=sagemaker_session |
| ) |
|
|
| docker_host = sagemaker_container._create_docker_host("host-1", {}, set(), "train", []) |
| assert "runtime" not in docker_host |
|
|
|
|
| @patch("sagemaker.local.image._HostingContainer.run", Mock()) |
| @patch("sagemaker.local.image._SageMakerContainer._prepare_serving_volumes", Mock(return_value=[])) |
| @patch("shutil.copy", Mock()) |
| @patch("shutil.copytree", Mock()) |
| def test_serve(tmpdir, sagemaker_session, caplog): |
| caplog.set_level(logging.INFO) |
| with patch( |
| "sagemaker.local.image._SageMakerContainer._create_tmp_folder", |
| return_value=str(tmpdir.mkdir("container-root")), |
| ): |
| image = "my-image" |
| sagemaker_container = _SageMakerContainer( |
| "local", 1, image, sagemaker_session=sagemaker_session |
| ) |
| environment = {"env1": 1, "env2": "b", "SAGEMAKER_SUBMIT_DIRECTORY": "s3://some/path"} |
|
|
| sagemaker_container.serve("/some/model/path", environment) |
| docker_compose_file = os.path.join( |
| sagemaker_container.container_root, "docker-compose.yaml" |
| ) |
|
|
| with open(docker_compose_file, "r") as f: |
| config = yaml.load(f, Loader=yaml.SafeLoader) |
|
|
| for h in sagemaker_container.hosts: |
| assert config["services"][h]["image"] == image |
| assert config["services"][h]["command"] == "serve" |
| assert "[Masked]" in caplog.text |
|
|
|
|
| @patch("sagemaker.local.image._HostingContainer.run", Mock()) |
| @patch("sagemaker.local.image._SageMakerContainer._prepare_serving_volumes", Mock(return_value=[])) |
| @patch("shutil.copy", Mock()) |
| @patch("shutil.copytree", Mock()) |
| def test_serve_local_code(tmpdir, sagemaker_session): |
| with patch( |
| "sagemaker.local.image._SageMakerContainer._create_tmp_folder", |
| return_value=str(tmpdir.mkdir("container-root")), |
| ): |
| image = "my-image" |
| sagemaker_container = _SageMakerContainer( |
| "local", 1, image, sagemaker_session=sagemaker_session |
| ) |
| environment = {"env1": 1, "env2": "b", "SAGEMAKER_SUBMIT_DIRECTORY": "file:///tmp/code"} |
|
|
| sagemaker_container.serve("/some/model/path", environment) |
| docker_compose_file = os.path.join( |
| sagemaker_container.container_root, "docker-compose.yaml" |
| ) |
|
|
| with open(docker_compose_file, "r") as f: |
| config = yaml.load(f, Loader=yaml.SafeLoader) |
|
|
| for h in sagemaker_container.hosts: |
| assert config["services"][h]["image"] == image |
| assert config["services"][h]["command"] == "serve" |
|
|
| volumes = config["services"][h]["volumes"] |
| assert "%s:/opt/ml/code" % "/tmp/code" in volumes |
| assert ( |
| "SAGEMAKER_SUBMIT_DIRECTORY=/opt/ml/code" |
| in config["services"][h]["environment"] |
| ) |
|
|
|
|
| @patch("sagemaker.local.image._HostingContainer.run", Mock()) |
| @patch("sagemaker.local.image._SageMakerContainer._prepare_serving_volumes", Mock(return_value=[])) |
| @patch("shutil.copy", Mock()) |
| @patch("shutil.copytree", Mock()) |
| def test_serve_local_code_no_env(tmpdir, sagemaker_session): |
| with patch( |
| "sagemaker.local.image._SageMakerContainer._create_tmp_folder", |
| return_value=str(tmpdir.mkdir("container-root")), |
| ): |
| image = "my-image" |
| sagemaker_container = _SageMakerContainer( |
| "local", 1, image, sagemaker_session=sagemaker_session |
| ) |
| sagemaker_container.serve("/some/model/path", {}) |
| docker_compose_file = os.path.join( |
| sagemaker_container.container_root, "docker-compose.yaml" |
| ) |
|
|
| with open(docker_compose_file, "r") as f: |
| config = yaml.load(f, Loader=yaml.SafeLoader) |
|
|
| for h in sagemaker_container.hosts: |
| assert config["services"][h]["image"] == image |
| assert config["services"][h]["command"] == "serve" |
|
|
|
|
| @patch("sagemaker.local.data.get_data_source_instance") |
| @patch("tarfile.is_tarfile") |
| @patch("tarfile.open", MagicMock()) |
| @patch("os.makedirs", Mock()) |
| def test_prepare_serving_volumes_with_s3_model( |
| is_tarfile, get_data_source_instance, sagemaker_session |
| ): |
| sagemaker_container = _SageMakerContainer( |
| "local", 1, "some-image", sagemaker_session=sagemaker_session |
| ) |
| sagemaker_container.container_root = "/tmp/container_root" |
|
|
| s3_data_source = Mock() |
| s3_data_source.get_root_dir.return_value = "/tmp/downloaded/data/" |
| s3_data_source.get_file_list.return_value = ["/tmp/downloaded/data/my_model.tar.gz"] |
| get_data_source_instance.return_value = s3_data_source |
| is_tarfile.return_value = True |
|
|
| volumes = sagemaker_container._prepare_serving_volumes("s3://bucket/my_model.tar.gz") |
| is_tarfile.assert_called_with("/tmp/downloaded/data/my_model.tar.gz") |
|
|
| assert len(volumes) == 1 |
| assert volumes[0].container_dir == "/opt/ml/model" |
| assert volumes[0].host_dir == "/tmp/downloaded/data/" |
|
|
|
|
| @patch("sagemaker.local.data.get_data_source_instance") |
| @patch("tarfile.is_tarfile", Mock(return_value=False)) |
| @patch("os.makedirs", Mock()) |
| def test_prepare_serving_volumes_with_local_model(get_data_source_instance, sagemaker_session): |
| sagemaker_container = _SageMakerContainer( |
| "local", 1, "some-image", sagemaker_session=sagemaker_session |
| ) |
| sagemaker_container.container_root = "/tmp/container_root" |
|
|
| local_file_data_source = Mock() |
| local_file_data_source.get_root_dir.return_value = "/path/to/my_model" |
| local_file_data_source.get_file_list.return_value = ["/path/to/my_model/model"] |
| get_data_source_instance.return_value = local_file_data_source |
|
|
| volumes = sagemaker_container._prepare_serving_volumes("file:///path/to/my_model") |
|
|
| assert len(volumes) == 1 |
| assert volumes[0].container_dir == "/opt/ml/model" |
| assert volumes[0].host_dir == "/path/to/my_model" |
|
|
|
|
| def test_ecr_login_non_ecr(): |
| session_mock = Mock() |
| result = sagemaker.local.image._ecr_login_if_needed(session_mock, "ubuntu") |
|
|
| session_mock.assert_not_called() |
| assert result is False |
|
|
|
|
| @patch("sagemaker.local.image._check_output", return_value="123451324") |
| @pytest.mark.parametrize( |
| "image", |
| [ |
| "520713654638.dkr.ecr.us-east-1.amazonaws.com/image-i-have:1.0", |
| "520713654638.dkr.ecr.us-iso-east-1.c2s.ic.gov/image-i-have:1.0", |
| ], |
| ) |
| def test_ecr_login_image_exists(_check_output, image): |
| session_mock = Mock() |
|
|
| result = sagemaker.local.image._ecr_login_if_needed(session_mock, image) |
|
|
| session_mock.assert_not_called() |
| _check_output.assert_called() |
| assert result is False |
|
|
|
|
| @patch("subprocess.Popen", return_value=Mock(autospec=subprocess.Popen)) |
| @patch("sagemaker.local.image._check_output", return_value="") |
| def test_ecr_login_needed(mock_check_output, popen): |
| session_mock = Mock() |
|
|
| token = "very-secure-token" |
| token_response = "AWS:%s" % token |
| b64_token = base64.b64encode(token_response.encode("utf-8")) |
| response = { |
| "authorizationData": [ |
| { |
| "authorizationToken": b64_token, |
| "proxyEndpoint": "https://520713654638.dkr.ecr.us-east-1.amazonaws.com", |
| } |
| ], |
| "ResponseMetadata": { |
| "RetryAttempts": 0, |
| "HTTPStatusCode": 200, |
| "RequestId": "25b2ac63-36bf-11e8-ab6a-e5dc597d2ad9", |
| }, |
| } |
| session_mock.client("ecr").get_authorization_token.return_value = response |
| image = "520713654638.dkr.ecr.us-east-1.amazonaws.com/image-i-need:1.1" |
| |
| popen.return_value.communicate.return_value = (None, None) |
|
|
| result = sagemaker.local.image._ecr_login_if_needed(session_mock, image) |
|
|
| mock_check_output.assert_called_with(f"docker images -q {image}") |
| expected_command = [ |
| "docker", |
| "login", |
| "https://520713654638.dkr.ecr.us-east-1.amazonaws.com", |
| "-u", |
| "AWS", |
| "--password-stdin", |
| ] |
| popen.assert_called_with(expected_command, stdin=subprocess.PIPE) |
| popen.return_value.communicate.assert_called_with(input=token.encode()) |
| session_mock.client("ecr").get_authorization_token.assert_called_with( |
| registryIds=["520713654638"] |
| ) |
|
|
| assert result is True |
|
|
|
|
| @patch("subprocess.check_output", return_value="".encode("utf-8")) |
| def test_pull_image(check_output): |
| image = "520713654638.dkr.ecr.us-east-1.amazonaws.com/image-i-need:1.1" |
|
|
| sagemaker.local.image._pull_image(image) |
|
|
| expected_command = "docker pull %s" % image |
|
|
| check_output.assert_called_once_with(expected_command.split()) |
|
|
|
|
| def test__aws_credentials_with_long_lived_credentials(): |
| credentials = Credentials(access_key=_random_string(), secret_key=_random_string(), token=None) |
| session = Mock() |
| session.get_credentials.return_value = credentials |
|
|
| aws_credentials = _aws_credentials(session) |
|
|
| assert aws_credentials == [ |
| "AWS_ACCESS_KEY_ID=%s" % credentials.access_key, |
| "AWS_SECRET_ACCESS_KEY=%s" % credentials.secret_key, |
| ] |
|
|
|
|
| @patch("sagemaker.local.image._aws_credentials_available_in_metadata_service") |
| def test__aws_credentials_with_short_lived_credentials_and_ec2_metadata_service_having_credentials( |
| mock, |
| ): |
| credentials = Credentials( |
| access_key=_random_string(), secret_key=_random_string(), token=_random_string() |
| ) |
| session = Mock() |
| session.get_credentials.return_value = credentials |
| mock.return_value = True |
| aws_credentials = _aws_credentials(session) |
|
|
| assert aws_credentials is None |
|
|
|
|
| @patch("sagemaker.local.image._aws_credentials_available_in_metadata_service") |
| def test__aws_credentials_with_short_lived_credentials_and_ec2_metadata_service_having_no_credentials( |
| mock, |
| ): |
| credentials = Credentials( |
| access_key=_random_string(), secret_key=_random_string(), token=_random_string() |
| ) |
| session = Mock() |
| session.get_credentials.return_value = credentials |
| mock.return_value = False |
| aws_credentials = _aws_credentials(session) |
|
|
| assert aws_credentials == [ |
| "AWS_ACCESS_KEY_ID=%s" % credentials.access_key, |
| "AWS_SECRET_ACCESS_KEY=%s" % credentials.secret_key, |
| "AWS_SESSION_TOKEN=%s" % credentials.token, |
| ] |
|
|
|
|
| def _random_string(size=6, chars=string.ascii_uppercase): |
| return "".join(random.choice(chars) for x in range(size)) |
|
|