| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| from __future__ import absolute_import |
|
|
| import datetime |
| import os |
| import uuid |
|
|
| import pytest |
| from mock import Mock |
|
|
| from sagemaker.analytics import ( |
| AnalyticsMetricsBase, |
| HyperparameterTuningJobAnalytics, |
| TrainingJobAnalytics, |
| ) |
|
|
| BUCKET_NAME = "mybucket" |
| REGION = "us-west-2" |
|
|
|
|
| @pytest.fixture() |
| def sagemaker_session(): |
| return create_sagemaker_session() |
|
|
|
|
| def create_sagemaker_session( |
| describe_training_result=None, |
| list_training_results=None, |
| metric_stats_results=None, |
| describe_tuning_result=None, |
| ): |
| boto_mock = Mock(name="boto_session", region_name=REGION) |
| sms = Mock( |
| name="sagemaker_session", |
| boto_session=boto_mock, |
| boto_region_name=REGION, |
| config=None, |
| local_mode=False, |
| ) |
| sms.default_bucket = Mock(name="default_bucket", return_value=BUCKET_NAME) |
| sms.sagemaker_client.describe_hyper_parameter_tuning_job = Mock( |
| name="describe_hyper_parameter_tuning_job", return_value=describe_tuning_result |
| ) |
| sms.sagemaker_client.describe_training_job = Mock( |
| name="describe_training_job", return_value=describe_training_result |
| ) |
| sms.sagemaker_client.list_training_jobs_for_hyper_parameter_tuning_job = Mock( |
| name="list_training_jobs_for_hyper_parameter_tuning_job", return_value=list_training_results |
| ) |
| cwm_mock = Mock(name="cloudwatch_client") |
| boto_mock.client = Mock(return_value=cwm_mock) |
| cwm_mock.get_metric_statistics = Mock(name="get_metric_statistics") |
| cwm_mock.get_metric_statistics.side_effect = cw_request_side_effect |
| return sms |
|
|
|
|
| def cw_request_side_effect( |
| Namespace, MetricName, Dimensions, StartTime, EndTime, Period, Statistics |
| ): |
| if _is_valid_request(Namespace, MetricName, Dimensions, StartTime, EndTime, Period, Statistics): |
| return _metric_stats_results() |
|
|
|
|
| def _is_valid_request(Namespace, MetricName, Dimensions, StartTime, EndTime, Period, Statistics): |
| could_watch_request = { |
| "Namespace": Namespace, |
| "MetricName": MetricName, |
| "Dimensions": Dimensions, |
| "StartTime": StartTime, |
| "EndTime": EndTime, |
| "Period": Period, |
| "Statistics": Statistics, |
| } |
| print(could_watch_request) |
| return could_watch_request == cw_request() |
|
|
|
|
| def cw_request(): |
| describe_training_result = _describe_training_result() |
| return { |
| "Namespace": "/aws/sagemaker/TrainingJobs", |
| "MetricName": "train:acc", |
| "Dimensions": [{"Name": "TrainingJobName", "Value": "my-training-job"}], |
| "StartTime": describe_training_result["TrainingStartTime"], |
| "EndTime": describe_training_result["TrainingEndTime"] + datetime.timedelta(minutes=1), |
| "Period": 60, |
| "Statistics": ["Average"], |
| } |
|
|
|
|
| def test_abstract_base_class(): |
| |
| with pytest.raises(TypeError) as _: |
| AnalyticsMetricsBase() |
|
|
|
|
| def test_tuner_name(sagemaker_session): |
| tuner = HyperparameterTuningJobAnalytics("my-tuning-job", sagemaker_session=sagemaker_session) |
| assert tuner.name == "my-tuning-job" |
| assert str(tuner).find("my-tuning-job") != -1 |
|
|
|
|
| @pytest.mark.parametrize("has_training_job_definition_name", [True, False]) |
| def test_tuner_dataframe(has_training_job_definition_name): |
| training_job_definition_name = "training_def_1" |
|
|
| def mock_summary(name="job-name", value=0.9): |
| summary = { |
| "TrainingJobName": name, |
| "TrainingJobStatus": "Completed", |
| "FinalHyperParameterTuningJobObjectiveMetric": {"Name": "awesomeness", "Value": value}, |
| "TrainingStartTime": datetime.datetime(2018, 5, 16, 1, 2, 3), |
| "TrainingEndTime": datetime.datetime(2018, 5, 16, 5, 6, 7), |
| "TunedHyperParameters": {"learning_rate": 0.1, "layers": 137}, |
| } |
|
|
| if has_training_job_definition_name: |
| summary["TrainingJobDefinitionName"] = training_job_definition_name |
| return summary |
|
|
| session = create_sagemaker_session( |
| list_training_results={ |
| "TrainingJobSummaries": [ |
| mock_summary(), |
| mock_summary(), |
| mock_summary(), |
| mock_summary(), |
| mock_summary(), |
| ] |
| } |
| ) |
|
|
| tuner = HyperparameterTuningJobAnalytics("my-tuning-job", sagemaker_session=session) |
| df = tuner.dataframe() |
| assert df is not None |
| assert len(df) == 5 |
| assert ( |
| len(session.sagemaker_client.list_training_jobs_for_hyper_parameter_tuning_job.mock_calls) |
| == 1 |
| ) |
|
|
| |
| tuner.clear_cache() |
| df = tuner.dataframe() |
| assert ( |
| len(session.sagemaker_client.list_training_jobs_for_hyper_parameter_tuning_job.mock_calls) |
| == 2 |
| ) |
| df = tuner.dataframe(force_refresh=True) |
| assert ( |
| len(session.sagemaker_client.list_training_jobs_for_hyper_parameter_tuning_job.mock_calls) |
| == 3 |
| ) |
|
|
| |
| assert len(df["layers"]) == 5 |
| assert min(df["layers"]) == 137 |
|
|
| |
| assert min(df["TrainingElapsedTimeSeconds"]) > 5 |
| assert max(df["TrainingElapsedTimeSeconds"]) < 86400 |
|
|
| if has_training_job_definition_name: |
| for index in range(0, 5): |
| assert df["TrainingJobDefinitionName"][index] == training_job_definition_name |
| else: |
| assert "TrainingJobDefinitionName" not in df |
|
|
| |
| tmp_name = "/tmp/unit-test-%s.csv" % uuid.uuid4() |
| assert not os.path.isfile(tmp_name) |
| tuner.export_csv(tmp_name) |
| assert os.path.isfile(tmp_name) |
| os.unlink(tmp_name) |
|
|
|
|
| def test_description(): |
| session = create_sagemaker_session( |
| describe_tuning_result={ |
| "HyperParameterTuningJobConfig": { |
| "ParameterRanges": { |
| "CategoricalParameterRanges": [], |
| "ContinuousParameterRanges": [ |
| {"MaxValue": "1", "MinValue": "0", "Name": "eta"}, |
| {"MaxValue": "10", "MinValue": "0", "Name": "gamma"}, |
| ], |
| "IntegerParameterRanges": [ |
| {"MaxValue": "30", "MinValue": "5", "Name": "num_layers"}, |
| {"MaxValue": "100", "MinValue": "50", "Name": "iterations"}, |
| ], |
| } |
| }, |
| "TrainingJobDefinition": { |
| "AlgorithmSpecification": { |
| "TrainingImage": "training_image", |
| "TrainingInputMode": "File", |
| } |
| }, |
| } |
| ) |
|
|
| tuner = HyperparameterTuningJobAnalytics("my-tuning-job", sagemaker_session=session) |
|
|
| d = tuner.description() |
| assert len(session.sagemaker_client.describe_hyper_parameter_tuning_job.mock_calls) == 1 |
| assert d is not None |
| assert d["HyperParameterTuningJobConfig"] is not None |
| tuner.clear_cache() |
| d = tuner.description() |
| assert len(session.sagemaker_client.describe_hyper_parameter_tuning_job.mock_calls) == 2 |
| d = tuner.description() |
| assert len(session.sagemaker_client.describe_hyper_parameter_tuning_job.mock_calls) == 2 |
| d = tuner.description(force_refresh=True) |
| assert len(session.sagemaker_client.describe_hyper_parameter_tuning_job.mock_calls) == 3 |
|
|
| |
| r = tuner.tuning_ranges |
| assert len(r) == 4 |
|
|
|
|
| def test_tuning_ranges_multi_training_job_definitions(): |
| session = create_sagemaker_session( |
| describe_tuning_result={ |
| "HyperParameterTuningJobConfig": {}, |
| "TrainingJobDefinitions": [ |
| { |
| "DefinitionName": "estimator_1", |
| "HyperParameterRanges": { |
| "CategoricalParameterRanges": [], |
| "ContinuousParameterRanges": [ |
| {"MaxValue": "1", "MinValue": "0", "Name": "eta"}, |
| {"MaxValue": "10", "MinValue": "0", "Name": "gamma"}, |
| ], |
| "IntegerParameterRanges": [ |
| {"MaxValue": "30", "MinValue": "5", "Name": "num_layers"}, |
| {"MaxValue": "100", "MinValue": "50", "Name": "iterations"}, |
| ], |
| }, |
| "AlgorithmSpecification": { |
| "TrainingImage": "training_image_1", |
| "TrainingInputMode": "File", |
| }, |
| }, |
| { |
| "DefinitionName": "estimator_2", |
| "HyperParameterRanges": { |
| "CategoricalParameterRanges": [ |
| {"Values": ["TF", "MXNet"], "Name": "framework"} |
| ], |
| "ContinuousParameterRanges": [ |
| {"MaxValue": "1.0", "MinValue": "0.2", "Name": "gamma"} |
| ], |
| "IntegerParameterRanges": [], |
| }, |
| "AlgorithmSpecification": { |
| "TrainingImage": "training_image_2", |
| "TrainingInputMode": "File", |
| }, |
| }, |
| ], |
| } |
| ) |
|
|
| expected_result = { |
| "estimator_1": { |
| "eta": {"MaxValue": "1", "MinValue": "0", "Name": "eta"}, |
| "gamma": {"MaxValue": "10", "MinValue": "0", "Name": "gamma"}, |
| "iterations": {"MaxValue": "100", "MinValue": "50", "Name": "iterations"}, |
| "num_layers": {"MaxValue": "30", "MinValue": "5", "Name": "num_layers"}, |
| }, |
| "estimator_2": { |
| "framework": {"Values": ["TF", "MXNet"], "Name": "framework"}, |
| "gamma": {"MaxValue": "1.0", "MinValue": "0.2", "Name": "gamma"}, |
| }, |
| } |
|
|
| tuner = HyperparameterTuningJobAnalytics("my-tuning-job", sagemaker_session=session) |
|
|
| assert expected_result == tuner.tuning_ranges |
|
|
|
|
| def test_trainer_name(): |
| describe_training_result = { |
| "TrainingStartTime": datetime.datetime(2018, 5, 16, 1, 2, 3), |
| "TrainingEndTime": datetime.datetime(2018, 5, 16, 5, 6, 7), |
| } |
| session = create_sagemaker_session(describe_training_result) |
| trainer = TrainingJobAnalytics("my-training-job", ["metric"], sagemaker_session=session) |
| assert trainer.name == "my-training-job" |
| assert str(trainer).find("my-training-job") != -1 |
|
|
|
|
| def _describe_training_result(): |
| return { |
| "TrainingStartTime": datetime.datetime(2018, 5, 16, 1, 2, 3), |
| "TrainingEndTime": datetime.datetime(2018, 5, 16, 5, 6, 7), |
| } |
|
|
|
|
| def _metric_stats_results(): |
| return { |
| "Datapoints": [ |
| {"Average": 77.1, "Timestamp": datetime.datetime(2018, 5, 16, 1, 3, 3)}, |
| {"Average": 87.1, "Timestamp": datetime.datetime(2018, 5, 16, 1, 8, 3)}, |
| {"Average": 97.1, "Timestamp": datetime.datetime(2018, 5, 16, 2, 3, 3)}, |
| ] |
| } |
|
|
|
|
| def test_trainer_dataframe(): |
| session = create_sagemaker_session( |
| describe_training_result=_describe_training_result(), |
| metric_stats_results=_metric_stats_results(), |
| ) |
| trainer = TrainingJobAnalytics("my-training-job", ["train:acc"], sagemaker_session=session) |
|
|
| df = trainer.dataframe() |
| assert df is not None |
| assert len(df) == 3 |
| assert min(df["value"]) == 77.1 |
| assert max(df["value"]) == 97.1 |
|
|
| |
| tmp_name = "/tmp/unit-test-%s.csv" % uuid.uuid4() |
| assert not os.path.isfile(tmp_name) |
| trainer.export_csv(tmp_name) |
| assert os.path.isfile(tmp_name) |
| os.unlink(tmp_name) |
|
|
|
|
| def test_start_time_end_time_and_period_specified(): |
| describe_training_result = { |
| "TrainingStartTime": datetime.datetime(2018, 5, 16, 1, 2, 3), |
| "TrainingEndTime": datetime.datetime(2018, 5, 16, 5, 6, 7), |
| } |
| session = create_sagemaker_session(describe_training_result) |
| start_time = datetime.datetime(2018, 5, 16, 1, 3, 4) |
| end_time = datetime.datetime(2018, 5, 16, 5, 1, 1) |
| period = 300 |
| trainer = TrainingJobAnalytics( |
| "my-training-job", |
| ["metric"], |
| sagemaker_session=session, |
| start_time=start_time, |
| end_time=end_time, |
| period=period, |
| ) |
|
|
| assert trainer._time_interval["start_time"] == start_time |
| assert trainer._time_interval["end_time"] == end_time |
| assert trainer._period == period |
|
|