| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| """Placeholder docstring""" |
| from __future__ import absolute_import |
|
|
| import os |
| import platform |
| import sys |
| import tempfile |
| from abc import ABCMeta |
| from abc import abstractmethod |
| from six import with_metaclass |
|
|
| from six.moves.urllib.parse import urlparse |
|
|
| import sagemaker.amazon.common |
| import sagemaker.local.utils |
| import sagemaker.utils |
|
|
|
|
| def get_data_source_instance(data_source, sagemaker_session): |
| """Return an Instance of :class:`sagemaker.local.data.DataSource`. |
| |
| The instance can handle the provided data_source URI. |
| |
| data_source can be either file:// or s3:// |
| |
| Args: |
| data_source (str): a valid URI that points to a data source. |
| sagemaker_session (:class:`sagemaker.session.Session`): a SageMaker Session to |
| interact with S3 if required. |
| |
| Returns: |
| sagemaker.local.data.DataSource: an Instance of a Data Source |
| |
| Raises: |
| ValueError: If parsed_uri scheme is neither `file` nor `s3` , raise an |
| error. |
| """ |
| parsed_uri = urlparse(data_source) |
| if parsed_uri.scheme == "file": |
| return LocalFileDataSource(parsed_uri.netloc + parsed_uri.path) |
| if parsed_uri.scheme == "s3": |
| return S3DataSource(parsed_uri.netloc, parsed_uri.path, sagemaker_session) |
| raise ValueError( |
| "data_source must be either file or s3. parsed_uri.scheme: {}".format(parsed_uri.scheme) |
| ) |
|
|
|
|
| def get_splitter_instance(split_type): |
| """Return an Instance of :class:`sagemaker.local.data.Splitter`. |
| |
| The instance returned is according to the specified `split_type`. |
| |
| Args: |
| split_type (str): either 'Line' or 'RecordIO'. Can be left as None to |
| signal no data split will happen. |
| |
| Returns |
| :class:`sagemaker.local.data.Splitter`: an Instance of a Splitter |
| """ |
| if split_type is None: |
| return NoneSplitter() |
| if split_type == "Line": |
| return LineSplitter() |
| if split_type == "RecordIO": |
| return RecordIOSplitter() |
| raise ValueError("Invalid Split Type: %s" % split_type) |
|
|
|
|
| def get_batch_strategy_instance(strategy, splitter): |
| """Return an Instance of :class:`sagemaker.local.data.BatchStrategy` according to `strategy` |
| |
| Args: |
| strategy (str): Either 'SingleRecord' or 'MultiRecord' |
| splitter (:class:`sagemaker.local.data.Splitter): splitter to get the data from. |
| |
| Returns |
| :class:`sagemaker.local.data.BatchStrategy`: an Instance of a BatchStrategy |
| """ |
| if strategy == "SingleRecord": |
| return SingleRecordStrategy(splitter) |
| if strategy == "MultiRecord": |
| return MultiRecordStrategy(splitter) |
| raise ValueError('Invalid Batch Strategy: %s - Valid Strategies: "SingleRecord", "MultiRecord"') |
|
|
|
|
| class DataSource(with_metaclass(ABCMeta, object)): |
| """Placeholder docstring""" |
|
|
| @abstractmethod |
| def get_file_list(self): |
| """Retrieve the list of absolute paths to all the files in this data source. |
| |
| Returns: |
| List[str]: List of absolute paths. |
| """ |
|
|
| @abstractmethod |
| def get_root_dir(self): |
| """Retrieve the absolute path to the root directory of this data source. |
| |
| Returns: |
| str: absolute path to the root directory of this data source. |
| """ |
|
|
|
|
| class LocalFileDataSource(DataSource): |
| """Represents a data source within the local filesystem.""" |
|
|
| def __init__(self, root_path): |
| super(LocalFileDataSource, self).__init__() |
|
|
| self.root_path = os.path.abspath(root_path) |
| if not os.path.exists(self.root_path): |
| raise RuntimeError("Invalid data source: %s does not exist." % self.root_path) |
|
|
| def get_file_list(self): |
| """Retrieve the list of absolute paths to all the files in this data source. |
| |
| Returns: |
| List[str] List of absolute paths. |
| """ |
| if os.path.isdir(self.root_path): |
| return [ |
| os.path.join(self.root_path, f) |
| for f in os.listdir(self.root_path) |
| if os.path.isfile(os.path.join(self.root_path, f)) |
| ] |
| return [self.root_path] |
|
|
| def get_root_dir(self): |
| """Retrieve the absolute path to the root directory of this data source. |
| |
| Returns: |
| str: absolute path to the root directory of this data source. |
| """ |
| if os.path.isdir(self.root_path): |
| return self.root_path |
| return os.path.dirname(self.root_path) |
|
|
|
|
| class S3DataSource(DataSource): |
| """Defines a data source given by a bucket and S3 prefix. |
| |
| The contents will be downloaded and then processed as local data. |
| """ |
|
|
| def __init__(self, bucket, prefix, sagemaker_session): |
| """Create an S3DataSource instance. |
| |
| Args: |
| bucket (str): S3 bucket name |
| prefix (str): S3 prefix path to the data |
| sagemaker_session (:class:`sagemaker.session.Session`): a sagemaker_session with the |
| desired settings |
| to talk to S3 |
| """ |
| super(S3DataSource, self).__init__() |
|
|
| |
| root_dir = sagemaker.utils.get_config_value( |
| "local.container_root", sagemaker_session.config |
| ) |
| if root_dir: |
| root_dir = os.path.abspath(root_dir) |
|
|
| working_dir = tempfile.mkdtemp(dir=root_dir) |
| |
| |
| |
| if root_dir is None and platform.system() == "Darwin": |
| working_dir = "/private{}".format(working_dir) |
|
|
| sagemaker.utils.download_folder(bucket, prefix, working_dir, sagemaker_session) |
| self.files = LocalFileDataSource(working_dir) |
|
|
| def get_file_list(self): |
| """Retrieve the list of absolute paths to all the files in this data source. |
| |
| Returns: |
| List[str]: List of absolute paths. |
| """ |
| return self.files.get_file_list() |
|
|
| def get_root_dir(self): |
| """Retrieve the absolute path to the root directory of this data source. |
| |
| Returns: |
| str: absolute path to the root directory of this data source. |
| """ |
| return self.files.get_root_dir() |
|
|
|
|
| class Splitter(with_metaclass(ABCMeta, object)): |
| """Placeholder docstring""" |
|
|
| @abstractmethod |
| def split(self, file): |
| """Split a file into records using a specific strategy |
| |
| Args: |
| file (str): path to the file to split |
| |
| Returns: |
| generator for the individual records that were split from the file |
| """ |
|
|
|
|
| class NoneSplitter(Splitter): |
| """Does not split records, essentially reads the whole file.""" |
|
|
| |
| _textchars = bytearray({7, 8, 9, 10, 12, 13, 27} | set(range(0x20, 0x100)) - {0x7F}) |
|
|
| def split(self, filename): |
| """Split a file into records using a specific strategy. |
| |
| For this NoneSplitter there is no actual split happening and the file |
| is returned as a whole. |
| |
| Args: |
| filename (str): path to the file to split |
| |
| Returns: generator for the individual records that were split from |
| the file |
| """ |
| with open(filename, "rb") as f: |
| buf = f.read() |
| if not self._is_binary(buf): |
| buf = buf.decode() |
| yield buf |
|
|
| def _is_binary(self, buf): |
| """Check whether `buf` contains binary data. |
| |
| Returns True if `buf` contains any non-utf-8 characters. |
| |
| Args: |
| buf (bytes): data to inspect |
| |
| Returns: |
| True if data is binary, otherwise False |
| """ |
| return bool(buf.translate(None, self._textchars)) |
|
|
|
|
| class LineSplitter(Splitter): |
| """Split records by new line.""" |
|
|
| def split(self, file): |
| """Split a file into records using a specific strategy |
| |
| This LineSplitter splits the file on each line break. |
| |
| Args: |
| file (str): path to the file to split |
| |
| Returns: generator for the individual records that were split from |
| the file |
| """ |
| with open(file, "r") as f: |
| for line in f: |
| yield line |
|
|
|
|
| class RecordIOSplitter(Splitter): |
| """Split using Amazon Recordio. |
| |
| Not useful for string content. |
| """ |
|
|
| def split(self, file): |
| """Split a file into records using a specific strategy |
| |
| This RecordIOSplitter splits the data into individual RecordIO |
| records. |
| |
| Args: |
| file (str): path to the file to split |
| |
| Returns: generator for the individual records that were split from |
| the file |
| """ |
| with open(file, "rb") as f: |
| for record in sagemaker.amazon.common.read_recordio(f): |
| yield record |
|
|
|
|
| class BatchStrategy(with_metaclass(ABCMeta, object)): |
| """Placeholder docstring""" |
|
|
| def __init__(self, splitter): |
| """Create a Batch Strategy Instance |
| |
| Args: |
| splitter (sagemaker.local.data.Splitter): A Splitter to pre-process |
| the data before batching. |
| """ |
| self.splitter = splitter |
|
|
| @abstractmethod |
| def pad(self, file, size): |
| """Group together as many records as possible to fit in the specified size. |
| |
| Args: |
| file (str): file path to read the records from. |
| size (int): maximum size in MB that each group of records will be |
| fitted to. passing 0 means unlimited size. |
| |
| Returns: |
| generator of records |
| """ |
|
|
|
|
| class MultiRecordStrategy(BatchStrategy): |
| """Feed multiple records at a time for batch inference. |
| |
| Will group up as many records as possible within the payload specified. |
| """ |
|
|
| def pad(self, file, size=6): |
| """Group together as many records as possible to fit in the specified size. |
| |
| Args: |
| file (str): file path to read the records from. |
| size (int): maximum size in MB that each group of records will be |
| fitted to. passing 0 means unlimited size. |
| |
| Returns: |
| generator of records |
| """ |
| buffer = "" |
| for element in self.splitter.split(file): |
| if _payload_size_within_limit(buffer + element, size): |
| buffer += element |
| else: |
| tmp = buffer |
| buffer = element |
| yield tmp |
| if _validate_payload_size(buffer, size): |
| yield buffer |
|
|
|
|
| class SingleRecordStrategy(BatchStrategy): |
| """Feed a single record at a time for batch inference. |
| |
| If a single record does not fit within the payload specified it will |
| throw a RuntimeError. |
| """ |
|
|
| def pad(self, file, size=6): |
| """Group together as many records as possible to fit in the specified size. |
| |
| This SingleRecordStrategy will not group any record and will return |
| them one by one as long as they are within the maximum size. |
| |
| Args: |
| file (str): file path to read the records from. |
| size (int): maximum size in MB that each group of records will be |
| fitted to. passing 0 means unlimited size. |
| |
| Returns: |
| generator of records |
| """ |
| for element in self.splitter.split(file): |
| if _validate_payload_size(element, size): |
| yield element |
|
|
|
|
| def _payload_size_within_limit(payload, size): |
| """Placeholder docstring.""" |
| size_in_bytes = size * 1024 * 1024 |
| if size == 0: |
| return True |
| return sys.getsizeof(payload) < size_in_bytes |
|
|
|
|
| def _validate_payload_size(payload, size): |
| """Check if a payload is within the size in MB threshold. |
| |
| Raise an exception if the payload is beyond the size in MB threshold. |
| |
| Args: |
| payload: data that will be checked |
| size (int): max size in MB |
| |
| Returns: |
| bool: True if within bounds. if size=0 it will always return True |
| |
| Raises: |
| RuntimeError: If the payload is larger a runtime error is thrown. |
| """ |
|
|
| if _payload_size_within_limit(payload, size): |
| return True |
| raise RuntimeError("Record is larger than %sMB. Please increase your max_payload" % size) |
|
|