| import json |
| import logging |
| import os |
| import subprocess |
| import sys |
| import textwrap |
| from pathlib import Path |
| from typing import Callable, Dict, Generator, Optional |
| from uuid import uuid4 |
|
|
| import boto3 |
| import pytest |
| from aws_cdk import App, CfnOutput, Environment, RemovalPolicy, Stack, aws_logs |
| from aws_cdk.aws_lambda import ( |
| Architecture, |
| Code, |
| Function, |
| LayerVersion, |
| Runtime, |
| Tracing, |
| ) |
| from filelock import FileLock |
|
|
| from tests.e2e.utils.base import InfrastructureProvider |
| from tests.e2e.utils.constants import ( |
| CDK_OUT_PATH, |
| PYTHON_RUNTIME_VERSION, |
| SOURCE_CODE_ROOT_PATH, |
| ) |
| from tests.e2e.utils.lambda_layer.powertools_layer import LocalLambdaPowertoolsLayer |
|
|
| logger = logging.getLogger(__name__) |
|
|
|
|
| class BaseInfrastructure(InfrastructureProvider): |
| RANDOM_STACK_VALUE: str = f"{uuid4()}" |
|
|
| def __init__(self) -> None: |
| self.feature_path = Path(sys.modules[self.__class__.__module__].__file__).parent |
| self.feature_name = self.feature_path.parts[-1].replace("_", "-") |
| self.stack_name = f"test{PYTHON_RUNTIME_VERSION}-{self.feature_name}-{self.RANDOM_STACK_VALUE}" |
| self.stack_outputs: Dict[str, str] = {} |
|
|
| |
| self.session = boto3.session.Session() |
| self.cfn = self.session.client("cloudformation") |
| self.account_id = self.session.client("sts").get_caller_identity()["Account"] |
| self.region = self.session.region_name |
|
|
| self.app = App() |
| self.stack = Stack(self.app, self.stack_name, env=Environment(account=self.account_id, region=self.region)) |
|
|
| |
| self._feature_infra_class_name = self.__class__.__name__ |
| self._feature_infra_module_path = self.feature_path / "infrastructure" |
| self._feature_infra_file = self.feature_path / "infrastructure.py" |
| self._handlers_dir = self.feature_path / "handlers" |
| self._cdk_out_dir: Path = CDK_OUT_PATH / self.feature_name |
| self._stack_outputs_file = f'{self._cdk_out_dir / "stack_outputs.json"}' |
|
|
| if not self._feature_infra_file.exists(): |
| raise FileNotFoundError( |
| "You must have your infrastructure defined in 'tests/e2e/<feature>/infrastructure.py'.", |
| ) |
|
|
| def create_lambda_functions( |
| self, |
| function_props: Optional[Dict] = None, |
| architecture: Architecture = Architecture.X86_64, |
| ) -> Dict[str, Function]: |
| """Create Lambda functions available under handlers_dir |
| |
| It creates CloudFormation Outputs for every function found in PascalCase. For example, |
| {handlers_dir}/basic_handler.py creates `BasicHandler` and `BasicHandlerArn` outputs. |
| |
| |
| Parameters |
| ---------- |
| function_props: Optional[Dict] |
| Dictionary representing CDK Lambda FunctionProps to override defaults |
| |
| architecture: Architecture |
| Used to create Lambda Layer and functions in a different architecture. Defaults to x86_64. |
| |
| Returns |
| ------- |
| output: Dict[str, Function] |
| A dict with PascalCased function names and the corresponding CDK Function object |
| |
| Examples |
| -------- |
| |
| Creating Lambda functions available in the handlers directory |
| |
| ```python |
| self.create_lambda_functions() |
| ``` |
| |
| Creating Lambda functions and override runtime to Python 3.13 |
| |
| ```python |
| from aws_cdk.aws_lambda import Runtime |
| |
| self.create_lambda_functions(function_props={"runtime": Runtime.PYTHON_3_13) |
| ``` |
| """ |
| if not self._handlers_dir.exists(): |
| raise RuntimeError(f"Handlers dir '{self._handlers_dir}' must exist for functions to be created.") |
|
|
| layer_build = LocalLambdaPowertoolsLayer(architecture=architecture).build() |
| layer = LayerVersion( |
| self.stack, |
| "aws-lambda-powertools-e2e-test", |
| layer_version_name="aws-lambda-powertools-e2e-test", |
| compatible_runtimes=[ |
| Runtime.PYTHON_3_8, |
| Runtime.PYTHON_3_9, |
| Runtime.PYTHON_3_10, |
| Runtime.PYTHON_3_11, |
| Runtime.PYTHON_3_12, |
| Runtime.PYTHON_3_13, |
| ], |
| compatible_architectures=[architecture], |
| code=Code.from_asset(path=layer_build), |
| ) |
|
|
| |
| |
| handlers = list(self._handlers_dir.rglob("*.py")) |
| source = Code.from_asset(f"{self._handlers_dir}") |
| logger.debug(f"Creating functions for handlers: {handlers}") |
|
|
| function_settings_override = function_props or {} |
| output: Dict[str, Function] = {} |
|
|
| for fn in handlers: |
| fn_name = fn.stem |
| fn_name_pascal_case = fn_name.title().replace("_", "") |
| logger.debug(f"Creating function: {fn_name_pascal_case}") |
| function_settings = { |
| "id": f"{fn_name}-lambda", |
| "code": source, |
| "handler": f"{fn_name}.lambda_handler", |
| "tracing": Tracing.ACTIVE, |
| "runtime": self._determine_runtime_version(), |
| "layers": [layer], |
| "architecture": architecture, |
| **function_settings_override, |
| } |
|
|
| function = Function(self.stack, **function_settings) |
|
|
| aws_logs.LogGroup( |
| self.stack, |
| id=f"{fn_name}-lg", |
| log_group_name=f"/aws/lambda/{function.function_name}", |
| retention=aws_logs.RetentionDays.ONE_DAY, |
| removal_policy=RemovalPolicy.DESTROY, |
| ) |
|
|
| |
| self.add_cfn_output(name=fn_name_pascal_case, value=function.function_name, arn=function.function_arn) |
|
|
| output[fn_name_pascal_case] = function |
|
|
| return output |
|
|
| def deploy(self) -> Dict[str, str]: |
| """Synthesize and deploy a CDK app, and return its stack outputs |
| |
| NOTE: It auto-generates a temporary CDK app to benefit from CDK CLI lookup features |
| |
| Returns |
| ------- |
| Dict[str, str] |
| CloudFormation Stack Outputs with output key and value |
| """ |
| stack_file = self._create_temp_cdk_app() |
| synth_command = f"npx cdk synth --app 'python {stack_file}' -o {self._cdk_out_dir}" |
| deploy_command = ( |
| f"npx cdk deploy --app '{self._cdk_out_dir}' -O {self._stack_outputs_file} " |
| "--require-approval=never --method=direct" |
| ) |
|
|
| |
| subprocess.check_output(synth_command, shell=True) |
| subprocess.check_output(deploy_command, shell=True) |
| return self._read_stack_output() |
|
|
| def delete(self) -> None: |
| """Delete CloudFormation Stack""" |
| logger.debug(f"Deleting stack: {self.stack_name}") |
| self.cfn.delete_stack(StackName=self.stack_name) |
|
|
| def _sync_stack_name(self, stack_output: Dict): |
| """Synchronize initial stack name with CDK final stack name |
| |
| When using `cdk synth` with context methods (`from_lookup`), |
| CDK can initialize the Stack multiple times until it resolves |
| the context. |
| |
| Parameters |
| ---------- |
| stack_output : Dict |
| CDK CloudFormation Outputs, where the key is the stack name |
| """ |
| self.stack_name = list(stack_output.keys())[0] |
|
|
| def _read_stack_output(self): |
| content = Path(self._stack_outputs_file).read_text() |
| outputs: Dict = json.loads(content) |
| self._sync_stack_name(stack_output=outputs) |
|
|
| |
| self.stack_outputs = list(outputs.values())[0] |
| return self.stack_outputs |
|
|
| def _create_temp_cdk_app(self): |
| """Autogenerate a CDK App with our Stack so that CDK CLI can deploy it |
| |
| This allows us to keep our BaseInfrastructure while supporting context lookups. |
| """ |
| |
| temp_file = self._cdk_out_dir / f"cdk_app_{PYTHON_RUNTIME_VERSION}.py" |
|
|
| if temp_file.exists(): |
| |
| return temp_file |
|
|
| |
| infra_module = str(self._feature_infra_module_path.relative_to(SOURCE_CODE_ROOT_PATH)).replace(os.sep, ".") |
|
|
| code = f""" |
| from {infra_module} import {self._feature_infra_class_name} |
| stack = {self._feature_infra_class_name}() |
| stack.create_resources() |
| stack.app.synth() |
| """ |
|
|
| if not self._cdk_out_dir.is_dir(): |
| self._cdk_out_dir.mkdir(parents=True, exist_ok=True) |
|
|
| with temp_file.open("w") as fd: |
| fd.write(textwrap.dedent(code)) |
|
|
| |
| temp_file.chmod(0o755) |
| return temp_file |
|
|
| def _determine_runtime_version(self) -> Runtime: |
| """Determine Python runtime version based on the current Python interpreter""" |
| version = sys.version_info |
| if version.major == 3 and version.minor == 8: |
| return Runtime.PYTHON_3_8 |
| elif version.major == 3 and version.minor == 9: |
| return Runtime.PYTHON_3_9 |
| elif version.major == 3 and version.minor == 10: |
| return Runtime.PYTHON_3_10 |
| elif version.major == 3 and version.minor == 11: |
| return Runtime.PYTHON_3_11 |
| elif version.major == 3 and version.minor == 12: |
| return Runtime.PYTHON_3_12 |
| elif version.major == 3 and version.minor == 13: |
| return Runtime.PYTHON_3_13 |
| else: |
| raise ValueError(f"Unsupported Python version: {version}") |
|
|
| def create_resources(self) -> None: |
| """Create any necessary CDK resources. It'll be called before deploy |
| |
| Examples |
| ------- |
| |
| Creating a S3 bucket and export name and ARN |
| |
| ```python |
| def created_resources(self): |
| s3 = s3.Bucket(self.stack, "MyBucket") |
| |
| # This will create MyBucket and MyBucketArn CloudFormation Output |
| self.add_cfn_output(name="MyBucket", value=s3.bucket_name, arn_value=bucket.bucket_arn) |
| ``` |
| |
| Creating Lambda functions available in the handlers directory |
| |
| ```python |
| def created_resources(self): |
| self.create_lambda_functions() |
| ``` |
| """ |
| raise NotImplementedError() |
|
|
| def add_cfn_output(self, name: str, value: str, arn: str = ""): |
| """Create {Name} and optionally {Name}Arn CloudFormation Outputs. |
| |
| Parameters |
| ---------- |
| name : str |
| CloudFormation Output Key |
| value : str |
| CloudFormation Output Value |
| arn : str |
| CloudFormation Output Value for ARN |
| """ |
| CfnOutput(self.stack, f"{name}", value=value) |
| if arn: |
| CfnOutput(self.stack, f"{name}Arn", value=arn) |
|
|
|
|
| def call_once( |
| task: Callable, |
| tmp_path_factory: pytest.TempPathFactory, |
| worker_id: str, |
| callback: Optional[Callable] = None, |
| ) -> Generator[object, None, None]: |
| """Call function and serialize results once whether CPU parallelization is enabled or not |
| |
| Parameters |
| ---------- |
| task : Callable |
| Function to call once and JSON serialize result whether parallel test is enabled or not. |
| tmp_path_factory : pytest.TempPathFactory |
| pytest temporary path factory to discover shared tmp when multiple CPU processes are spun up |
| worker_id : str |
| pytest-xdist worker identification to detect whether parallelization is enabled |
| callback : Callable |
| Function to call when job is complete. |
| |
| Yields |
| ------ |
| Generator[object, None, None] |
| Callable output when called |
| """ |
|
|
| try: |
| if worker_id == "master": |
| |
| yield task() |
| else: |
| |
| root_tmp_dir = tmp_path_factory.getbasetemp().parent |
| cache = root_tmp_dir / f"{PYTHON_RUNTIME_VERSION}_cache.json" |
|
|
| with FileLock(f"{cache}.lock"): |
| |
| |
| |
| if cache.is_file(): |
| callable_result = json.loads(cache.read_text()) |
| else: |
| callable_result: Dict = task() |
| cache.write_text(json.dumps(callable_result)) |
| yield callable_result |
| finally: |
| if callback is not None: |
| callback() |
|
|