import json import logging import os from dataclasses import dataclass import boto3 from common.response import ok logger = logging.getLogger('delete_training_jobs') logger.setLevel(logging.INFO) dynamodb = boto3.resource('dynamodb') training_job_table = dynamodb.Table(os.environ.get('TRAINING_JOB_TABLE')) s3_bucket_name = os.environ.get('S3_BUCKET_NAME') s3 = boto3.resource('s3') bucket = s3.Bucket(s3_bucket_name) @dataclass class DeleteTrainingJobsEvent: training_job_list: [str] def handler(event, ctx): logger.info(f'event: {event}') logger.info(f'ctx: {ctx}') body = DeleteTrainingJobsEvent(**json.loads(event['body'])) # unique list for preventing duplicate delete training_job_list = list(set(body.training_job_list)) for training_job in training_job_list: training = training_job_table.get_item(Key={'id': training_job}) if 'Item' not in training: continue logger.info(f'training: {training}') training = training['Item'] if 'input_s3_location' in training: prefix = training['input_s3_location'].replace(f"s3://{s3_bucket_name}/", "") logger.info(f'delete prefix: {prefix}') response = bucket.objects.filter(Prefix=prefix).delete() logger.info(f'delete response: {response}') if 'params' in training: if 'training_params' in training['params']: if 's3_model_path' in training['params']['training_params']: s3_model_path = training['params']['training_params']['s3_model_path'] prefix = s3_model_path.replace(f"s3://{s3_bucket_name}/", "") logger.info(f'delete prefix: {prefix}') response = bucket.objects.filter(Prefix=prefix).delete() logger.info(f'delete response: {response}') training_job_table.delete_item(Key={'id': training_job}) return ok(message='training jobs deleted')