"amzn...123" attributes (Map) pref_name : "大阪府" pop_start : "12:00" pop_end : "18:00" ... Source Storage - S3 Full s3://{bucket_name}/user_setting/full/{date}/AWSDynamoDB/{export_id}/data/{fil e_id}.json.gz { "Item": { "id": "amzn...123", "attributes": { "pop_start": "12:00", ... } } } Incremental s3://{bucket_name}/user_setting/incremental/{date}/AWSDynamoDB/data/{file_i d}.jsonl.gz { "Keys": { "id": "amzn...123" }, "OldImage": { "attributes": { "pop_start": "12:00", ... } }, "NewImage": { "attributes": { "pop_start": "06:00", ... } }, ... } Bronze (ST) Full テーブル b_user_setting_full record source_path ... {"Ite m":...} s3://.../full/04-1 6/... ... Incremental テーブル b_user_setting_incremental record source_path ... {"Key s":...} s3://.../incrementa l/04-17/... ... DynamoDB to S3 - Lambda // incremental response = dynamodb.export_table_to_point_in_time( TableArn=TABLE_ARN, S3Bucket=S3_BUCKET, S3Prefix=f"user_setting/incremental/{now_jst.strftime('%Y-%m-%d')}", ExportFormat="DYNAMODB_JSON", ExportType="INCREMENTAL_EXPORT", IncrementalExportSpecification={ "ExportFromTime": export_from, "ExportToTime": export_to } ) S3 to Bronze - Lakeflow SDP // incremental CREATE STREAMING TABLE b_user_setting_incremental AS SELECT value AS record, _metadata.file_path AS source_path, current_timestamp() AS ingest_timestamp FROM cloud_files( 's3://{bucket_name}/user_setting/incremental/*/AWSDynamoDB/data/', 'text' );