@@ -232,6 +232,12 @@ def test_sagemaker_pyspark_sse_kms_s3(role, image_uri, sagemaker_session, region
232232 if not kms_key_id :
233233 raise ValueError ("AWS managed s3 kms key(alias: aws/s3) does not exist" )
234234
235+ # TODO: PDT is the only case requires different partition at this time,
236+ # in the future we need to change it to fixture
237+ aws_partition = "aws"
238+ if region == "us-gov-west-1" :
239+ aws_partition = "aws-us-gov"
240+
235241 bucket = sagemaker_session .default_bucket ()
236242 timestamp = datetime .now ().isoformat ()
237243 input_data_key = f"spark/input/sales/{ timestamp } /data.jsonl"
@@ -253,7 +259,7 @@ def test_sagemaker_pyspark_sse_kms_s3(role, image_uri, sagemaker_session, region
253259 "Classification" : "core-site" ,
254260 "Properties" : {
255261 "fs.s3a.server-side-encryption-algorithm" : "SSE-KMS" ,
256- "fs.s3a.server-side-encryption.key" : f"arn:aws :kms:{ region } :{ account_id } :key/{ kms_key_id } " ,
262+ "fs.s3a.server-side-encryption.key" : f"arn:{ aws_partition } :kms:{ region } :{ account_id } :key/{ kms_key_id } " ,
257263 },
258264 },
259265 )
@@ -270,7 +276,7 @@ def test_sagemaker_pyspark_sse_kms_s3(role, image_uri, sagemaker_session, region
270276 for s3_object in s3_objects :
271277 object_metadata = s3_client .get_object (Bucket = bucket , Key = s3_object ["Key" ])
272278 assert object_metadata ["ServerSideEncryption" ] == "aws:kms"
273- assert object_metadata ["SSEKMSKeyId" ] == f"arn:aws :kms:{ region } :{ account_id } :key/{ kms_key_id } "
279+ assert object_metadata ["SSEKMSKeyId" ] == f"arn:{ aws_partition } :kms:{ region } :{ account_id } :key/{ kms_key_id } "
274280
275281
276282def test_sagemaker_scala_jar_multinode (role , image_uri , configuration , sagemaker_session , sagemaker_client ):
0 commit comments