cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
cancel
Showing results for 
Search instead for 
Did you mean: 

Databricks-Autoloader-S3-KMS

shivam-singh
New Contributor

Hi, I am working on a requirement where I am using autoloader in a DLT pipeline to ingest new files as they come.
This flow is working fine. However I am facing an issue, when we have the source bucket an s3 location, since the bucket is having a SSE-KMS setup and when it tries to write schemaLocation, it is throwing an error.

What extra changes need to be done, I have provided the kms related permissions to the instance profile that is getting used. What changes I need to make in my notebook/pipeline to get it working. 

1 REPLY 1

kulkpd
Contributor

Can you please paste the exact errors and check below things:

check following if its related to KMS:
1. IAM role policy and KMS policy should have allow permissions

2. Did you use extraConfig while mounting the source-s3 bucket:
If you have used IAM role while mounting the bucket. Verify iam-role-arn has KMS related permissions::

dbutils.fs.mount(
"s3a://src-bucket",
"/mnt/my-mount",
"sse-s3"
extraConfigs = Map(
"fs.s3a.credentialsType" -> "AssumeRole",
"fs.s3a.stsAssumeRole.arn" -> "iam-role-arn"
)
)
Welcome to Databricks Community: Lets learn, network and celebrate together

Join our fast-growing data practitioner and expert community of 80K+ members, ready to discover, help and collaborate together while making meaningful connections. 

Click here to register and join today! 

Engage in exciting technical discussions, join a group with your peers and meet our Featured Members.