cancel
Showing results forย 
Search instead forย 
Did you mean:ย 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results forย 
Search instead forย 
Did you mean:ย 

Databricks-Autoloader-S3-KMS

shivam-singh
New Contributor

Hi, I am working on a requirement where I am using autoloader in a DLT pipeline to ingest new files as they come.
This flow is working fine. However I am facing an issue, when we have the source bucket an s3 location, since the bucket is having a SSE-KMS setup and when it tries to write schemaLocation, it is throwing an error.

What extra changes need to be done, I have provided the kms related permissions to the instance profile that is getting used. What changes I need to make in my notebook/pipeline to get it working. 

1 REPLY 1

kulkpd
Contributor

Can you please paste the exact errors and check below things:

check following if its related to KMS:
1. IAM role policy and KMS policy should have allow permissions

2. Did you use extraConfig while mounting the source-s3 bucket:
If you have used IAM role while mounting the bucket. Verify iam-role-arn has KMS related permissions::

dbutils.fs.mount(
"s3a://src-bucket",
"/mnt/my-mount",
"sse-s3"
extraConfigs = Map(
"fs.s3a.credentialsType" -> "AssumeRole",
"fs.s3a.stsAssumeRole.arn" -> "iam-role-arn"
)
)

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local communityโ€”sign up today to get started!

Sign Up Now