cancel
Showing results forย 
Search instead forย 
Did you mean:ย 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results forย 
Search instead forย 
Did you mean:ย 

Databricks-Autoloader-S3-KMS

shivam-singh
New Contributor

Hi, I am working on a requirement where I am using autoloader in a DLT pipeline to ingest new files as they come.
This flow is working fine. However I am facing an issue, when we have the source bucket an s3 location, since the bucket is having a SSE-KMS setup and when it tries to write schemaLocation, it is throwing an error.

What extra changes need to be done, I have provided the kms related permissions to the instance profile that is getting used. What changes I need to make in my notebook/pipeline to get it working. 

1 REPLY 1

kulkpd
Contributor

Can you please paste the exact errors and check below things:

check following if its related to KMS:
1. IAM role policy and KMS policy should have allow permissions

2. Did you use extraConfig while mounting the source-s3 bucket:
If you have used IAM role while mounting the bucket. Verify iam-role-arn has KMS related permissions::

dbutils.fs.mount(
"s3a://src-bucket",
"/mnt/my-mount",
"sse-s3"
extraConfigs = Map(
"fs.s3a.credentialsType" -> "AssumeRole",
"fs.s3a.stsAssumeRole.arn" -> "iam-role-arn"
)
)

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you wonโ€™t want to miss the chance to attend and share knowledge.

If there isnโ€™t a group near you, start one and help create a community that brings people together.

Request a New Group