cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Maverick1
by Valued Contributor II
  • 4828 Views
  • 6 replies
  • 6 kudos

How to infer the online feature store table via an mlflow registered model, which is deployed to a sagemaker endpoint?

Can an mlflow registered model automatically infer the online feature store table, if that model is trained and logged via a databricks feature store table and the table is pushed to an online feature store (like AWS RDS)?

  • 4828 Views
  • 6 replies
  • 6 kudos
Latest Reply
Janifer45
New Contributor II
  • 6 kudos

Thanks for this

  • 6 kudos
5 More Replies
Prabakar
by Databricks Employee
  • 2709 Views
  • 1 replies
  • 2 kudos

Accessing the regions that are disabled by default in AWS from Databricks. In AWS we have 4 regions that are disabled by default. You must first enabl...

Accessing the regions that are disabled by default in AWS from Databricks.In AWS we have 4 regions that are disabled by default. You must first enable it before you can create and manage resources. The following Regions are disabled by default:Africa...

  • 2709 Views
  • 1 replies
  • 2 kudos
Latest Reply
AndreaCuda
New Contributor II
  • 2 kudos

Hello - We are looking to deploy and run Databricks in AWS in Bahrain, or UAE. Is this possible? This post is older so wondering if this is a viable option.

  • 2 kudos
Anonymous
by Not applicable
  • 6899 Views
  • 9 replies
  • 2 kudos

Resolved! Issue in creating workspace - Custom AWS Configuration

We have tried to create new workspace using "Custom AWS Configuration" and we have given our own VPC (Customer managed VPC) and tried but workspace failed to launch. We are getting below error which couldn't understand where the issue is in.Workspace...

  • 6899 Views
  • 9 replies
  • 2 kudos
Latest Reply
Briggsrr
New Contributor II
  • 2 kudos

Experiencing workspace launch failures with custom AWS configuration is frustrating. The "MALFORMED_REQUEST" error and failed network validation checks suggest a VPC configuration issue. It feels like playing Infinite Craft, endlessly combining eleme...

  • 2 kudos
8 More Replies
jlb0001
by New Contributor III
  • 3553 Views
  • 4 replies
  • 2 kudos

[AWS] How do you replace the Account Admin?

I need to remove an older admin that previously set up the Databricks Account. However, I get an error (even through I am also an Account Admin).How do I replace a prior account admin? Or at least remove their admin status and/or disable the accoun...

Databricks Permssion Error - Cannot Disable Original Account
  • 3553 Views
  • 4 replies
  • 2 kudos
Latest Reply
SunnySharma
New Contributor II
  • 2 kudos

Was this ever fixed as I have got the exact same issue as our original AWS Account Admin has left the business and I now cannot remove his Account Admin role (but I have disabled access etc)!

  • 2 kudos
3 More Replies
Anonymous
by Not applicable
  • 51676 Views
  • 7 replies
  • 12 kudos

How to connect and extract data from sharepoint using Databricks (AWS) ?

We are using Databricks (on AWS). We need to connect to SharePoint and extract & load data to Databricks Delta table. Any possible solution on this ?

  • 51676 Views
  • 7 replies
  • 12 kudos
Latest Reply
yliu
New Contributor III
  • 12 kudos

Wondering the same.. Can we use Sharepoint REST API to download the file and save to dbfs/external location and read it? 

  • 12 kudos
6 More Replies
boskicl
by New Contributor III
  • 33164 Views
  • 6 replies
  • 10 kudos

Resolved! Table write command stuck "Filtering files for query."

Hello all,Background:I am having an issue today with databricks using pyspark-sql and writing a delta table. The dataframe is made by doing an inner join between two tables and that is the table which I am trying to write to a delta table. The table ...

filtering job_info spill_memory
  • 33164 Views
  • 6 replies
  • 10 kudos
Latest Reply
timo199
New Contributor II
  • 10 kudos

Even if I vacuum and optimize, it keeps getting stuck.cluster type is r6gd.xlarge min:4, max:6driver type is r6gd.2xlarge

  • 10 kudos
5 More Replies
costi9992
by New Contributor III
  • 3311 Views
  • 3 replies
  • 2 kudos

Access Databricks API using IDP token

Hello,We have a databricks account & workspace, provided by AWS with SSO enabled. Is there any way to access databricks workspace API ( jobs/clusters, etc ) using a token retrieved from IdentityProvider ? We can access databricks workspace API with A...

  • 3311 Views
  • 3 replies
  • 2 kudos
Latest Reply
fpopa
New Contributor II
  • 2 kudos

Hey - Costin and Anonymous user, have you managed to get this working, do you have examples by any chance?I'm also trying something similar but I haven't been able to make it work.> authenticate and access the Databricks REST API by setting the Autho...

  • 2 kudos
2 More Replies
bricksdata
by New Contributor
  • 11354 Views
  • 4 replies
  • 0 kudos

Unable to authenticate against https://accounts.cloud.databricks.com as an account admin.

ProblemI'm unable to authenticate against the https://accounts.cloud.databricks.com endpoint even though I'm an account admin. I need it to assign account level groups to workspaces via the workspace assignment api (https://api-docs.databricks.com/re...

  • 11354 Views
  • 4 replies
  • 0 kudos
Latest Reply
137292
New Contributor II
  • 0 kudos

From this doc: To automate Databricks account-level functionality, you cannot use Databricks personal access tokens. Instead, you must use either OAuth tokens for Databricks account admin users or service principals. For more information, see:Use a s...

  • 0 kudos
3 More Replies
MoJaMa
by Databricks Employee
  • 10747 Views
  • 7 replies
  • 2 kudos
  • 10747 Views
  • 7 replies
  • 2 kudos
Latest Reply
User15848365773
New Contributor II
  • 2 kudos

Hi @amitca71 @atanu .. yes you can associate as many vpcs(workspace deployment fundamental) across regions and aws accounts to one single databricks aws account infact its one of the super powers of databricks platform and you can even track all thei...

  • 2 kudos
6 More Replies
AndyAtINX
by New Contributor III
  • 4146 Views
  • 4 replies
  • 1 kudos

Resolved! Error inviting user to workspace "Failed to add user: A user with email ... or username ... in different cases already exist in the account"

We have 3 workspaces - 1 old version in one AWS account, 2 latest versions in another.We are PAYG full edition, not using SSO.Our admins (existing DBX users in the `admins` group) can invite new users via the Admin Console from the 1 old and 1 new wo...

  • 4146 Views
  • 4 replies
  • 1 kudos
Latest Reply
Schneider-Elect
New Contributor II
  • 1 kudos

We are facing same issue, We are on azure. @AndyAtINX you mean if user exist in workspace with abc@gmail.com we should add the user in workspace2 with abc@gmail.com not ABC@GMAIL.COM. if this the case we tried this and its not working for us.

  • 1 kudos
3 More Replies
Fast_Lanes
by New Contributor II
  • 1671 Views
  • 2 replies
  • 3 kudos

Why am I getting shown estimated costs and charges (DBU in $) during community trial edition?

I recently signed up for the 14 day community trial and I noticed it was incurring or showing estimate costs on my Usage page shortly after I created my first workspace linked through AWS. Is this only for monitoring purposes or am I actually going t...

  • 1671 Views
  • 2 replies
  • 3 kudos
Latest Reply
Ajay-Pandey
Esteemed Contributor III
  • 3 kudos

Hi @Fast_Lanes No, you will not be charged for the trial period on databricks usage.It's just showing the usual charges that will be charged when you will not be using trial version.

  • 3 kudos
1 More Replies
danatsafe
by New Contributor
  • 7341 Views
  • 3 replies
  • 0 kudos

Amazon returns a 403 error code when trying to access an S3 Bucket

Hey! So far I have followed along with the Configure S3 access with instance profiles article to grant my cluster access to an S3 bucket. I have also made sure to disable IAM role passthrough on the cluster. Upon querying the bucket through a noteboo...

  • 7341 Views
  • 3 replies
  • 0 kudos
Latest Reply
winojoe
New Contributor III
  • 0 kudos

I had the same issue and I found a solutionFor me, the permission problems only exist when the Cluster's (compute's) Access mode is "Shared No Isolation".  When the Access Mode is either "Shared" or "Single User" then the IAM configuration seems to a...

  • 0 kudos
2 More Replies
parthsalvi
by Contributor
  • 3597 Views
  • 1 replies
  • 2 kudos

Amazon SES : boto3 credentials not found. DBR 11.2 Shared mode

We're trying to send email using Amazon SES using boto3.client in python. We've added SES Full access in clusters IAM Role.   We were able to send email in "No isolation shared" mode in DBR 11.2 using ses = boto3.client('ses', region_name='us-****-2'...

image
  • 3597 Views
  • 1 replies
  • 2 kudos
Latest Reply
JameDavi_51481
Contributor
  • 2 kudos

This appears to be an intentional design choice to prevent users from using the credentials of the host machine to carry out arbitrary AWS API calls. I really wish there was a workaround or setting to disable this behavior because we put a lot of wor...

  • 2 kudos
ivanychev
by Contributor II
  • 2445 Views
  • 1 replies
  • 1 kudos

Is there a way to avoid using EBS drives on workers with local NVMe SSD?

The Databricks on AWS docs claim that 30G + 150G EBS drives are mounter to every node by default. But if I use instance type like r5d.2xlarge, it already has local disk so I want to avoid mounting the 150G EBS drive to it. Is there a way to do it?We ...

  • 2445 Views
  • 1 replies
  • 1 kudos
Latest Reply
pabloanzorenac
New Contributor II
  • 1 kudos

Hey Ivan, did you find a way to do this?

  • 1 kudos
Thaw
by New Contributor III
  • 2333 Views
  • 2 replies
  • 3 kudos

Resolved! How to change Instance Family in CloudFormation in a Databricks trial mood?

I implemented Databrick on AWS and the template is used i3.xlarge. Could I use it for down Instance Family for cost optimization? Is i3.xlarge the minimum size to use Databricks in a trial mood? Thanks

  • 2333 Views
  • 2 replies
  • 3 kudos
Latest Reply
Thaw
New Contributor III
  • 3 kudos

Thank you so much for your reply to my question, @Vidula Khanna​ @Kaniz Fatma​ . After I took some study time, I understood the basics, and then I am on the way to Databricks.

  • 3 kudos
1 More Replies
Labels