cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

nchittampelly
by New Contributor II
  • 3770 Views
  • 3 replies
  • 0 kudos

What is the best way to connect Oracle CRM cloud from databricks?

What is the best way to connect Oracle CRM cloud from databricks?

  • 3770 Views
  • 3 replies
  • 0 kudos
Latest Reply
nchittampelly
New Contributor II
  • 0 kudos

Oracle CRM on Demand is a Cloud platform not a relational database.Is there any proven solution for this requirement?

  • 0 kudos
2 More Replies
prasadvaze
by Valued Contributor II
  • 9879 Views
  • 5 replies
  • 6 kudos

Resolved! Limit on number of result rows displayed on databricks SQL UI

Databricks SQL UI currently limits the query results display to 64000 rows. When will this limit go away? Using SSMS I get 40MM rows results in the UI and my users won't switch to databricks SQL for this reason

  • 9879 Views
  • 5 replies
  • 6 kudos
Latest Reply
vsrmerc
New Contributor II
  • 6 kudos

want to understand the reason behind it. retrieving 500k records is not a problem, is it rendering over the http thats the problematic? 

  • 6 kudos
4 More Replies
Nidhig
by Contributor
  • 406 Views
  • 4 replies
  • 5 kudos

Lakeflow jobs

 Hi I am currently working on migrating all ADF jobs to LakeFlow jobs. I have a few questions:Pipeline cost: What is the cost model for running LakeFlow pipelines? Any documentation available? ADF vs Lakeflow Job?Job reuse: Do LakeFlow jobs reuse the...

  • 406 Views
  • 4 replies
  • 5 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 5 kudos

Hi @Nidhig ,1. Regarding pipeline cost - here you're mostly paying for compute usage. So the exact price depends on which plan you are at and which cloud provider you are using. For instance, for Azure premium plan and US East region you have followi...

  • 5 kudos
3 More Replies
RIDBX
by Contributor
  • 215 Views
  • 2 replies
  • 1 kudos

How to make streaming files?

Thanks for reviewing my threads.I am trying to test streaming table /files in databricks FREE edition.-- Create test streaming tableCREATE OR REFRESH STREAMING TABLE user.demo.test_bronze_st ASSELECT * FROM STREAM read_files('/Volumes/xxx_ws/demo/raw...

  • 215 Views
  • 2 replies
  • 1 kudos
Latest Reply
RIDBX
Contributor
  • 1 kudos

Thanks for weighing in. Are you saying CREATE OR REFRESH STREAMING TABLE user.demo.test_bronze_st  cannot be used in FREE Edition?If we can use it, how do to create STREAM read_files('/Volumes/xxx_ws/demo/raw_files/test.csv'),where .csv sitting on lo...

  • 1 kudos
1 More Replies
William_Scardua
by Valued Contributor
  • 399 Views
  • 1 replies
  • 3 kudos

Resolved! What the best Framework/Package for data quality

Hi everyone,I’m currently looking for a data-quality solution for my environment. I don’t have DTL tables or a Unity Catalog in place.In your opinion, what is the best framework or package to implement reliable data-quality checks under these conditi...

  • 399 Views
  • 1 replies
  • 3 kudos
Latest Reply
nayan_wylde
Esteemed Contributor
  • 3 kudos

Here are few DQ packages for DLT or LDP that you can try.1. Databricks Labs DQXPurpose-built for Spark and Databricks.Rule-based checks on DataFrames (batch & streaming).Supports quarantine and profiling.Lightweight and easy to integrate.2. Great Exp...

  • 3 kudos
ShivMukesh
by New Contributor
  • 4162 Views
  • 3 replies
  • 0 kudos

Upgrade HSM to UC using Ucx tool - workspace to workspace migration

Hello team,I understand that an automatic upgrade to UC utilizing the UCx tool (Databricks Lab project) is now available to complete this migration from HSM to UC in automate way. But does this tool allow workspace to workspace catalog/artifact migra...

  • 4162 Views
  • 3 replies
  • 0 kudos
Latest Reply
nayan_wylde
Esteemed Contributor
  • 0 kudos

@ShivMukesh I have used UCX to migrate to Unity catalog. It is a great tool. But it also involves lot of workarounds specially in group migration and table migration. In group migration it renames the old workspace group and assigns the same permissi...

  • 0 kudos
2 More Replies
Han_bbb
by New Contributor
  • 158 Views
  • 1 replies
  • 0 kudos

Need to restore my scripts from the legacy version

Dear support team,The last time I used databricks was back in 2024 and I have several scripts stored in it. I really need to get access to them now but I can't login in with message "User is not a member of this workspace." Please help. Thanks

  • 158 Views
  • 1 replies
  • 0 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 0 kudos

Hi @Han_bbb ,Could you provide more details? Is it workspace in your work or your private one? Which cloud provider? 

  • 0 kudos
Naveenkumar1811
by New Contributor III
  • 313 Views
  • 3 replies
  • 0 kudos

Resolved! Can we Change the ownership of Databricks Managed Secret to SP in Azure Data Bricks?

Hi Team,Earlier we faced an Issue where the jar file(Created by a old employee) in workspace directory is used as library in the cluster which is run from a SP. Since the employee left the org and the id got removed even though the SP is part of ADMI...

  • 313 Views
  • 3 replies
  • 0 kudos
Latest Reply
Coffee77
Contributor III
  • 0 kudos

I think there is no other way. In any case, here is how I usually configure my (all-purpose and jobs compute) clusters to access secrets via environment variables so that you don't have to update all references if some similar issue arises again. The...

  • 0 kudos
2 More Replies
DarioB
by New Contributor III
  • 328 Views
  • 1 replies
  • 1 kudos

Resolved! Issues recreating Tables with enableRowTracking and DBR16.4 and below

We are running a Deep Clone script to copy Catalogs between Environments; this script is run through a job (run by SP) with DBR 16.4.12.Some tables are Deep Cloned and other ones are Dropped and Recreated to load partial data. The ones dropped are re...

  • 328 Views
  • 1 replies
  • 1 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 1 kudos

Happy Monday @DarioB , I did some digging and would like to provide you with some helpful hints/tips. Thanks for the detailed context—this is a known rough edge in DBR 16.x when recreating tables that have row tracking materialized. What’s happening ...

  • 1 kudos
Volker
by Contributor
  • 5651 Views
  • 2 replies
  • 0 kudos

Structured Streaming schemaTrackingLocation does not work with starting_version

Hello Community,I came across a strange behviour when using structured streaming on top of a delta table. I have a stream that I wanted to start from a specific version  of a delta table using the option option("starting_version", x) because I did no...

Data Engineering
Delta Lake
schemaTrackingLocation
starting_version
structured streaming
  • 5651 Views
  • 2 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

This issue is related to how Delta Lake’s structured streaming interacts with schema evolution and options like startingVersion and schemaTrackingLocation. The behavior you've observed has been noted by other users, and can be subtle due to how check...

  • 0 kudos
1 More Replies
stevenayers-bge
by Contributor
  • 4664 Views
  • 2 replies
  • 1 kudos

Querying Unity Managed Tables from Redshift

I built a script about 6 months ago to make our Delta Tables accessible in Redshift for another team, but it's a bit nasty...Generate a delta lake manifest each time the databricks delta table is updatedRecreate the redshift external table (incase th...

  • 4664 Views
  • 2 replies
  • 1 kudos
Latest Reply
mark_ott
Databricks Employee
  • 1 kudos

There is indeed a better and more integrated way to make Delta Lake tables accessible in Redshift without manually generating manifests and dynamically creating external tables or partitions. Some important points and options: Databricks Delta Lake ...

  • 1 kudos
1 More Replies
Mangeysh
by New Contributor
  • 4229 Views
  • 2 replies
  • 0 kudos

Azure data bricks API for JSON output , displaying on UI

Hello AllI am new to Azure Data Bricks and trying to show the Azure data bricks table data onto UI using react JS. Lets say there 2 tables Emplyee and Salary , I need to join these two tables with empid and generate JSON out put and calling API (end ...

  • 4229 Views
  • 2 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

The most effective way to display joined data from Azure Databricks tables (like Employee and Salary) in a React JS UI involves exposing your Databricks data through an API and then consuming that API in your frontend. Flask can work, but there are b...

  • 0 kudos
1 More Replies
rvo19941
by New Contributor II
  • 4972 Views
  • 2 replies
  • 0 kudos

Auto Loader File Notification Mode not working with ADLS Gen2 and files written as a stream

Dear,I am working on a real-time use case and am therefore using Auto Loader with file notification to ingest json files from a Gen2 Azure Storage Account in real-time. Full refreshes of my table work fine but I noticed Auto Loader was not picking up...

Data Engineering
ADLS
Auto Loader
Event Subscription
File Notification
Queue Storage
  • 4972 Views
  • 2 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

Auto Loader file notification in Databricks relies on Azure Event Grid’s BlobCreated event to trigger notifications for newly created files in Azure Data Lake Gen2. The issue you’re experiencing is a known limitation when files are written via certai...

  • 0 kudos
1 More Replies
achntrl
by New Contributor
  • 5498 Views
  • 1 replies
  • 0 kudos

CI/CD - Databricks Asset Bundles - Deploy/destroy only bundles with changes after Merge Request

Hello everyone,We're in the process of migrating to Databricks and are encountering challenges implementing CI/CD using Databricks Asset Bundles. Our monorepo houses multiple independent bundles within a "dabs" directory, with only one team member wo...

  • 5498 Views
  • 1 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

Your challenge—reliably determining the subset of changed Databricks Asset Bundles after a Merge Request (MR) is merged into main for focused deploy/destroy CI/CD actions—is common in complex monorepo, multi-environment setups. Let’s break down the p...

  • 0 kudos
alesventus
by Contributor
  • 6047 Views
  • 1 replies
  • 0 kudos

Effectively refresh Power BI report based on Delta Lake

Hi, I have several Power BI reports based on Delta Lake tables that are refreshed every 4 hours. ETL process in Databricks is much cheaper that refresh of these Power BI reports. My questions are: if approach described below is correct and if there i...

alesventus_0-1723191725173.png alesventus_1-1723192163389.png
  • 6047 Views
  • 1 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

Current Approach Assessment Power BI Import Mode: Importing all table data results in full dataset refreshes, driving up compute and data transfer costs during each refresh. Delta Lake as Source: Databricks clusters are used for both ETL and respon...

  • 0 kudos

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels