cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

stacihoover109
by Visitor
  • 14 Views
  • 0 replies
  • 0 kudos

The Rise of Human-Centered AI Design

This topic explores how artificial intelligence is shifting from purely technical efficiency toward human-centered design principles. It covers inclusive UX, ethical frameworks, and neurodiverse empowerment, showing how AI tools can be crafted to sup...

  • 14 Views
  • 0 replies
  • 0 kudos
ralphchan
by New Contributor II
  • 4062 Views
  • 6 replies
  • 0 kudos

Connect Oracle Fusion (ERP / HCM) to Databricks

Any suggestion to connect Oracle Fusion (ERP/HCM) to Databricks?I have explored a few options including the use of Oracle Integration Cloud but it requires a lot of customization.

  • 4062 Views
  • 6 replies
  • 0 kudos
Latest Reply
Shankar-Raj
  • 0 kudos

Check out BI Connector(https://www.biconnector.com/oracle-fusion-data-warehouse-integration/), which offers direct connectivity from Oracle Fusion/FDI/OAC to DW/Lakehouse/Power BI/Tableau and more.   It's widely deployed among Oracle Fusion & Analyti...

  • 0 kudos
5 More Replies
Pratikmsbsvm
by Contributor
  • 146 Views
  • 4 replies
  • 3 kudos

Resolved! Data Pipeline for Bringing Data from Oracle Fusion to Azure Databricks

I am trying to bring Oracle Fusion (SCM, HCM, Finance) Data and push to ADLS Gen2. Databricks used for Data Transformation and Powerbi used for Reports Visualization.I have 3 Option.Option 1 :Option 2 : Option 3May someone please help me which is bes...

Option1.png Option2.png Option3.png
  • 146 Views
  • 4 replies
  • 3 kudos
Latest Reply
Shankar-Raj
  • 3 kudos

BICC is suitable for certain use cases, but it has several limitations and is not particularly user-friendly.  BICC uses PVOs which cause a huge operational gap among the users:IT/DW teams: Its a multi-hop(BICC-file system-OCI-ADW-delta share) proces...

  • 3 kudos
3 More Replies
saicharandeepb
by New Contributor III
  • 426 Views
  • 5 replies
  • 1 kudos

How to Retrieve DBU Count per Compute Type for Accurate Cost Calculation?

Hello Everyone,We are currently working on a cost analysis initiative to gain deeper insights into our Databricks usage. As part of this effort, we are trying to calculate the hourly cost of each Databricks compute instance by utilizing the Azure Ret...

  • 426 Views
  • 5 replies
  • 1 kudos
Latest Reply
yogeshsingh
Databricks Employee
  • 1 kudos

Databricks doesn’t expose a “DBU/hour per node type” lookup in system tables, but you can derive it empirically by dividing total DBUs recorded in the billable usage system table by total node runtime from the compute node timeline for each instance ...

  • 1 kudos
4 More Replies
Mathias_Peters
by Contributor II
  • 21 Views
  • 0 replies
  • 0 kudos

Streamed DLT Pipeline using a lookup table

Hi, I need to join three streams/streamed data sets in a DLT pipeline. I am reading from a Kinesis data stream a sequence of events per group key. The logically first of the events per group contains a marker which determines whether that group is re...

  • 21 Views
  • 0 replies
  • 0 kudos
Surya-Prathap
by New Contributor II
  • 93 Views
  • 4 replies
  • 2 kudos

Support Request: Issue Running Multiple Ingestion Gateway Concurrently

Hi, we are ingesting data using Databricks Lake flow SQL connector from two different SQL Server databases hosted on separate servers. As part of the setup:We created two separate Ingestion Gateways.We created two separate ingestion pipelines.Both pi...

  • 93 Views
  • 4 replies
  • 2 kudos
Latest Reply
HarishPrasath25
  • 2 kudos

Hi @Louis_Frolio , I’ve successfully ingested one SQL database using the Lakeflow SQL connector. As part of the setup, I created an ingestion pipeline along with a gateway, and it is working as expected - when I run or re-run the pipeline, it picks u...

  • 2 kudos
3 More Replies
dpc
by Contributor
  • 47 Views
  • 3 replies
  • 3 kudos

disabling a task in a databricks job

Hello I have jobs that perform a number of activities.Sometimes I want to disable one or more of these activities.I can do that easily in an app like ADF.Reading around I cannot find an easy way to do this although, what I've read suggest that it was...

  • 47 Views
  • 3 replies
  • 3 kudos
Latest Reply
dpc
Contributor
  • 3 kudos

Thanks @Raman_Unifeye Do you know if there are any plans to add this feature?

  • 3 kudos
2 More Replies
tnorlund
by New Contributor
  • 61 Views
  • 2 replies
  • 0 kudos

Serverless compute cannot access internet

I'm experiencing "Network is unreachable" errors when trying to access external APIs from serverless compute jobs, despite having a network policy configured to allow access to all destinations. I have the "Serverless egress control" in the account s...

  • 61 Views
  • 2 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

The ongoing "Network is unreachable" errors in Databricks serverless compute suggest that network egress is being blocked, even though you have "Full access" set for Serverless Egress Control and a permissive network policy. The partial DNS resolutio...

  • 0 kudos
1 More Replies
v1b3_c0d3rXXX
by Databricks MVP
  • 3268 Views
  • 6 replies
  • 2 kudos

Resolved! Trouble Enabling File Events For An External Location

Hello all,I am trying to enable file events on my Azure Workspace for the File Arrival Trigger trigger mode for Databricks Workflows. I'm following this documentation exactly (I think) but I'm not seeing the option to enable them. As you can see here...

Dnt_TchTheRolex_1-1746752104483.png Dnt_TchTheRolex_0-1746751988442.png
  • 3268 Views
  • 6 replies
  • 2 kudos
Latest Reply
Drupe
New Contributor II
  • 2 kudos

I also have an external location that is a ADLS v2 however from databricks I only want to grant read access to the managed identity. I want to use file arrival trigger but want to have option of using more than 50 such triggers. Is it possible to onl...

  • 2 kudos
5 More Replies
adihc
by New Contributor II
  • 5489 Views
  • 12 replies
  • 1 kudos

Resolved! Options to access files in the community edition

As of now DBFS option is disabled in the Databricks community edition. What are the other ways to use file in the Databricks notebooks for learning? When I go to catalog it show default option only with the AWS S3. Is it the only option to access the...

  • 5489 Views
  • 12 replies
  • 1 kudos
Latest Reply
MastanRaoM
New Contributor II
  • 1 kudos

Hi Chandra,Is it possible in new community edition.

  • 1 kudos
11 More Replies
Yogesh_Verma_
by Contributor
  • 77 Views
  • 1 replies
  • 2 kudos

Databricks Architecture Center

 Databricks Architecture Center — Your Blueprint for Building Modern Data & AI PlatformsThe Databricks Architecture Center is a centralized knowledge hub that provides:End-to-end reference architecturesIndustry-specific patternsArchitecture decision ...

Yogesh_Verma__0-1763628453509.png
  • 77 Views
  • 1 replies
  • 2 kudos
Latest Reply
Raman_Unifeye
Contributor III
  • 2 kudos

Its very useful. I suppose you missed to provide the link. Here is the link for easy accesshttps://www.databricks.com/resources/architectures 

  • 2 kudos
Suheb
by New Contributor II
  • 46 Views
  • 3 replies
  • 0 kudos

How do I choose between a standard cluster and a shared cluster in Databricks?

When should I use a single-user cluster, and when should I use a multi-user/shared cluster? What’s the difference and how do I pick the right one?

  • 46 Views
  • 3 replies
  • 0 kudos
Latest Reply
Coffee77
Contributor III
  • 0 kudos

Agree with previous responses BUT consider using a dedicated cluster if using it under an "automation security principal" identity for a subset of your jobs, if that makes sense. To summarize, you need to run a job, under a non-human identity and you...

  • 0 kudos
2 More Replies
RevanthV
by New Contributor III
  • 73 Views
  • 3 replies
  • 0 kudos

POC on spark 4.x

I need to do some POC with spark 3.5.7 and 4.x and need some local setup with some sample Kafka source. The POC would read data from Kafka via streaming job and write to delta table and I would like to do this on spark-4.x ..Do you know of any quick ...

  • 73 Views
  • 3 replies
  • 0 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 0 kudos

Hello @RevanthV , I did some digging and here are some helpful tips. Got it — here are fast, reproducible ways to stand up a local Kafka source and run a Spark Structured Streaming job that writes to a Delta table, plus the common fixes for the conne...

  • 0 kudos
2 More Replies
Deenar
by New Contributor II
  • 622 Views
  • 2 replies
  • 1 kudos

Dashboard Filters (Showing Description) but selecting the id value for use query parameters

HiIs there a way to have a Single or multiple filter in a Dashboard that shows titles/names but on selection passes in the identifier as query parameter filters. The cost management dashboard shipped by Databricks seems to do this for the workspace s...

Deenar_0-1740567476099.png
  • 622 Views
  • 2 replies
  • 1 kudos
Latest Reply
200052
New Contributor II
  • 1 kudos

I wanted to share my solution to this, hope this helps. Create a SQL function that accepts an ARRAY<STRING>, extracts a numeric ID enclosed in parentheses from the end of each string, and returns an ARRAY<BIGINT>.Example UsageInput: ["Item (1)", "Ite...

  • 1 kudos
1 More Replies

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels