cancel
Showing results forย 
Search instead forย 
Did you mean:ย 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results forย 
Search instead forย 
Did you mean:ย 

Data Pipeline for Bringing Data from Oracle Fusion to Azure Databricks

Pratikmsbsvm
Contributor

I am trying to bring Oracle Fusion (SCM, HCM, Finance) Data and push to ADLS Gen2. Databricks used for Data Transformation and Powerbi used for Reports Visualization.

I have 3 Option.

Option 1 :

Option1.png

Option 2 : 

Option2.png

Option 3

Option3.png

May someone please help me which is best enterprise cost effective approach and why. Or any Other way to achieve this effectively.

Thanks a lot

1 ACCEPTED SOLUTION

Accepted Solutions

bianca_unifeye
New Contributor III

My preference is option 1 

 

  • Delta Sharing is the most efficient and secure integration between Databricks and external systems.

  • No JDBC bottlenecks (no long-running queries, no network saturation).

  • Data shared as Delta format, which is natively optimized for Databricks.

  • Lower operational overhead โ€” Databricks reads the Delta Shares directly.

  • Good for large volumes (Finance, SCM, HCM typically generate big datasets).

  • Strong governance and lineage support.

Also I don't like to use JDBC, I avoid using it unless there are no other options

 

  • Not scalable for large Oracle Fusion workloads.

  • JDBC pulls are:

    • slow

    • stateful

    • prone to timeouts

    • difficult to parallelize

    • expensive for large history loads

  • High latency for production-grade pipelines.

  • You must manage incremental logic manually (ROWIDs, timestamps, etc.).

 

 

View solution in original post

3 REPLIES 3

bianca_unifeye
New Contributor III

@Raman_Unifeye this one is for you ๐Ÿ˜

bianca_unifeye
New Contributor III

My preference is option 1 

 

  • Delta Sharing is the most efficient and secure integration between Databricks and external systems.

  • No JDBC bottlenecks (no long-running queries, no network saturation).

  • Data shared as Delta format, which is natively optimized for Databricks.

  • Lower operational overhead โ€” Databricks reads the Delta Shares directly.

  • Good for large volumes (Finance, SCM, HCM typically generate big datasets).

  • Strong governance and lineage support.

Also I don't like to use JDBC, I avoid using it unless there are no other options

 

  • Not scalable for large Oracle Fusion workloads.

  • JDBC pulls are:

    • slow

    • stateful

    • prone to timeouts

    • difficult to parallelize

    • expensive for large history loads

  • High latency for production-grade pipelines.

  • You must manage incremental logic manually (ROWIDs, timestamps, etc.).

 

 

Raman_Unifeye
Contributor III

Option-1 using Oracle's Bulk extraction utility BICC. It can directly export the extracted data files (typically CSV) to Oracle cloud storage destination and then you could use ADF to get it copied over to ADLS.