cancel
Showing results for 
Search instead for 
Did you mean: 
Get Started Discussions
Start your journey with Databricks by joining discussions on getting started guides, tutorials, and introductory topics. Connect with beginners and experts alike to kickstart your Databricks experience.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Phani1
by Valued Contributor II
  • 3844 Views
  • 5 replies
  • 1 kudos

Azure Synapse vs Databricks

 Hi team,Could you kindly provide your perspective on the cost and performance comparison between Azure Synapse and Databricks SQL Warehouse/serverless, as well as their respective use cases? Thank you.

  • 3844 Views
  • 5 replies
  • 1 kudos
Latest Reply
Witold
Honored Contributor
  • 1 kudos

@Suncat There hasn't been any major changes for than a year: https://learn.microsoft.com/en-us/azure/synapse-analytics/spark/apache-spark-version-support E.g. I don't believe we will see support for Spark 3.5 at all. At least, apparently it's support...

  • 1 kudos
4 More Replies
Phani1
by Valued Contributor II
  • 1760 Views
  • 4 replies
  • 0 kudos

Talend ETL code to Databricks

Hi Team,What is the best way to transfer Talend ETL code to Databricks and what are the best methods/practices for migrating Talend ETL's to Databricks (notebook, code conversion/migration strategy, workflow's etc)?Regards,Janga

  • 1760 Views
  • 4 replies
  • 0 kudos
Latest Reply
thelogicplus
Contributor
  • 0 kudos

if you're planning to migrate from your current technology to Databricks, Travinto Technologies' Code Converter Tool is here to make the process seamless. This powerful tool enables you to migrate data, ETL workflows, and reports across platforms eff...

  • 0 kudos
3 More Replies
Phani1
by Valued Contributor II
  • 468 Views
  • 3 replies
  • 1 kudos

Delta sharing vs CosmosDB

 Hi All,We have a situation where we write data to CosmosDB and create JSON data for a transaction table, which includes a mini statement in JSON format.Now, we want to introduce the concept of delta sharing and share the transaction table. The Java ...

  • 468 Views
  • 3 replies
  • 1 kudos
Latest Reply
Phani1
Valued Contributor II
  • 1 kudos

Thanks for  your reply,Right now, the team is transferring data from Databricks to Cosmos DB, and then they're using REST APIs to access that data. They handle about 100 requests per minute, with some tables needing around 100 requests per second due...

  • 1 kudos
2 More Replies
Phani1
by Valued Contributor II
  • 407 Views
  • 1 replies
  • 0 kudos

code vulnerabilities, code smells, and bugs

Hi Team, is there a way in Databricks to check for code vulnerabilities, code smells, and bugs?Note :Databricks native functionality only  

  • 407 Views
  • 1 replies
  • 0 kudos
Latest Reply
SathyaSDE
Contributor
  • 0 kudos

Hi,As far as I am aware, for security scanning/monitoring at Databricks account level, we have belowSAT - https://github.com/databricks-industry-solutions/security-analysis-toolhttps://www.databricks.com/trust/trusthttps://learn.microsoft.com/en-us/a...

  • 0 kudos
Phani1
by Valued Contributor II
  • 720 Views
  • 1 replies
  • 0 kudos

incremental loads without date column

Hi All,We are facing a situation where our data source is Snowflake, and the data is saved in a storage location(adls) in parquet format. However, the tables or data lack a date column or any incremental column for performing incremental loads to Dat...

  • 720 Views
  • 1 replies
  • 0 kudos
Latest Reply
-werners-
Esteemed Contributor III
  • 0 kudos

Ideally you would have some change tracking system (cdc f.e.) on the source tables (Streams in the case of Snowflake, Introduction to Streams | Snowflake Documentation).But that is not the case.So I think you approach is ok.  You cannot track what is...

  • 0 kudos
Phani1
by Valued Contributor II
  • 647 Views
  • 1 replies
  • 0 kudos

Oracle -> Oracle Golden Gate ->Databricks Delta lake

Hi All,We have a situation where we are collecting data from different Oracle instances.The customer is using Oracle GoldenGate to replicate this data into a storage location.From there, we can use Auto Loader or Delta Live Tables to read Avro files ...

  • 647 Views
  • 1 replies
  • 0 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 0 kudos

Hi @Phani1 ,In my opinion this is really good setup. You have push scenario where Oracle GoldenGate  is responsible for delivering data into storage, so you don't have to bother about extraction part. And autoloader is the best choice when it comes t...

  • 0 kudos
Phani1
by Valued Contributor II
  • 484 Views
  • 0 replies
  • 0 kudos

Denodo Connection Parameters.

 Hi All,We are establishing a connection from Denodo to Databricks. During the development phase, we utilized a personal access token associated with  developer account. However, this approach is not considered a best practice for production environm...

  • 484 Views
  • 0 replies
  • 0 kudos
Phani1
by Valued Contributor II
  • 769 Views
  • 1 replies
  • 0 kudos

Databricks serverless Vs snowflake

Hi All,We want to switch from Snowflake to Databricks SQL Warehouse/serverless to simplify our data layers and reduce data copies before the reporting layer. Please share the benefits of using serverless over Snowflake and any limitations you see. We...

  • 769 Views
  • 1 replies
  • 0 kudos
Latest Reply
-werners-
Esteemed Contributor III
  • 0 kudos

one big pro is that you do not need to copy data to the dwh. also your transformations and analytics queries reside on the same platform (databricks).If databricks can cover all the requirementsm compared to snowflakem is hard to tell.  Probably ther...

  • 0 kudos
Phani1
by Valued Contributor II
  • 1181 Views
  • 1 replies
  • 0 kudos

best practices for implementing early arriving fact handling

 Hi All,Can you please share us the best practices for implementing early arriving fact handling in databricks for streaming data processed in near real time using structured streaming.There are many ways to handle this use case in batch/mini batch. ...

Phani1_0-1724754033290.png
  • 1181 Views
  • 1 replies
  • 0 kudos
Latest Reply
Phani1
Valued Contributor II
  • 0 kudos

Greetings Team, I would like to inquire if any of you have suggestions regarding the query.

  • 0 kudos
Phani1
by Valued Contributor II
  • 707 Views
  • 1 replies
  • 0 kudos

autoloader data processing

 Hi Team,Can you share the best practices for designing the autoloader data processing?We have data from 30 countries data coming in various files. Currently, we are thinking of using a root folder i.e country, and with subfolders for the individual ...

  • 707 Views
  • 1 replies
  • 0 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 0 kudos

Hi @Phani1 ,Structure of folders that you are going to use make sense to me. Since you've mentioned that there will be thousands of files, the best practice will be to use autoloader with file notification mode.  Also, you can read about databricks r...

  • 0 kudos
Phani1
by Valued Contributor II
  • 1556 Views
  • 2 replies
  • 1 kudos

classic cluster vs serverless cost

Hi Team,Can you help me the cost comparison between classic cluster and serverless?

  • 1556 Views
  • 2 replies
  • 1 kudos
Latest Reply
Rishabh_Tiwari
Databricks Employee
  • 1 kudos

Hi @Phani1 , Thank you for reaching out to our community! We're here to help you.  To ensure we provide you with the best support, could you please take a moment to review the response and choose the one that best answers your question? Your feedback...

  • 1 kudos
1 More Replies
Labels