cancel
Showing results for 
Search instead for 
Did you mean: 
Community Discussions
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

JordiDekker
by New Contributor III
  • 1590 Views
  • 2 replies
  • 0 kudos

ClassCastException when attempting to timetravel (databricks-connect)

Hi all,Using databricks-connect 11.3.19, I get an "java.lang.ClassCastException" when attempting to timetravel. The exact same statement works fine when executed in the databricks GUI directly. Any ideas on what's going on? Is this a known limitation...

  • 1590 Views
  • 2 replies
  • 0 kudos
Latest Reply
SusanaD
New Contributor II
  • 0 kudos

Did you find a solution?

  • 0 kudos
1 More Replies
kartik-chandra
by New Contributor III
  • 857 Views
  • 2 replies
  • 0 kudos

Resolved! Spark read with format as "delta" isn't working with Java multithreading

0I have a Spark application (using Java library) which needs to replicate data from one blob storage to another. I have created a readStream() within it which is listening continuously to a Kafka topic for incoming events. The corresponding writeStre...

  • 857 Views
  • 2 replies
  • 0 kudos
Latest Reply
kartik-chandra
New Contributor III
  • 0 kudos

The problem was indeed with the way ClassLoader was being set in the ForkJoinPool (common Pool used) thread. Spark in SparkClassUtils uses Thread.currentThread().getContextClassLoader which might behave differently in another thread.To solve it I cre...

  • 0 kudos
1 More Replies
A1459
by New Contributor
  • 498 Views
  • 0 replies
  • 0 kudos

Execute delete query from notebook on azure synapse

Hello Everyone, Is there a way we can execute the delete query from azure notebook on azure synapse database.I tried using read api method with option "query" but getting error like jdbc connector not able to handle code.Can any suggest how we can de...

  • 498 Views
  • 0 replies
  • 0 kudos
Sujitha
by Community Manager
  • 3171 Views
  • 1 replies
  • 5 kudos

New how-to guide to data warehousing with the Data Intelligence Platform

Just launched: The Big Book of Data Warehousing and BI, a new hands-on guide focused on real-world use cases from governance, transformation, analytics and AI.As the demand for data becomes insatiable in every company, the data infrastructure has bec...

Screenshot 2023-12-05 at 11.04.08 AM.png
  • 3171 Views
  • 1 replies
  • 5 kudos
Latest Reply
Edward3
New Contributor II
  • 5 kudos

lol beans It used to take me a long time to regain my equilibrium, but recently I learned that a website really leads this layout when you may find delight after a stressful day here. Since then, I've been able to find my equilibrium much more quickl...

  • 5 kudos
invalidargument
by New Contributor II
  • 360 Views
  • 2 replies
  • 0 kudos

Create new workbooks with code

Is is possible to create new notebooks from a notevbook in databricks? I have tried this code. But all of them are generic files, not notebooks.notebook_str = """# Databricks notebook source import pyspark.sql.functions as F import numpy as np # CO...

  • 360 Views
  • 2 replies
  • 0 kudos
Latest Reply
invalidargument
New Contributor II
  • 0 kudos

Unfortunaly %run does not help me since I can't %run a .py file. I still need my code in notebooks.I am transpiling propriatary code to python using jinja templates. I would like to have the output as notebooks since those are most convenient to edit...

  • 0 kudos
1 More Replies
SamGreene
by Contributor
  • 813 Views
  • 1 replies
  • 0 kudos

Resolved! DLT Pipeline Graph is not detecting dependencies

Hi,This is my first databricks project.  I am loading data from a UC external volume in ADLS into tables and then split one of the tables into two tables based on a column.  When I create a pipeline, the tables don't have any dependencies and this is...

  • 813 Views
  • 1 replies
  • 0 kudos
Latest Reply
SamGreene
Contributor
  • 0 kudos

While re-implementing my pipeline to publish to dev/test/prod instead of bronze/silver/gold, I think I found the answer.  The downstream tables need to use the LIVE schema. 

  • 0 kudos
SamGreene
by Contributor
  • 871 Views
  • 2 replies
  • 1 kudos

Resolved! Unpivoting data in live tables

I am loading data from CSV into live tables.  I have a live delta table with data like this:WaterMeterID, ReadingDateTime1, ReadingValue1, ReadingDateTime2, ReadingValue2It needs to be unpivoted into this:WaterMeterID, ReadingDateTime1, ReadingValue1...

  • 871 Views
  • 2 replies
  • 1 kudos
Latest Reply
Kaniz
Community Manager
  • 1 kudos

Hi @SamGreene, The stack function allows you to unpivot columns by rotating their values into rows. It’s available both in Scala and PySpark.

  • 1 kudos
1 More Replies
stackoftuts
by New Contributor
  • 263 Views
  • 0 replies
  • 0 kudos

AI uses

Delve into the transformative realm of AI applications, where innovation merges seamlessly with technology's limitless possibilities.Explore the multifaceted landscape of AI uses and its dynamic impact on diverse industries at StackOfTuts. 

  • 263 Views
  • 0 replies
  • 0 kudos
Kroy
by Contributor
  • 575 Views
  • 2 replies
  • 0 kudos

Resolved! Multi Customer setup

We are trying to do POC to have shared resource like compute across multiple customer, Storage will be different, Is this possible ?    

Kroy_0-1702375949921.png
  • 575 Views
  • 2 replies
  • 0 kudos
Latest Reply
Kaniz
Community Manager
  • 0 kudos

Hi @Kroy , When it comes to shared compute resources in Databricks, there are some best practices and options you can consider:   Shared Access Mode for Clusters: Databricks allows you to create clusters in shared access mode. This means that multipl...

  • 0 kudos
1 More Replies
patojo94
by New Contributor II
  • 1265 Views
  • 2 replies
  • 3 kudos

Resolved! Stream failure JsonParseException

Hi all! I am having the following issue with a couple of pyspark streams. I have some notebooks running each of them an independent file structured streaming using  delta bronze table  (gzip parquet files) dumped from kinesis to S3 in a previous job....

Community Discussions
Photon
streaming aggregations
  • 1265 Views
  • 2 replies
  • 3 kudos
Latest Reply
Kaniz
Community Manager
  • 3 kudos

Hi @patojo94, You're encountering an issue with malformed records in your PySpark streams.    Let's explore some potential solutions:   Malformed Record Handling: The error message indicates that there are malformed records during parsing. By default...

  • 3 kudos
1 More Replies
Jay_adb
by New Contributor
  • 674 Views
  • 1 replies
  • 0 kudos

Resolved! Databricks Certification Exam Got Suspended. Need help in resolving the issue

Hi @Cert-Team ,My Databricks exam got suspended on December 9, 2023, at 11:30, and it is still in the suspended state.During the exam, it was initially paused due to poor lighting, but after addressing that, it worked fine. However, after some time, ...

  • 674 Views
  • 1 replies
  • 0 kudos
Latest Reply
Cert-Team
Honored Contributor III
  • 0 kudos

Hi @Jay_adb I'm sorry to hear you had this issue. Thanks for filing a ticket with the support team. I have sent a message to them to look into your ticket and resolve asap.

  • 0 kudos
JordanYaker
by Contributor
  • 451 Views
  • 0 replies
  • 0 kudos

DAB "bundle deploy" Dry Run

Is there a way to perform a dry-run with "bundle deploy" in order to see the job configuration changes for an environment without actually deploying the changes?

  • 451 Views
  • 0 replies
  • 0 kudos
Sujitha
by Community Manager
  • 680 Views
  • 0 replies
  • 1 kudos

🌟 End-of-Year Community Survey 🌟

Hello Community Members,  We value your experience and want to make it even better! Help us shape the future by sharing your thoughts through our quick Survey. Ready to have your voice heard? Click   here  and take a few moments to complete the surv...

Screenshot 2023-12-11 at 2.38.17 PM.png
  • 680 Views
  • 0 replies
  • 1 kudos
DBEnthusiast
by New Contributor III
  • 798 Views
  • 3 replies
  • 1 kudos

Resolved! More than expected number of Jobs created in Databricks

Hi Databricks Gurus !I am trying to run a very simple snippet :data_emp=[["1","sarvan","1"],["2","John","2"],["3","Jose","1"]]emp_columns=["EmpId","Name","Dept"]df=spark.createDataFrame(data=data_emp, schema=emp_columns)df.show() --------Based on a g...

  • 798 Views
  • 3 replies
  • 1 kudos
Latest Reply
Kaniz
Community Manager
  • 1 kudos

I want to express my gratitude for your effort in selecting the most suitable solution. It's great to hear that your query has been successfully resolved. Thank you for your contribution.

  • 1 kudos
2 More Replies
Soma
by Valued Contributor
  • 287 Views
  • 1 replies
  • 0 kudos

df.queryExecution.redactedSql is not working with Spark sql Listener

We are trying to capture the query executed by spark .We are trying to use df.queryExecution.redactedSql to get the SQL from query execution but it is not working in sqlListener

  • 287 Views
  • 1 replies
  • 0 kudos
Latest Reply
Kaniz
Community Manager
  • 0 kudos

Hi @Soma, In PySpark, when you execute a query and want to capture the SQL from the query execution, you can use the explain() method. 

  • 0 kudos