cancel
Showing results for 
Search instead for 
Did you mean: 
Get Started Discussions
Start your journey with Databricks by joining discussions on getting started guides, tutorials, and introductory topics. Connect with beginners and experts alike to kickstart your Databricks experience.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

patojo94
by New Contributor II
  • 4104 Views
  • 1 replies
  • 2 kudos

Stream failure JsonParseException

Hi all! I am having the following issue with a couple of pyspark streams. I have some notebooks running each of them an independent file structured streaming using  delta bronze table  (gzip parquet files) dumped from kinesis to S3 in a previous job....

Get Started Discussions
Photon
streaming aggregations
  • 4104 Views
  • 1 replies
  • 2 kudos
Xyrion
by New Contributor II
  • 2080 Views
  • 1 replies
  • 0 kudos

Constraint options usage

I am trying to use the constraints options:NOT ENFORCEDDEFERRABLEINITIALLY DEFERREDNORELYHowever it seems I am not able to use them successfully. When I try to use them with PRIMARY KEYS (not sure if it is possible), I am not able to enforce any key....

  • 2080 Views
  • 1 replies
  • 0 kudos
Latest Reply
Xyrion
New Contributor II
  • 0 kudos

BTW the forum is bugged I can't paste code..

  • 0 kudos
Jay_adb
by New Contributor
  • 2079 Views
  • 1 replies
  • 0 kudos

Resolved! Databricks Certification Exam Got Suspended. Need help in resolving the issue

Hi @Cert-Team ,My Databricks exam got suspended on December 9, 2023, at 11:30, and it is still in the suspended state.During the exam, it was initially paused due to poor lighting, but after addressing that, it worked fine. However, after some time, ...

  • 2079 Views
  • 1 replies
  • 0 kudos
Latest Reply
Cert-Team
Databricks Employee
  • 0 kudos

Hi @Jay_adb I'm sorry to hear you had this issue. Thanks for filing a ticket with the support team. I have sent a message to them to look into your ticket and resolve asap.

  • 0 kudos
JordanYaker
by Contributor
  • 1685 Views
  • 0 replies
  • 2 kudos

DAB "bundle deploy" Dry Run

Is there a way to perform a dry-run with "bundle deploy" in order to see the job configuration changes for an environment without actually deploying the changes?

  • 1685 Views
  • 0 replies
  • 2 kudos
Sujitha
by Databricks Employee
  • 8523 Views
  • 0 replies
  • 1 kudos

🌟 End-of-Year Community Survey 🌟

Hello Community Members,  We value your experience and want to make it even better! Help us shape the future by sharing your thoughts through our quick Survey. Ready to have your voice heard? Click   here  and take a few moments to complete the surv...

Screenshot 2023-12-11 at 2.38.17 PM.png
  • 8523 Views
  • 0 replies
  • 1 kudos
Borkadd
by New Contributor II
  • 2175 Views
  • 1 replies
  • 0 kudos

Multi Task Job creation through Pulumi

I am trying to create a multi-task Databricks Job in Azure Cloud with its own cluster.Although I was able to create a single task job without any issues, the code to deploy the multi-task job fails due to the following cluster validation error:error:...

  • 2175 Views
  • 1 replies
  • 0 kudos
Latest Reply
Borkadd
New Contributor II
  • 0 kudos

Hello @Retired_mod, thanks for your answer, but the problem keeps the same. I had already tested with different cluster configurations, single-node and multi-node, including those cluster configurations which worked with single task jobs, but the err...

  • 0 kudos
TechMG
by New Contributor II
  • 8822 Views
  • 0 replies
  • 0 kudos

power Bi paginate

Hello,I am facing similar kind of issue. I am working on Power BI paginated report and databricks is my source for the report. I was trying to pass the parameter by passing the query in expression builder as mentioned below. https://community.databri...

  • 8822 Views
  • 0 replies
  • 0 kudos
Aria
by New Contributor III
  • 10543 Views
  • 2 replies
  • 2 kudos

Databricks Asset bundle

Hi,I am new to databricks, We are trying to use Databricks asset bundles for code deployment .I have spect a lot of time but still so many things are not clear to me.Can we change the target path of the notebooks deployed from /shared/.bundle/* to so...

  • 10543 Views
  • 2 replies
  • 2 kudos
Latest Reply
Еmil
New Contributor III
  • 2 kudos

Hi @Retired_mod,Thank you for your post, I thought it will solve my issues too, however after reading your suggestion it was nothing new for me, because l already have done exactly that.. Here is what I have dome so you or anyone can replicate it:1. ...

  • 2 kudos
1 More Replies
Lazloo
by New Contributor III
  • 2351 Views
  • 1 replies
  • 0 kudos

Using nested dataframes with databricks-connect>13.x

 We needed to move to databricks-connect>13.x. Now I facing the issue that when I work with a nested dataframe of the structure```root|-- a: string (nullable = true)|-- b: array (nullable = true)| |-- element: struct (containsNull = true)| | |-- c: s...

  • 2351 Views
  • 1 replies
  • 0 kudos
Latest Reply
Lazloo
New Contributor III
  • 0 kudos

In addition here is the full stack trace23/12/07 14:51:56 ERROR SerializingExecutor: Exception while executing runnable grpc_shaded.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1MessagesAvailable@33dfd6ecgrpc_shaded.io.grpc...

  • 0 kudos
Phani1
by Valued Contributor II
  • 765 Views
  • 0 replies
  • 0 kudos

Auto Loader notebook for multiple tables

Hi Team,My requirement is ,i do have File A from source A which needs to write into Multiple Delta tables i.e DeltaTableA,DeltaTableB,DeltaTableC. Is it possible to have a single instance of an autoloader script. (multiple write streams). Could you p...

  • 765 Views
  • 0 replies
  • 0 kudos
mvmiller
by New Contributor III
  • 3111 Views
  • 1 replies
  • 0 kudos

How to facilitate incremental updates to an SCD Type 1 table that uses SCD Type 2 source tables

I have an SCD Type 1 delta table (target) for which I am trying to figure out how to facilitate insert, updates, and deletes.  This table is sourced by multiple delta tables, with an SCD Type 2 structure, which are joined together to create the targe...

  • 3111 Views
  • 1 replies
  • 0 kudos
Latest Reply
mvmiller
New Contributor III
  • 0 kudos

Correction (I can't seem to edit or remove original post):- "... trying to think through an process" --> *a* process- "Thoughts and advice or much appreciated" --> Thoughts and/or advice are much appreciated.

  • 0 kudos
Gembo
by New Contributor III
  • 4880 Views
  • 4 replies
  • 0 kudos

Cluster Access mode set to Shared on Databricks, results in connection refused on Exasol

I am trying to run a TRUNCATE command on my exasol DWH from Databricks using the pyexasol. This works perfectly fine when I have the cluster access mode as "No Isolation Shared" which does not have access to our Unity Catalog. When I change the clust...

  • 4880 Views
  • 4 replies
  • 0 kudos
Latest Reply
SSundaram
Contributor
  • 0 kudos

Interesting. Did you try with "Single User" mode, which also has UC support? 

  • 0 kudos
3 More Replies
margutie
by New Contributor
  • 1226 Views
  • 0 replies
  • 0 kudos

Error from Knime trought proxy

I want to connect to Databricks from Knime on a company computer that uses a proxy. The error I'm encountering is as follows: ERROR Create Databricks Environment 3:1 Execute failed: Could not open the client transport with JDBC URI: jdbc:hive2://adb-...

  • 1226 Views
  • 0 replies
  • 0 kudos
Sreekanth_N
by New Contributor II
  • 4185 Views
  • 2 replies
  • 0 kudos

'NotebookHandler' object has no attribute 'setContext' in pyspark streaming in AWS

I am facing issue while calling dbutils.notebook.run() inside of pyspark streaming with concurrent.executor. At first the error is "pyspark.sql.utils.IllegalArgumentException: Context not valid. If you are calling this outside the main thread,you mus...

  • 4185 Views
  • 2 replies
  • 0 kudos
Latest Reply
Kevin3
New Contributor III
  • 0 kudos

The error message you're encountering in PySpark when using dbutils.notebook.run() suggests that the context in which you are attempting to call the run() method is not valid. PySpark notebooks in Databricks have certain requirements when it comes to...

  • 0 kudos
1 More Replies

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels