cancel
Showing results for 
Search instead for 
Did you mean: 
Get Started Discussions
Start your journey with Databricks by joining discussions on getting started guides, tutorials, and introductory topics. Connect with beginners and experts alike to kickstart your Databricks experience.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

aerofish
by New Contributor III
  • 4088 Views
  • 3 replies
  • 1 kudos

drop duplicates within watermark

Recently we are using structured streaming to ingest data. We want to use watermark to drop duplicated event. But We encountered some wired behavior and unexpected exception. Anyone can help me to explain what is the expected behavior and how should ...

  • 4088 Views
  • 3 replies
  • 1 kudos
Latest Reply
aerofish
New Contributor III
  • 1 kudos

Any maintainer can help me on this question??

  • 1 kudos
2 More Replies
bigt23
by New Contributor II
  • 6032 Views
  • 2 replies
  • 1 kudos

Resolved! Read zstd file from Databricks

I just started to read `zstd` compressed file in Databricks on Azure, Runtime 14.1 on Spark 3.5.0I've set PySpark commands as followspath = f"wasbs://{container}@{storageaccount}.blob.core.windows.net/test-zstd" schema = "some schema" df = spark.read...

  • 6032 Views
  • 2 replies
  • 1 kudos
Latest Reply
-werners-
Esteemed Contributor III
  • 1 kudos

The available compression types are format dependent.For json, zstd is not (yet) available, whereas for parquet it is.

  • 1 kudos
1 More Replies
floringrigoriu
by New Contributor II
  • 3364 Views
  • 0 replies
  • 0 kudos

Can Error Message be un Redacted

I there a way to un-redact the logging of error message ?Alternatively would be nice to have access to the source code of involved classes like : com.databricks.backend.common.util.CommandLineHelper or com.databricks.util.UntrustedUtils I'm getting t...

  • 3364 Views
  • 0 replies
  • 0 kudos
Abhiqa
by New Contributor II
  • 5705 Views
  • 1 replies
  • 1 kudos

How to schedule/refresh databricks alerts using REST API?

Hi, I am deploying Databricks SQL alerts using REST API. But I can't seem to figure out how to schedule their refresh task.I went through the documentation it says "Alerts can be scheduled using the sql_task type of the Jobs API, e.g. Jobs/Create"How...

Abhiqa_0-1697550139434.png Abhiqa_1-1697550638337.png
Get Started Discussions
Alerts
REST API
sql query
sql_task
  • 5705 Views
  • 1 replies
  • 1 kudos
Latest Reply
btafur
Databricks Employee
  • 1 kudos

What they mention in the API docs is that you can create a job with sql_task of type Alert. To make it easier you can try creating the job first in the UI first and downloading the JSON config. Here is an example with the main parameters that should ...

  • 1 kudos
bfrank1972
by New Contributor III
  • 1116 Views
  • 0 replies
  • 0 kudos

Small files and discrepancy in S3 vs catalog

Hello all,I'm in the process of optimizing my tables and I'm running into a confusing situation. I have a table named "trace_messages_fg_streaming_event". If I navigate to the Databricks catalog, it shows stats:Size: 6.7GB, Files: 464But when I look ...

bfrank1972_0-1697559008309.png
  • 1116 Views
  • 0 replies
  • 0 kudos
naga_databricks
by Contributor
  • 12028 Views
  • 1 replies
  • 1 kudos

Shared access vs Single user access mode

I am running a notebook to get secret value from GCP Secret Manager. This is working well with Single user Access Mode, however it fail when i use a cluster with Shared Access mode. I have specified the same GCP service account on both of these clust...

  • 12028 Views
  • 1 replies
  • 1 kudos
Latest Reply
naga_databricks
Contributor
  • 1 kudos

Thanks for your response.I am using a cloud service account (same account that was used to create the workspace) on the cluster properties in case of both the single user cluster and on the shared user cluster. This service account has all the necess...

  • 1 kudos
alesventus
by Contributor
  • 17381 Views
  • 6 replies
  • 0 kudos

Specify bottleneck for databricks cluster

Hi, Im trying to find out what is bottleneck on cluster when running loading process.Scenario: Loading CDC changes from sql server to Raw zone and merge changes into Bronze zone and then merge Bronze to Silver. All is orchestrated in data factory as ...

  • 17381 Views
  • 6 replies
  • 0 kudos
Latest Reply
Debayan
Databricks Employee
  • 0 kudos

stdout and stderr looks okay, do you have the log4j to share? You can make a doc out of it and share the doc here. 

  • 0 kudos
5 More Replies
AbhiJ
by New Contributor III
  • 30186 Views
  • 3 replies
  • 3 kudos

Facing Issues with Databricks JDBC Connectivity after Idle time

Hello team, I am using commons(commons-dbcp2) Datasource which supports default connection pooling in Spring Java application (rest services to fetch databricks data via JDBC template).Initially all works fine and can get the data from databricks via...

  • 30186 Views
  • 3 replies
  • 3 kudos
Latest Reply
ash42
New Contributor II
  • 3 kudos

I am seeing the same issue with hikari. When a pooled connection is created then the databricks cluster is terminated (or restarted), the HikariDataSource retains a stale session handle.Why does connection.isValid() returns true then executing any qu...

  • 3 kudos
2 More Replies
smehta_0908
by New Contributor II
  • 4796 Views
  • 2 replies
  • 0 kudos

Resolved! Unable to edit Catalog Owner

I created a Catalog and ownership was assigned to meI created databricks account-group on UC, added my user to this account-group, Assigned ownership of the catalog to this account-group.I deleted the account-groupNow, the catalog ownership is showin...

  • 4796 Views
  • 2 replies
  • 0 kudos
Latest Reply
Debayan
Databricks Employee
  • 0 kudos

Hi, In addition to the previous message, you can refer to https://docs.databricks.com/en/data-governance/unity-catalog/manage-privileges/index.html#assign-a-metastore-admin to get more information on metastore and etc.

  • 0 kudos
1 More Replies
Data_Analytics1
by Contributor III
  • 3017 Views
  • 2 replies
  • 0 kudos

Merge version data files of Delta table

Hi,I am having one CDC enabled Delta table. In 256th version, table is having 50 data files. I want all to merge and create a single file. How can I merge all 50 data file and when I query for 256th version, I should get 1 data file? Is there any com...

  • 3017 Views
  • 2 replies
  • 0 kudos
Latest Reply
Debayan
Databricks Employee
  • 0 kudos

Hi, ae you talking about merging CSV files? https://community.databricks.com/t5/machine-learning/merge-12-csv-files-in-databricks/td-p/3551#:~:text=Use%20Union()%20method%20to,from%20the%20specified%20set%2Fs.

  • 0 kudos
1 More Replies
sai_sathya
by New Contributor III
  • 2224 Views
  • 1 replies
  • 0 kudos

validating record count at SQL server database tabbles with migrated azure data lake gen2

we are migrating out project from on-premise to azure , so on-premise database is the SQL server that we are using and azure data lake gen2 is the storage location where store data currently and so far we are currently validating record count of each...

  • 2224 Views
  • 1 replies
  • 0 kudos
Ankita1
by New Contributor
  • 1825 Views
  • 0 replies
  • 0 kudos

Deleting external table takes 8 hrs

Hi,I am trying to delete the data from the external partitioned table, it has around 3 years of data, and the partition is created on the date column.I am trying to delete each partition first and then the schema of the table, which takes around 8hrs...

  • 1825 Views
  • 0 replies
  • 0 kudos
THIAM_HUATTAN
by Valued Contributor
  • 2456 Views
  • 0 replies
  • 0 kudos

why the code breaks below?

from pyspark.sql import SparkSessionfrom pyspark.ml.regression import LinearRegressionfrom pyspark.ml.feature import VectorAssemblerfrom pyspark.ml.evaluation import RegressionEvaluatorfrom pyspark.ml import Pipelineimport numpy as np# Create a Spark...

  • 2456 Views
  • 0 replies
  • 0 kudos
Isolated
by New Contributor
  • 2046 Views
  • 2 replies
  • 0 kudos

Having trouble with ARC (Automated Record Connector) Python Notebook

I'm trying to use Databricks ARC (Automated Record Connector) and running into an object issue. I assume I'm missing something rather trivial that's not related to ARC. #Databricks Python notebook #CMD1 import AutoLinker from arc.autolinker import A...

  • 2046 Views
  • 2 replies
  • 0 kudos
Latest Reply
Debayan
Databricks Employee
  • 0 kudos

https://www.databricks.com/blog/improving-public-sector-decision-making-simple-automated-record-linking and https://github.com/databricks-industry-solutions/auto-data-linkage#databricks-runtime-requirements

  • 0 kudos
1 More Replies
AkhilInala
by New Contributor II
  • 2759 Views
  • 1 replies
  • 0 kudos

Databricks Certified Data Engineer Associate Exam suspended

Initially, My exam was suspended after 20 mins on me into the exam. I stated that there was some internet issue from my end(however there wasn't any). However, I immediately reached out to Kryterion folks but they responded after 50 minutes and allow...

  • 2759 Views
  • 1 replies
  • 0 kudos
Latest Reply
AkhilInala
New Contributor II
  • 0 kudos

Adding my username - akhil.inala@tredence.comexam - Databricks Certified Data Engineer Associate Exam

  • 0 kudos

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels