- 807 Views
- 0 replies
- 0 kudos
Hi everybody, I'm trying to install on a cluster (Azure Databricks, DBR 7.3LTS) texlive-full using apt-get install texlive-full in an init script.The issue is that, most of the times (not always), I get a 404 when downloading packages from security.u...
- 807 Views
- 0 replies
- 0 kudos
by
aimas
• New Contributor III
- 4033 Views
- 8 replies
- 5 kudos
Hi, i try to create a table using UI, but i keep getting the error "error creating table <table name> create a cluster first" even when i have a cluster alread running. what is the problem?
- 4033 Views
- 8 replies
- 5 kudos
Latest Reply
Be sure that cluster is selected (arrow in database) and at least there is Default database.
7 More Replies
- 3150 Views
- 5 replies
- 4 kudos
I'm trying to execute this writeStream data_frame.writeStream.format("delta") \
.option("checkpointLocation", checkpoint_path) \
.trigger(processingTime="1 second") \
.option("mergeSchema", "true") \
.o...
- 3150 Views
- 5 replies
- 4 kudos
Latest Reply
You can remove that folder so it will be recreated automatically.Additionally every new job run should have new (or just empty) checkpoint location.You can add in your code before running streaming:dbutils.fs.rm(checkpoint_path, True)Additionally you...
4 More Replies
- 3699 Views
- 9 replies
- 3 kudos
Hi team, I'm getting weird error in one of my jobs when connecting to Snowflake. All my other jobs (I've got plenty) work fine. The current one also works fine when I have only one coding step (except installing needed libraries in my very first step...
- 3699 Views
- 9 replies
- 3 kudos
Latest Reply
Dan_Z
Honored Contributor
@marchello​ I suggest you contact Snowflake to move forward on this one.
8 More Replies
- 3222 Views
- 3 replies
- 3 kudos
I’m trying to use sql query on azure-databricks with distinct sort and aliasesSELECT DISTINCT album.ArtistId AS my_alias
FROM album ORDER BY album.ArtistIdThe problem is that if I add an alias then I can not use not aliased name in the order by ...
- 3222 Views
- 3 replies
- 3 kudos
Latest Reply
The code from above is worked in both cases.
2 More Replies
- 4822 Views
- 17 replies
- 18 kudos
Please help resolve this error :org.neo4j.driver.exceptions.SecurityException: Failed to establish secured connection with the serverThis occurs when I want to establish a connection to neo4j aura to my cluster .Thank you.
- 4822 Views
- 17 replies
- 18 kudos
Latest Reply
@Werner Stinckens​ and @Nyaribo Maseru​ - You two are awesome! Thank you for working so hard together.
16 More Replies
- 732 Views
- 1 replies
- 0 kudos
Hi, I noticed unexpected behavior for Date type. If year value is less then 1000 then filtering do not work.
Steps:create table test (date Date); insert into test values ('0001-01-01'); select * from test where date = '0001-01-01'
Returns 0 rows....
- 732 Views
- 1 replies
- 0 kudos
- 2232 Views
- 1 replies
- 0 kudos
This means that no trial completed successfully. This almost always means that there is a bug in the objective function, and every invocation is resulting in an error. See the error output in the logs for details. In Databricks, the underlying error ...
- 2232 Views
- 1 replies
- 0 kudos
Latest Reply
The fmin function should be of the form:def evaluate_hyperparams(params):
"""
This method will be passed to `hyperopt.fmin()`. It fits and evaluates the model using the given hyperparameters to get the validation loss.
:param params: This d...
- 16355 Views
- 17 replies
- 3 kudos
This is an expensive and long-running job that gets about halfway done before failing. The stack trace is included below, but here is the salient part:
Caused by: org.apache.spark.SparkException: Job aborted due to stage failure: Task 4881 in stage...
- 16355 Views
- 17 replies
- 3 kudos
Latest Reply
According to https://docs.databricks.com/jobs.html#jar-job-tips:"Job output, such as log output emitted to stdout, is subject to a 20MB size limit. If the total output has a larger size, the run will be canceled and marked as failed."That was my prob...
16 More Replies
- 19940 Views
- 1 replies
- 1 kudos
Code : from pyspark.sql.functions import *acDF = spark.read.format('csv').options(header='true', inferschema='true').load("/mnt/rahulmnt/Insurance_Info1.csv"); acDF.write.option("overwriteSchema", "true").format("delta").mode("overwrite").save("/delt...
- 19940 Views
- 1 replies
- 1 kudos
Latest Reply
1. using Spark SQL Context in python, scala notebooks :
sql("SET spark.databricks.delta.formatCheck.enabled=false")
2. In SQL dbc notebooks:
SET spark.databricks.delta.formatCheck.enabled=false
- 4132 Views
- 3 replies
- 0 kudos
Hi everyone.
I am trying to learn the keyword OPTIMIZE from this blog using scala: https://docs.databricks.com/delta/optimizations/optimization-examples.html#delta-lake-on-databricks-optimizations-scala-notebook.
But my local spark seems not able t...
- 4132 Views
- 3 replies
- 0 kudos
Latest Reply
Hi Jigao,
OPTIMIZE isn't in the open source delta API, so won't run on your local Spark instance - https://docs.delta.io/latest/api/scala/io/delta/tables/index.html?search=optimize
2 More Replies