cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 
Data + AI Summit 2024 - Data Engineering & Streaming

Forum Posts

Lon_Fortes
by New Contributor III
  • 5407 Views
  • 2 replies
  • 1 kudos

Resolved! How can I check that column on a delta table has a "NOT NULL" constraint or not?

Title pretty much says it all - I'm trying to determine whether or not a column on my existing delta table was defined as NOT NULL or not. It does not show up in any of the metadata (describe detail, describe history, show tblproperties). Thanks in...

  • 5407 Views
  • 2 replies
  • 1 kudos
Latest Reply
Matthew8
New Contributor II
  • 1 kudos

walgreenslistens Wrote:A UNIQUE constraint defines a set of columns that uniquely identify rows in a table only if all the key values are not NULL. If one or more key parts are NULL, duplicate keys are allowed.

  • 1 kudos
1 More Replies
User16857281869
by New Contributor II
  • 691 Views
  • 1 replies
  • 0 kudos

We want to do demand forecasting for our supply chain. How should we benefit from Spark in the Usercase development?

We have a series of blogs on the topic which describe the challenges and the best practices on development of demand forecasting usecases on Databricks. Please refer to this blog and the references in it for more info.

  • 691 Views
  • 1 replies
  • 0 kudos
Latest Reply
Anonymous
Not applicable
  • 0 kudos

We have a series of blogs on the topic which describe the challenges and the best practices on development of demand forecasting usecases on Databricks. Please refer to this blog and the references in it for more info.

  • 0 kudos
irfanaziz
by Contributor II
  • 1707 Views
  • 1 replies
  • 1 kudos

Resolved! How to keep the original Swedish/Finnish character in the file?

The files are in ANSI format as it shows in the notepad, I could manuelly convert the files to utf8 and read it. But the files are really large. I dont want to download and upload the files. Is there way so i could keep the swedish/finnish characte...

  • 1707 Views
  • 1 replies
  • 1 kudos
Latest Reply
irfanaziz
Contributor II
  • 1 kudos

So the answer was using the option("charset", "iso-8859-1")

  • 1 kudos
SarahDorich
by New Contributor II
  • 2734 Views
  • 3 replies
  • 0 kudos

How to register datasets for Detectron2

I'm trying to run a Detectron2 model in Databricks and cannot figure out how to register my train, val and test datasets. My datasets live in an Azure data lake. I have tried the following with no luck. Any help is appreciated. 1) Specifying full p...

  • 2734 Views
  • 3 replies
  • 0 kudos
Latest Reply
Thurman
New Contributor II
  • 0 kudos

Register your dataset Optionally, register metadata for your dataset.

  • 0 kudos
2 More Replies
hmcdowelle
by New Contributor II
  • 11718 Views
  • 18 replies
  • 0 kudos

i just can't seem to make a cluster without error

I have been trying to create a cluster and this is my first time using databricks. I have tried across multiple resources and am getting frustrated. Each time the cluster comes up with an error. I have no idea what I am doing wrong. I use default se...

  • 11718 Views
  • 18 replies
  • 0 kudos
Latest Reply
kwayebgh
New Contributor II
  • 0 kudos

I have a Free Azure student account and I was facing similar challenges. This is how I solved mine after many hours of trial and error. Mine is working now. When creating the Azure Databricks resource group: Select Premium. Don't use the 14 day tri...

  • 0 kudos
17 More Replies
user_b22ce5eeAl
by New Contributor II
  • 1348 Views
  • 2 replies
  • 0 kudos

pandas udf type grouped map fails

Hello, I am trying to get the shap values for my whole dataset using pandas udf for each category of a categorical variable. It runs well when I run it on a few categories but when I want to run the function on the whole dataset my job fails. I see ...

  • 1348 Views
  • 2 replies
  • 0 kudos
Latest Reply
Jackson
New Contributor II
  • 0 kudos

I want to use data.groupby.apply() to apply a function to each row of my Pyspark Dataframe per group.I used The Grouped Map Pandas UDFs. However I can't figure out how to add another argument to my function. DGCustomerFirst SurveyI tried using the ar...

  • 0 kudos
1 More Replies
StephanieAlba
by Valued Contributor III
  • 1634 Views
  • 1 replies
  • 0 kudos

Is it possible to turn off the redaction of secrets? Is there a better way to solve this?

As part of our Azure Data Factory pipeline, we utilize Databricks to run some scripts that identify which files we need to load from a certain source. This list of files is then passed back into Azure Data Factory utilizing the Exit status from the n...

  • 1634 Views
  • 1 replies
  • 0 kudos
Latest Reply
StephanieAlba
Valued Contributor III
  • 0 kudos

No, it is not possible to turn off redaction. No, there is not another way to return values from a notebook.1) Using a native Databricks feature such as Autoloader is suggested.2) They could write the list of files to be processed to a delta table an...

  • 0 kudos
guruv
by New Contributor III
  • 634 Views
  • 0 replies
  • 0 kudos

Transactional approach to write to Azure ADLS gen2 storage

Hi, what is the recommended way to read data from delta table and write to ADLS gen2 storage in parquet format. In my case i use a notebook to read data do some processing and write it to storage and update delta table with detail of last written da...

  • 634 Views
  • 0 replies
  • 0 kudos
User16826992666
by Valued Contributor
  • 1419 Views
  • 2 replies
  • 0 kudos

Resolved! Can I convert parquet files to Delta?

I am already storing my data as parquet files and have registered them as a table in Databricks. If I want to convert the table to be a Delta table, do I have to do a full read of the data and rewrite it in the Delta format?

  • 1419 Views
  • 2 replies
  • 0 kudos
Latest Reply
User16752244127
Contributor
  • 0 kudos

more details and programmatic options can be found in the Porting Guide.

  • 0 kudos
1 More Replies
MoJaMa
by Valued Contributor II
  • 1925 Views
  • 2 replies
  • 0 kudos
  • 1925 Views
  • 2 replies
  • 0 kudos
Latest Reply
User16752244127
Contributor
  • 0 kudos

Kinesis streams is the Kinesis streaming service. Select this!Kinesis Firehose reads data from a Kinesis stream and writes it e.g. to S3 or Redshift. or Splunk (more details here)

  • 0 kudos
1 More Replies
User16826994223
by Honored Contributor III
  • 1210 Views
  • 2 replies
  • 0 kudos

What is the differentiator between delta sharing and other cloud sharing platform

What is differentiator between delta sharing and other cloud sharing platform.

  • 1210 Views
  • 2 replies
  • 0 kudos
Latest Reply
User16752244127
Contributor
  • 0 kudos

Also, unlike other servers, Delta Sharing internally uses pre-signed URLs to S3, GCS, or ADSL, so data transfer from a client happens at the bandwidth of the underlying cloud object-store. This way the Delta Sharing server scales extremely well and d...

  • 0 kudos
1 More Replies
MatthewLau
by New Contributor
  • 737 Views
  • 0 replies
  • 0 kudos

Logging Lifetime Plot_history_alive as a model

Hi Databricks Community, I have followed the CLV Databricks accelator (https://databricks.com/notebooks/CLV_Part_1_Customer_Lifetimes.html) to do an initial CLV analysis. Thank you for sharing this with the community. My question is that in the note...

0693f000007OoRXAA0
  • 737 Views
  • 0 replies
  • 0 kudos

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.

If there isn’t a group near you, start one and help create a community that brings people together.

Request a New Group
Labels