cancel
Showing results for 
Search instead for 
Did you mean: 
Community Articles
Dive into a collaborative space where members like YOU can exchange knowledge, tips, and best practices. Join the conversation today and unlock a wealth of collective wisdom to enhance your experience and drive success.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

alysson_souza
by Databricks Employee
  • 8154 Views
  • 1 replies
  • 5 kudos

Configuring DNS resolution for Private Databricks Workspaces (AWS)

Intro For customers on the E2 Platform, Databricks has a feature that allows them to use AWS PrivateLink to provision secure private workspaces by creating VPC endpoints to both the front-end and back-end interfaces of the Databricks infrastructure. ...

AWS Inbound DNS Endpoints for Workspaces - Copy of Page 1.png alysson_souza_1-1718062496966.png AWS Inbound DNS Endpoints for Workspaces - Copy of Page 1.png
  • 8154 Views
  • 1 replies
  • 5 kudos
Latest Reply
Sujitha
Databricks Employee
  • 5 kudos

@alysson_souza love it! Thank you for sharing

  • 5 kudos
SashankKotta
by Databricks Employee
  • 4193 Views
  • 0 replies
  • 2 kudos

CICD for databricks workflow jobs

This post is to set up Databricks workflow jobs as a CI/CD. Below are the two essential components needed for a complete CI/CD setup of workflow jobs. Databricks Asset Bundles(DABs)AzureDevOps pipeline. Databricks Asset Bundle ( From local terminal )...

Screenshot 2024-06-14 at 4.27.27 PM.png Screenshot 2024-06-14 at 3.59.46 PM.png Screenshot 2024-06-14 at 4.19.59 PM.png
  • 4193 Views
  • 0 replies
  • 2 kudos
RamkannanA
by New Contributor II
  • 3543 Views
  • 1 replies
  • 3 kudos

Resolved! RamK - Certification Update

Hi Team,My name is Ram based out of Singapore. I am new to this Community . Recently I have completed my certification in Databricks starting from Data Analyst , Data Engineering and Gen AI. Looking forward to get connected in serving the Data and AI...

RamkannanA_1-1718261550063.png
  • 3543 Views
  • 1 replies
  • 3 kudos
Latest Reply
Rishabh-Pandey
Databricks MVP
  • 3 kudos

Happy to see you here .

  • 3 kudos
sudhirgarg
by New Contributor II
  • 2499 Views
  • 0 replies
  • 1 kudos

Free Databricks Professional Data Engineer Practice Tests

Hi All,I came across a very good set of Practice tests on Databricks Professional Data Engineer Certification.For time being It is being given for free by instructor as promotional activity . Enroll if you are planning to go for the certificationhttp...

  • 2499 Views
  • 0 replies
  • 1 kudos
NandiniN
by Databricks Employee
  • 1831 Views
  • 0 replies
  • 1 kudos

How to deal with Slow Jobs?

Definitely configure job timeouts, and configure notifications. This will help you to identify slowness due to various factors. It is crucial to also investigate and fix the issue causing the slowness. The first step is to identify the problem. This ...

  • 1831 Views
  • 0 replies
  • 1 kudos
NandiniN
by Databricks Employee
  • 1719 Views
  • 0 replies
  • 0 kudos

Monitoring a Streaming Job

If you have a streaming job, you need to check the batch metrics to be able to understand the stream progress. However, here are some other suggestions which we can use to monitor a streaming job and be stuck in a "hung" state. Streaming Listeners sp...

  • 1719 Views
  • 0 replies
  • 0 kudos
NandiniN
by Databricks Employee
  • 1914 Views
  • 0 replies
  • 0 kudos

Why configure a job timeout?

If you use Databricks Jobs for your workloads, it is possible you might have run into a situation where you find your jobs to be in "hung" state. Before cancelling the job it is important to collect the thread dump as I described here to be able to f...

  • 1914 Views
  • 0 replies
  • 0 kudos
MichTalebzadeh
by Valued Contributor
  • 2227 Views
  • 1 replies
  • 0 kudos

A handy tool called spark-column-analyser

I just wanted to share a tool I built called spark-column-analyzer. It's a Python package that helps you dig into your Spark DataFrames with ease.Ever spend ages figuring out what's going on in your columns? Like, how many null values are there, or h...

Community Articles
Generative AI
python
spark
  • 2227 Views
  • 1 replies
  • 0 kudos
Latest Reply
MichTalebzadeh
Valued Contributor
  • 0 kudos

An example added to README in GitHubDoing analysis for column PostcodeJson formatted output{"Postcode": {"exists": true,"num_rows": 93348,"data_type": "string","null_count": 21921,"null_percentage": 23.48,"distinct_count": 38726,"distinct_percentage"...

  • 0 kudos
youssefmrini
by Databricks Employee
  • 1615 Views
  • 0 replies
  • 2 kudos

Schema evolution clause added to SQL merge syntax

You can now add the WITH SCHEMA EVOLUTION clause to a SQL merge statement to enable schema evolution for the operation. For more information: https://docs.databricks.com/en/delta/update-schema.html#sql-evo   #Databricks

  • 1615 Views
  • 0 replies
  • 2 kudos
Hubert-Dudek
by Databricks MVP
  • 1470 Views
  • 0 replies
  • 2 kudos

VariantType + Parse_json()

In Spark 4.0, there are no more data type mismatches when converting dynamic JSONs, as the new data type VariantType comes with a new function to parse JSONs. Stay tuned for 4.0 release.

variant.png
  • 1470 Views
  • 0 replies
  • 2 kudos
youssefmrini
by Databricks Employee
  • 2201 Views
  • 0 replies
  • 1 kudos

Type widening is in Public Preview

You can now enable type widening on tables backed by Delta Lake. Tables with type widening enabled allow changing the type of columns to a wider data type without rewriting underlying data files.   For more information:https://docs.databricks.co...

  • 2201 Views
  • 0 replies
  • 1 kudos
Yassine_bens
by New Contributor
  • 1857 Views
  • 1 replies
  • 0 kudos

How to convert txt files to delta tables

Hello members of Databricks's comunity,I am currently working on a project where we collect data from machines, that data is in .txt format. The data is currently in an Azure container, I need to clean the files and convert them to delta tables, how ...

  • 1857 Views
  • 1 replies
  • 0 kudos
Latest Reply
feiyun0112
Honored Contributor
  • 0 kudos

https://docs.databricks.com/en/ingestion/add-data/upload-data.html 

  • 0 kudos
Hubert-Dudek
by Databricks MVP
  • 908 Views
  • 0 replies
  • 0 kudos

RocksDB for storing state stream

Now, you can keep the state of stateful streaming in RocksDB. For example, retrieving keys from memory to check for duplicate records inside the watermark is now faster. #databricks

state2.png
  • 908 Views
  • 0 replies
  • 0 kudos
legobricks
by New Contributor II
  • 2538 Views
  • 4 replies
  • 0 kudos

Unable to mount GCS bucket with underscores in the name

I have two buckets with the same configurations and labels.One is named my-bucket and the other is my_bucket. I am able to mount my-bucket but get an opaque error message when trying to mount my_bucket. Is this known/expected behavior? Are underscore...

  • 2538 Views
  • 4 replies
  • 0 kudos
Latest Reply
NandiniN
Databricks Employee
  • 0 kudos

Hi @legobricks , Curious on the error that you are getting. However, for GCS - https://cloud.google.com/storage/docs/buckets#naming I do see underscores are allowed but there is also a note below: You can use a bucket name in a DNS record as part of ...

  • 0 kudos
3 More Replies

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels