cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

amarnathpal
by New Contributor III
  • 1369 Views
  • 4 replies
  • 0 kudos

Adding a New Column for Updated Date in Pipeline

I've successfully set up my pipeline and everything is working fine. I'd like to add a new column to our table that records the date whenver any records got updated. Could you advise on how to go about this?

  • 1369 Views
  • 4 replies
  • 0 kudos
Latest Reply
nikhilj0421
Databricks Employee
  • 0 kudos

Do you want to add dates for the historical data as well?

  • 0 kudos
3 More Replies
Ramakrishnan83
by New Contributor III
  • 3701 Views
  • 2 replies
  • 0 kudos

Optimize and Vaccum Command

Hi team,I am running a weekly purge process from databricks notebooks that cleans up chunk of records from my tables used for audit purposes. Tables are external tables. I need clarification on below items1.Should I need to  run Optimize and Vacuum c...

  • 3701 Views
  • 2 replies
  • 0 kudos
Latest Reply
JaimeAnders
New Contributor II
  • 0 kudos

That's a valid point about minimal read queries! However, while immediate storage reduction might not be necessary, consistent data integrity and potential future reporting needs might still warrant occasional optimize and vacuuming, even with extern...

  • 0 kudos
1 More Replies
jeremy98
by Honored Contributor
  • 3186 Views
  • 6 replies
  • 2 kudos

Resolved! Catch Metadata Workflow databricks

Hello community,Is it possible to get metadata workflow of a databricks job that is running? Like the start time, end time, triggered by etc.? Using dbutils.widgets.get()?

  • 3186 Views
  • 6 replies
  • 2 kudos
Latest Reply
Juan_Cardona
Databricks Partner
  • 2 kudos

Now The best practice for this is not using the API (some functions were deprecated for this objective) instead you should use job parameters job_id = dbutils.widgets.get("job parameter name with job_id") job_run = dbutils.widgets.get("job parameter ...

  • 2 kudos
5 More Replies
Ankit_Kothiya
by Databricks Partner
  • 1556 Views
  • 2 replies
  • 1 kudos

Databricks JDBC Driver Version 42 Limitations

We found that the Databricks JDBC driver does not support:Connection.setAutoCommit(false)Connection.commit()Connection.rollback()Execution of BEGIN TRANSACTIONCan you help us understand why these operations are not supported by the Databricks JDBC dr...

  • 1556 Views
  • 2 replies
  • 1 kudos
Latest Reply
Ankit_Kothiya
Databricks Partner
  • 1 kudos

Thank you, @SP_6721 , for your input!Could you please share an example snippet demonstrating how to handle batch processing, similar to what we typically do in a relational database?

  • 1 kudos
1 More Replies
venkad
by Contributor
  • 14779 Views
  • 5 replies
  • 7 kudos

Passing proxy configurations with databricks-sql-connector python?

Hi,I am trying to connect to databricks workspace which has IP Access restriction enabled using databricks-sql-connector. Only my Proxy server IPs are added in the allow list.from databricks import sql   connection = sql.connect( server_hostname ='...

  • 14779 Views
  • 5 replies
  • 7 kudos
Latest Reply
ss2025
New Contributor II
  • 7 kudos

Is there any resolution for the above setting up proxy with databricks sql connector

  • 7 kudos
4 More Replies
Upendra_Dwivedi
by Databricks Partner
  • 3578 Views
  • 4 replies
  • 0 kudos

Resolved! How to enable Databricks Apps User Authorization?

Hi All,I am working on implementation of user authorization in my databricks app. but to enable user auth it is asking:"A workspace admin must enable this feature to be able to request additional scopes. The user's API downscoped access token is incl...

  • 3578 Views
  • 4 replies
  • 0 kudos
Latest Reply
Upendra_Dwivedi
Databricks Partner
  • 0 kudos

Hi All, We can find this setting under Previews. Go to workspace>click your username>Previews

  • 0 kudos
3 More Replies
Ipshi
by New Contributor
  • 1010 Views
  • 1 replies
  • 0 kudos

databricks Data Engineer associate

Hi everyone , can anyone guide me about any test papers or any test materials anyone can go through for the databricks data engineer associate exam 

  • 1010 Views
  • 1 replies
  • 0 kudos
Latest Reply
Advika
Community Manager
  • 0 kudos

Hello @Ipshi! You can find resources for the Databricks Certified Data Engineer Associate exam in the Getting Ready for the Exam section of the exam-specific webpage on the website. This section includes a detailed list of topics covered and sample q...

  • 0 kudos
lawrence009
by Contributor
  • 3197 Views
  • 4 replies
  • 0 kudos

Blank Page after Logging In

On Feb 8 Singapore time, our Singapore workspace displayed a blank page (no interface or content) after login. Meanwhile our workspace in Tokyo reason worked normally. This lasted whole day and none of our troubleshooting yielded any clues. Then ever...

  • 3197 Views
  • 4 replies
  • 0 kudos
Latest Reply
ciro
New Contributor II
  • 0 kudos

After logging in, I’m getting a white screen, and it won’t load. I’ve tried clearing my cache and switching browsers, but nothing seems to work. This feels like something that really needs to be looked into. Has anyone figured out a way to fix it?

  • 0 kudos
3 More Replies
pargit2
by New Contributor II
  • 937 Views
  • 1 replies
  • 0 kudos

feature store delta sharing

Hi, I have 2 workspaces one for data engineers and one for data science team and I need to create in data engineering workspace the bronze and silver.I want to built them a feature store should I do it from data science workspace or data engineering ...

  • 937 Views
  • 1 replies
  • 0 kudos
Latest Reply
ciro
New Contributor II
  • 0 kudos

I like the idea of using Feature Store with Delta Sharing, but I’m a bit worried about its limits like no partition filtering and no streaming support. These could cause problems with performance and scaling in real situations.

  • 0 kudos
thisisadarshsin
by New Contributor II
  • 7471 Views
  • 12 replies
  • 0 kudos

Permission issue in Fundamentals of the Databricks Lakehouse Platform Quiz

Hi ,I am getting this Error,when i am trying to give the exam ofFundamentals of the Databricks Lakehouse Platform.403FORBIDDENYou don't have permission to access this page2023-05-20 12:37:41 | Error 403 | https://customer-academy.databricks.com/I al...

  • 7471 Views
  • 12 replies
  • 0 kudos
Latest Reply
Advika_
Databricks Employee
  • 0 kudos

Hello, everyone! We are sorry to hear you're having trouble accessing the Quiz. Please note that the Lakehouse Fundamentals course has been replaced by Databricks Fundamentals along with the updated content.Try logging into your account directly by u...

  • 0 kudos
11 More Replies
NathanC0926
by New Contributor
  • 1326 Views
  • 1 replies
  • 0 kudos

Delta Live Table (Streaming Tables) for excel (.xlsx, .xls)

What's the native way to ingest excel files using a streaming table? I wish that when the excel files land in unity catalog, it can pick up those and load it in to the Streaming Table. Data is Small, so we can afford some kind of UDF, but we really n...

  • 1326 Views
  • 1 replies
  • 0 kudos
Latest Reply
lingareddy_Alva
Esteemed Contributor
  • 0 kudos

Hi @NathanC0926 Ingesting Excel files with streaming tables requires a combination of Databricks Autoloader(for file discovery and exactly-once processing) and a custom UDF for Excel parsing.Here's the native approachKey Features of This Solution1. E...

  • 0 kudos
lumen
by New Contributor II
  • 1485 Views
  • 3 replies
  • 1 kudos

Notebook ID in DESCRIBE HISTORY not showing

We've recently installed Databricks 14.3 LTS with Unity Catalog and for some reason that is escaping me the Notebook Id is not showing up when I execute the DESCRIBE HISTORY SQL command. Example below for table test_catalog.lineagedemo.lm_lineage_tes...

Image 23-05-2025 at 14.18.png
  • 1485 Views
  • 3 replies
  • 1 kudos
Latest Reply
lumen
New Contributor II
  • 1 kudos

Hi @RameshRetnasamy first off thank you so much for taking the time to reply to my question. In my case they were indeed created via Notebooks, but I'll re-evaluate on my end as I might've missed something. If the issue persists, I'll re-assert the q...

  • 1 kudos
2 More Replies
Vinoth_nirmal
by New Contributor II
  • 1407 Views
  • 4 replies
  • 0 kudos

Not able create and start a cluster

Hi Team,I am trying to use community edition for learning below is my URL detailshttps://community.cloud.databricks.com/compute/interactive?o=2059657917292434Due to some reason my clusters are taking nearly 45 to 60 minutes for creation and after if ...

  • 1407 Views
  • 4 replies
  • 0 kudos
Latest Reply
Vinoth_nirmal
New Contributor II
  • 0 kudos

Hi @nikhilj0421  still the same issue, i could see its working till spark 15.4 LTS above 15.4LTS whatever version i useits not woring i am not able to create any cluster.

  • 0 kudos
3 More Replies
hao-uit
by New Contributor
  • 3207 Views
  • 1 replies
  • 0 kudos

Spark Streaming Job gets stuck in the "Stream Initializing"

Hello all,I am having an issue with my Spark Streaming Job. It is stuck at "Stream Initializing" stage.Need your help here to understand what is happening inside the "Stream Initializing" stage of Spark Streaming job which is taking so long. Here are...

  • 3207 Views
  • 1 replies
  • 0 kudos
Latest Reply
nikhilj0421
Databricks Employee
  • 0 kudos

Hi @hao-uit, do you see any kind of load on the driver and event logs? Also, what libraries you have installed on your cluster? 

  • 0 kudos
dev_puli
by New Contributor III
  • 56245 Views
  • 7 replies
  • 8 kudos

how to read the CSV file from users workspace

Hi!I have been carrying out a POC, so I created the CSV file in my workspace and tried to read the content using the techniques below in a Python notebook, but did not work.Option1:repo_file = "/Workspace/Users/u1@org.com/csv files/f1.csv"tmp_file_na...

  • 56245 Views
  • 7 replies
  • 8 kudos
Latest Reply
MujtabaNoori
New Contributor III
  • 8 kudos

Hi @Dev ,Generally, What happens spark reader APIs point to the DBFS by default. And, to read the file from User workspace, we need to append 'file:/' in the prefix.Thanks

  • 8 kudos
6 More Replies
Labels