Certifications
Join dynamic discussions on Databricks certifications within the Community. Exchange insights, tips,...
The forum board you are trying to access is permanently deleted.
Join dynamic discussions on Databricks certifications within the Community. Exchange insights, tips,...
Explore discussions on Databricks training programs and offerings within the Community. Get insights...
Start your journey with Databricks by joining discussions on getting started guides, tutorials, and ...
Engage in discussions about the Databricks Free Edition within the Databricks Community. Share insig...
How to make sure you will be sala Uddin Ahmed to deewane season of the same hai bhai send your number and the first to make it to you too baby
With managed tables, you can reduce your storage and compute costs thanks to predictive optimization or file list caching. Now it is really time to migrate external tables to managed ones, thanks to ALTER SET MANAGED functionality. Read more: - h...
Hi team,I'm currently diving into the Partnered Academy learning plan (e.g., Data Engineer Associate, ML Practitioner, etc.) and I'm looking to schedule my certification exam.I'm wondering about the official policy: Is it mandatory to complete the en...
From my experience everything is independent from each other
Something very strange has happened. When trying to login to my databricks community edition account. I'm getting the email with my verification code. but after entring that code, I'm getting the error message : "User is not a member of this workspac...
I'm also facing the same error on registering to community edition. After providing correct verification code, I'm getting error "User is not a member of this workspace". The url -> https://community.cloud.databricks.com/verify-code?tuuid=2cd7f036-38...
What are some best practices for optimizing Spark jobs in Databricks, especially when dealing large datasets? Any tips or resources would be greatly appreciated! I’m trying to analyze data on restaurant menu prices so that insights would be especiall...
Their menu is full of creative options — it’s hard to pick just one favorite because everything tastes amazing!”
I am looking for some help on getting databricks cluster metrics such as memory utilization, CPU utilization, memory swap utilization, free file system using REST API.I am trying it in postman using databricks token and with my Service Principal bear...
Is there any solution found to get cpu, memory metrics for Hive meta store backed workloads ? We are not using UC. So can't use system tables
Hi allI'm switching to evaluating Databricks after being unhappy with Microsoft Fabric.I'm looking for a platform to aggregate external files and tables, do processing and push into 3rd party using GraphQL. That push is going to be quite sophisticate...
hi @Dimitry , I'm not sure about this specific library, but I was able to query a graphQL endpoint with requests.post. Have you tested that at all? Happy to dig in deeper if you need to use this specific library for any reason, but if not - you may w...
After completing all the relevant courses for the certification, I haven’t received the coupon code yet.
Thank you for confirming and taking the time to clarify this. I really appreciate your help and guidance!
I noticed that unlike "Alter Table" there is no "Alter View" command to add comment on a column in the existing view. This is a regular view created on Tables (and not Materialized view). If the underlying table column has comment then the View inh...
Use COMMENT ONCOMMENT ON | Databricks on AWS
Hello AllAm getting the below error when trying to create ODBC DSN Simba 64 in local system to connect Databricks Server using the token and enabled SSL System trust store & Thrift Transport: HTTP.any helping hand really appreciated . [Simba][ThriftE...
Solved for my case. Still not sure why/how it was working on one server but not the other.Final fix was to add HTTPPath value to the Connection String I listed above.
I’m building a dashboard in Power BI’s Pro Workspace, connecting data via Direct Query from Databricks (around 60 million rows from 15 combined tables), using a SQL Serverless (small size and 4 clusters).The problem is that the dashboard is taking to...
@viniciuscini have you managed to get it working well for you?
If you were creating Unity Catalogs again, what would you do differently based on your past experience?
@nayan_wylde no don't do that hehe. It was example of extreme approach. Usually use catalog to separate environment + in enterprises to separate divisions like customer tower, marketing tower, finance tower etc
Let’s say we have big data application where data loss is not an option.Having GZRS (geo-zone-redundant storage) redundancy we would achieve zero data loss if primary region is alive – writer is waiting for acks from two or more Azure availability zo...
Databricks is working on improvements and new functionality related to that. For now, the only solution is a DEEP CLONE. You can run it more frequently or implement your own replication based on a change data feed. You could use delta sharing for tha...
How much time does it take to ingest two billion rows using the free databricks edition?https://www.databricks.com/blog/learn-experiment-and-build-databricks-free-edition
Thanks for sharing @Hubert-Dudek . Really cool idea to test performance of free edition
I am designing structured streaming job in Azure data bricks(using Scala) which will consume messages from two event hubs, lets call them source and target.I would like your feedback on below flow, whether it is will survive the production load and ...
It is hard to understand what the source is and what the target is. Some charts could be useful. Also, information on how long the state is kept. My solution usually is:- Use declarative lakeflow pipelines if possible (dlt) - if not, consider handlin...
User | Count |
---|---|
212 | |
174 | |
93 | |
75 | |
45 |