- 2256 Views
- 3 replies
- 0 kudos
In which format the Checkpoints are stored in storage and , how does it help in delta to increase performance.
- 2256 Views
- 3 replies
- 0 kudos
Latest Reply
Great points above on how checkpointing helps with performance. In additional Delta Lake also provides other data organization strategies such as compaction, Z-ordering to help with both read and write performance of Delta Tables. Additional details ...
2 More Replies
- 2531 Views
- 1 replies
- 0 kudos
It depends. If you specify the schema it will be zero, otherwise it will do a full file scan which doesn’t work well processing Big Data at a large scale.CSV files Dataframe Reader https://spark.apache.org/docs/latest/api/python/reference/api/pyspark...
- 2531 Views
- 1 replies
- 0 kudos
Latest Reply
As indicated there are ways to manage the amount of data being sampled for inferring schema. However as a best practice for production workloads its always best to define the schema explicitly for consistency, repeatability and robustness of the pipe...
- 3057 Views
- 2 replies
- 0 kudos
Is it possible to embed Databricks SQL Dashboards or specific widgets/visualization into a webpage?
- 3057 Views
- 2 replies
- 0 kudos
Latest Reply
Databricks SQL also integrates with several popular BI tools over JDBC/ODBC which you can use as a mechanism to embed visualizations into a webpage
1 More Replies
- 920 Views
- 1 replies
- 1 kudos
I want to test out different APIs directly from a Databricks notebook instead of using Postman or CURL. Is this possible?
- 920 Views
- 1 replies
- 1 kudos
Latest Reply
If you're question is about using the Databricks API from within a databricks notebook, then the answer is yes of course, you can definitely orchestrate anything and invoke the REST API from a python notebook using the `requests` library already bake...
- 725 Views
- 1 replies
- 0 kudos
I am trying to migrate my workload to another workspace ( from ST to E2), I am planning to use data bricks sync, but still I am not sure, will it migrate everything like , currents, user , groups, job, notebook etc or has some limitations which I s...
- 725 Views
- 1 replies
- 0 kudos
Latest Reply
Here is the support matrix for import/export operations for databricks-syncAlso checkout https://github.com/databrickslabs/migrate
- 621 Views
- 1 replies
- 0 kudos
I want to know how databricks maintain data recency in databricks
- 621 Views
- 1 replies
- 0 kudos
Latest Reply
When using delta tables in databricks, you have the advantage of delta cache which accelerates data reads by creating copies of remote files in nodes’ local storage using a fast intermediate data format. At the beginning of each query delta tables au...