cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

brickster_2018
by Databricks Employee
  • 3032 Views
  • 1 replies
  • 1 kudos
  • 3032 Views
  • 1 replies
  • 1 kudos
Latest Reply
sajith_appukutt
Honored Contributor II
  • 1 kudos

 G1GC can solve problems in some cases where garbage collection is a bottleneck. checkout https://databricks.com/blog/2015/05/28/tuning-java-garbage-collection-for-spark-applications.html

  • 1 kudos
User16790091296
by Contributor II
  • 1377 Views
  • 1 replies
  • 0 kudos
  • 1377 Views
  • 1 replies
  • 0 kudos
Latest Reply
Taha
Databricks Employee
  • 0 kudos

As of this comment, SQL analytics still requires a few additional enablement steps. You will need to ask your Databricks account team to help turn this on in your workspace.

  • 0 kudos
brickster_2018
by Databricks Employee
  • 1807 Views
  • 1 replies
  • 0 kudos
  • 1807 Views
  • 1 replies
  • 0 kudos
Latest Reply
aladda
Databricks Employee
  • 0 kudos

You could potentially do this through a Global Init Script - https://docs.databricks.com/clusters/init-scripts.html

  • 0 kudos
User16790091296
by Contributor II
  • 3741 Views
  • 3 replies
  • 0 kudos
  • 3741 Views
  • 3 replies
  • 0 kudos
Latest Reply
Mooune_DBU
Valued Contributor
  • 0 kudos

Dy doing a `GET` call using the cluster idcurl --netrc -X GET \ https://dbc-a1b2345c-d6e7.cloud.databricks.com/api/2.0/clusters/get \ --data '{ "cluster_id": "1234-567890-myclustID" }' \ | jq .The response json will have a `state` tag which will look...

  • 0 kudos
2 More Replies
User16826992666
by Valued Contributor
  • 2414 Views
  • 1 replies
  • 0 kudos

Can I move some partitions of a Delta table to a different location?

I am partitioning my Delta table by date. Older data is rarely accessed, so I am wondering if I can move some of the files off to colder storage options. What would happen if I did this? Is this a supported pattern or would it break the table?

  • 2414 Views
  • 1 replies
  • 0 kudos
Latest Reply
sajith_appukutt
Honored Contributor II
  • 0 kudos

You could look at S3 Intelligent-Tiering - https://aws.amazon.com/about-aws/whats-new/2018/11/s3-intelligent-tiering/

  • 0 kudos
brickster_2018
by Databricks Employee
  • 2247 Views
  • 1 replies
  • 0 kudos
  • 2247 Views
  • 1 replies
  • 0 kudos
Latest Reply
aladda
Databricks Employee
  • 0 kudos

Deleting the Delta log directory would cause you to lose the underlying transaction history on the delta table and other delta related optimizations. In effect the table would be converted to a Parquet table at that point

  • 0 kudos
User16790091296
by Contributor II
  • 1542 Views
  • 2 replies
  • 0 kudos
  • 1542 Views
  • 2 replies
  • 0 kudos
Latest Reply
Taha
Databricks Employee
  • 0 kudos

Also, a lot of examples here: https://docs.databricks.com/administration-guide/clusters/policies.html#cluster-policy-examples

  • 0 kudos
1 More Replies
brickster_2018
by Databricks Employee
  • 1963 Views
  • 1 replies
  • 0 kudos
  • 1963 Views
  • 1 replies
  • 0 kudos
Latest Reply
aladda
Databricks Employee
  • 0 kudos

Global: run on every cluster in the workspace. They can help you to enforce consistent cluster configurations across your workspace. Use them carefully because they can cause unanticipated impacts, like library conflicts. Only admin users can create ...

  • 0 kudos
brickster_2018
by Databricks Employee
  • 1003 Views
  • 1 replies
  • 1 kudos
  • 1003 Views
  • 1 replies
  • 1 kudos
Latest Reply
aladda
Databricks Employee
  • 1 kudos

For Databricks Runtime 5.5 LTS, Spark jobs, Python notebook cells, and library installation all support both Python 2 and 3.The default Python version for clusters created using the UI is Python 3. In Databricks Runtime 5.5 LTS the default version fo...

  • 1 kudos

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.

If there isn’t a group near you, start one and help create a community that brings people together.

Request a New Group
Labels