cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

NOOR_BASHASHAIK
by Contributor
  • 1956 Views
  • 1 replies
  • 1 kudos

Unity Catalog - addition of account groups/users to workspaces

Hi allwe have set-up metastore, and were doing certain activities as part of MVP.we realized in a particular databricks workspace that was enabled with UC, in admin settings > "Add Groups" section, user groups from other platforms/projects which leve...

  • 1956 Views
  • 1 replies
  • 1 kudos
Latest Reply
Ismail1
New Contributor III
  • 1 kudos

Good question, didn't think of it that way, from my understanding UC uses users pushed from the account console and not from workspaces, One way to restrict would be to restrict other workspaces from using said catalog and also control ACLs with the ...

  • 1 kudos
del1000
by New Contributor III
  • 936 Views
  • 0 replies
  • 0 kudos

Problem with sparkContext.parallelize and volatile functions?

I have a code:from time import sleep from random import random from operator import add   def f(a: int) -> float: sleep(0.1) return random() rdd1 = sc.parallelize(range(20), 2) rdd2 = sc.parallelize(range(20), 2) rdd3 = sc.parallelize(rang...

  • 936 Views
  • 0 replies
  • 0 kudos
MohamedThanveer
by New Contributor II
  • 1703 Views
  • 2 replies
  • 0 kudos

Databricks Certified Associate Developer for Apache Spark 3.0 - Python refund

I have scheduled an examination on 1st June 2023 and due to personal reason, I have cancelled the examination on 26th May 2023 (more than 72 hours) but I am yet to receive the refund amount. In the auto generated mail it is mentioned that the refund ...

  • 1703 Views
  • 2 replies
  • 0 kudos
Latest Reply
Anonymous
Not applicable
  • 0 kudos

Hi @Mohamed Thanveer​ Thank you for reaching out! Please submit a ticket to our Training Team here: https://help.databricks.com/s/contact-us?ReqType=training  and our team will get back to you shortly. 

  • 0 kudos
1 More Replies
yhyhy3
by New Contributor III
  • 19088 Views
  • 1 replies
  • 2 kudos

Best way to make a scrollable DataFrame in a ipywidgets.Output?

I have a pretty complex Jupyter widgets UI in a databricks notebook that has a dataframe that1. will be modified by some Jupyter widget callbacks2. needs to be displayed to the user and updated as it is modified3. is large and needs to support vertic...

  • 19088 Views
  • 1 replies
  • 2 kudos
Latest Reply
Anonymous
Not applicable
  • 2 kudos

Hi @Yushi Homma​ Thank you for posting your question in our community! We are happy to assist you.To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best answers y...

  • 2 kudos
NOOR_BASHASHAIK
by Contributor
  • 8110 Views
  • 3 replies
  • 3 kudos

Resolved! Notebooks & Jobs Migration from one workspace to another

Hi allI have a need to migrate just notebooks & jobs from one workspace to another. Is there an utility to do so?

  • 8110 Views
  • 3 replies
  • 3 kudos
Latest Reply
karthik_p
Esteemed Contributor
  • 3 kudos

@NOOR BASHA SHAIK​ you can also try databricks connect in addition to @Artem Sheiko​ information that was provided .

  • 3 kudos
2 More Replies
Erik
by Valued Contributor III
  • 2140 Views
  • 2 replies
  • 2 kudos

Create python modules for both repos and workspace

We are using the "databricks_notebook" terraform resource to deploy our notebooks into the "Workspace" as part of our CICD run, and our jobs run notebooks from the workspace. For development we clone the repo into "Repos". At the moment the only modu...

  • 2140 Views
  • 2 replies
  • 2 kudos
Latest Reply
RobiTakToRobi
New Contributor II
  • 2 kudos

You can create your own Python package and host it in Azure Artifacts. https://learn.microsoft.com/en-us/azure/devops/artifacts/quickstarts/python-packages?view=azure-devops

  • 2 kudos
1 More Replies
qwerty1
by Contributor
  • 1008 Views
  • 0 replies
  • 0 kudos

Why am I not able to view all table properties?

We have a live streaming table created using the commandCREATE OR REFRESH STREAMING LIVE TABLE foo TBLPROPERTIES ( "pipelines.autoOptimize.zOrderCols" = "c1,, c2, c3, c4", "delta.randomizeFilePrefixes" = "true" );But when I run the show table propert...

  • 1008 Views
  • 0 replies
  • 0 kudos
Priyag1
by Honored Contributor II
  • 5888 Views
  • 7 replies
  • 3 kudos

Resolved! Community Reward points

Hi Data bricks team , It is informed that all the lifetime points on community will be retired by June 2023. Is it the same for brightsites reward store also .Will the brightsites get restored or lost along with community life time points ? @Vidula...

  • 5888 Views
  • 7 replies
  • 3 kudos
Latest Reply
Sujitha
Databricks Employee
  • 3 kudos

@Priyadarshini G​ So, we are updating our community platform to provide a better experience for our users and members. As part of the updates, we plan on introducing a new recognition protocol that will better celebrate members for their contribution...

  • 3 kudos
6 More Replies
Satty
by New Contributor
  • 7004 Views
  • 1 replies
  • 0 kudos

Solution for ConnectException error: This is often caused by an OOM error that causes the connection to the Python REPL to be closed. Check your query's memory usage.

When ever I am trying to run and load multiple files in single dataframe for processing (overall file size is more than 15 gb in single dataframe at the end of the loop, my code is crashing everytime with the below error...ConnectException error: Thi...

  • 7004 Views
  • 1 replies
  • 0 kudos
Latest Reply
pvignesh92
Honored Contributor
  • 0 kudos

@Satish Agarwal​ It seems your system memory is not sufficient to load the 15GB file. I believe you are using Python Pandas data frame for loading 15GB file and not using Spark. Is there any particular reason that you cannot use Spark for this.

  • 0 kudos
pvignesh92
by Honored Contributor
  • 1436 Views
  • 0 replies
  • 0 kudos

Very often, we need to know how many files my table path contains and the overall size of the path for various optimizations. In the past, I had to wr...

Very often, we need to know how many files my table path contains and the overall size of the path for various optimizations. In the past, I had to write my own logic to accomplish this.Delta Lake is making life easier. See how simple it is to obtain...

1684878098472
  • 1436 Views
  • 0 replies
  • 0 kudos
Databricks3
by Contributor
  • 2636 Views
  • 1 replies
  • 0 kudos

%run is not working

I have created two notebooks in my workspace one(A) is having function and other one(B) is having the main code.I am trying to use the %run magic function to use the functions available in notebook A from notebook B. Both notebooks are on the same pa...

  • 2636 Views
  • 1 replies
  • 0 kudos
Latest Reply
Ajay-Pandey
Esteemed Contributor III
  • 0 kudos

@SK ASIF ALI​ â€‹It's not supported in community if you are using premium version then this should workCan you please send the snippet of the error message? #DAIS2023​ 

  • 0 kudos
carlosst01
by New Contributor II
  • 2467 Views
  • 0 replies
  • 0 kudos

java.lang.SecurityException: Could not verify permissions for DeleteFromTable ERROR IN EXTERNAL SYNAPSE TABLE

Hi guys, I am trying to delete an external table in Databricks that is connected to a Synapse table (via the com.databricks.spark.sqldw connector) but got the error shown below:Apart from that, when I try to execute an DELETE and INSERT actions (like...

WhatsApp Image 2023-05-25 at 12.02.52 PM WhatsApp Image 2023-05-25 at 12.08.16 PM WhatsApp Image 2023-05-25 at 12.13.14 PM
  • 2467 Views
  • 0 replies
  • 0 kudos
darkraisisi
by New Contributor
  • 1152 Views
  • 0 replies
  • 0 kudos

Is there a way to manually update the cuda required file in the db runtime? There are some rather annoying bugs still in TF 2.11 that have been fixed ...

Is there a way to manually update the cuda required file in the db runtime?There are some rather annoying bugs still in TF 2.11 that have been fixed in TF 2.12.Sadly the latest DB runtime 13.1 (beta) only supports the older TF 2.11 even tho 2.12 was ...

  • 1152 Views
  • 0 replies
  • 0 kudos
Brian61
by New Contributor
  • 1222 Views
  • 0 replies
  • 0 kudos
  • 1222 Views
  • 0 replies
  • 0 kudos

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels