cancel
Showing results for 
Search instead for 
Did you mean: 
Community Platform Discussions
Connect with fellow community members to discuss general topics related to the Databricks platform, industry trends, and best practices. Share experiences, ask questions, and foster collaboration within the community.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Lucifer
by New Contributor
  • 1142 Views
  • 1 replies
  • 0 kudos

displaying unity catalog metadata and other information in sharePoint

Is there any connectors or api which we can use to display metadata information stored in Unity catalog to business users using SharePoint.

  • 1142 Views
  • 1 replies
  • 0 kudos
Latest Reply
Ajay-Pandey
Esteemed Contributor III
  • 0 kudos

Hi @Lucifer As metadata are stored in system schema as table that means you can use databricks to extract the data from databricks and display it to Sharepoint Docs - Statement Execution API: Run SQL on warehouses | Databricks on AWS

  • 0 kudos
_databreaks
by New Contributor II
  • 1085 Views
  • 0 replies
  • 0 kudos

DLT to push data instead of a pull

I am relatively new to Databricks, and from my recent experience it appears that every step in a DLT Pipeline, we define each LIVE TABLES (be it streaming or not) to pull data upstream.I have yet to see an implementation where data from upstream woul...

  • 1085 Views
  • 0 replies
  • 0 kudos
NarenderKumar
by New Contributor III
  • 3698 Views
  • 1 replies
  • 1 kudos

Resolved! Unable to provide access in unity catalog using SQL commands

I am trying to provide access in unity catalog using the SQL commands.I am following the below documentation:https://learn.microsoft.com/en-us/azure/databricks/data-governance/unity-catalog/manage-privileges/It suggests to create SQL commands in belo...

NarenderKumar_2-1715253228379.png NarenderKumar_1-1715253206943.png NarenderKumar_3-1715253369442.png
  • 3698 Views
  • 1 replies
  • 1 kudos
Latest Reply
mhiltner
Databricks Employee
  • 1 kudos

Try this. For some reason the quotes are crazy when using these commands.  GRANT USAGE ON CATALOG `uda_dev` TO `your-group`   GRANT SELECT ON SCHEMA uda_dev.default TO `your-group` (without quotes for the schema) 

  • 1 kudos
Kuchnhi
by New Contributor
  • 804 Views
  • 1 replies
  • 0 kudos

Databricks workflow jobs run is taking Double time in EU Region

We have a scheduled job in Databricks workflow, This Job run is taking aroud 5 hours Previously before 1 month it was tasking on 2.5 hours. Can any one tell what may be the reason behind this. Note: There is no change has been made in this period of ...

  • 804 Views
  • 1 replies
  • 0 kudos
Latest Reply
Wojciech_BUK
Valued Contributor III
  • 0 kudos

You can check if you are using spot instances on your Job Cluster.btw. if you are using Azure West Europe is on very high demand and sometimes it takes time to provision compute.But it should be matter of minutes, not hours.Check maybe if your data v...

  • 0 kudos
ducng
by New Contributor II
  • 8910 Views
  • 3 replies
  • 0 kudos

Resolved! How to pass variables to a python file job

Hi everyone,It's relatively straight forward to pass a value to a key-value pair in notebook job. But for the python file job however, I couldn't figure out how to do it. Does anyone have any idea?Have been tried out different variations for a job wi...

ducng_0-1696256847837.png ducng_1-1696256883016.png
  • 8910 Views
  • 3 replies
  • 0 kudos
Latest Reply
Bruno-Castro
New Contributor II
  • 0 kudos

Thanks so much for this! By the way, is there a way to do it with the JSON interface? I am struggling to get the parameters if entered in this way  

  • 0 kudos
2 More Replies
Nagrjuna
by New Contributor II
  • 1070 Views
  • 1 replies
  • 0 kudos

Chat Bot with Azure blob and databricks

Hi Team, I am thinking to start a chat bot application for teams to query data from Azure blob and data bricks tables in python programming language.Please help me out on how i can start and which tools i can use for this requirement.Thanks in advanc...

  • 1070 Views
  • 1 replies
  • 0 kudos
Latest Reply
Yeshwanth
Databricks Employee
  • 0 kudos

@Nagrjuna , that's a great idea! Although we do not know about your use case completely, I am sure you would definitely fall in love with our AI/ML Products. To create a Python chat bot application that can pull data from Azure Blob Storage and Datab...

  • 0 kudos
benitoski
by New Contributor II
  • 1704 Views
  • 1 replies
  • 1 kudos

Resolved! Workspace FileNotFoundExecption

I have a model created with catboost and exported in onnx format in workspace and I want to download that model to my local machine.I tried to use the Export that is in the three points to the right of the model, but the model weighs more than 10 Mb ...

  • 1704 Views
  • 1 replies
  • 1 kudos
Latest Reply
feiyun0112
Honored Contributor
  • 1 kudos

you need put file to FileStorehttps://docs.databricks.com/en/dbfs/filestore.html#save-a-file-to-filestore 

  • 1 kudos
Benedetta
by New Contributor III
  • 2077 Views
  • 1 replies
  • 0 kudos

What happened to the ephemeral notebook links????? and the job ids????

Hey Databricks,      Why did you remove the ephemeral notebook links and job Ids from the parallel runs? This has created a huge gap for us. We can no longer view the ephemeral notebooks, and also the Jobids are missing from the output. Waccha doing?...

  • 2077 Views
  • 1 replies
  • 0 kudos
Latest Reply
Benedetta
New Contributor III
  • 0 kudos

Hi Kaniz,    It's funny you mention these things - we are doing some of those - the problem now is that the JobId is obscured from the output meaning we can't tell which ephemeral notebook goes with which JobId.  It looks like the ephemeral notebook ...

  • 0 kudos
Sudheer2
by New Contributor III
  • 3549 Views
  • 0 replies
  • 0 kudos

Updating Databricks SQL Warehouse using Terraform

 We can Update SQL Warehouse manually in Databricks.Click SQL Warehouses in the sidebarIn Advanced optionsWe can find Unity Catalog toggle button there! While Updating Existing SQL Warehouse in Azure to enable unity catalog using terraform, I couldn'...

  • 3549 Views
  • 0 replies
  • 0 kudos
liormayn
by New Contributor III
  • 3324 Views
  • 5 replies
  • 3 kudos

OSError: [Errno 78] Remote address changed

Hello:)as part of deploying an app that previously ran directly on emr to databricks, we are running experiments using LTS 9.1, and getting the following error: PythonException: An exception was thrown from a UDF: 'pyspark.serializers.SerializationEr...

  • 3324 Views
  • 5 replies
  • 3 kudos
Latest Reply
NandiniN
Databricks Employee
  • 3 kudos

Hi @liormayn , I can understand. I see the fix went on 20 March 2024, you would have to restart the clusters. Thanks!

  • 3 kudos
4 More Replies
TinasheChinyati
by New Contributor III
  • 2643 Views
  • 0 replies
  • 0 kudos

Stream to stream join NullPointerException

I have a DLT pipeline running in continous mode. I have a stream to stream join which runs for the first 5hrs but then fails with a Null Pointer Exception. I need assistance to know what I need to do to handle this. my code is structured as below:@dl...

  • 2643 Views
  • 0 replies
  • 0 kudos
Ikanip
by New Contributor II
  • 3417 Views
  • 4 replies
  • 2 kudos

Resolved! How to choose a compute, and how to find alternatives for the current compute being used?

We are using a compute for an Interactive Cluster in Production which incurs X amount of cost. We want to know what are the options available to use with near about the same processing power as the current compute but incur a cost of Y, which is less...

  • 3417 Views
  • 4 replies
  • 2 kudos
Latest Reply
raphaelblg
Databricks Employee
  • 2 kudos

Hello @Ikanip , You can utilize the Databricks Pricing Calculator to estimate costs. For detailed information on compute capacity, please refer to your cloud provider's documentation regarding Virtual Machine instance types.

  • 2 kudos
3 More Replies
scottbisaillon
by New Contributor
  • 1209 Views
  • 0 replies
  • 0 kudos

Databricks Running Jobs and Terraform

What happens to a currently running job when a workspace is deployed again using Terraform? Are the jobs paused/resumed, or are they left unaffected without any down time? Searching for this specific scenario doesn't seem to come up with anything and...

  • 1209 Views
  • 0 replies
  • 0 kudos
Abhay_1002
by New Contributor
  • 644 Views
  • 0 replies
  • 0 kudos

Archive file support in Jar Type application

In my spark application, I am using set of python libraries. I am submitting spark application as Jar Task. But I am not able to find any option provide Archive Files.So, in order to handle python dependencies, I am using approach:Create archive file...

  • 644 Views
  • 0 replies
  • 0 kudos

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.

If there isn’t a group near you, start one and help create a community that brings people together.

Request a New Group
Top Kudoed Authors