cancel
Showing results for 
Search instead for 
Did you mean: 
Get Started Discussions
Start your journey with Databricks by joining discussions on getting started guides, tutorials, and introductory topics. Connect with beginners and experts alike to kickstart your Databricks experience.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

traillog
by New Contributor
  • 1533 Views
  • 0 replies
  • 0 kudos

Response code 400 received when using VSCode on Windows 10 but no issue while using Ubuntu

I use VSCode on Windows 10 for building and deploying a workflow from my system and always encounter response code 400 when trying to deploy it. I am able to deploy the workflows via Ubuntu, but not via Windows. Has anyone encountered this issue befo...

  • 1533 Views
  • 0 replies
  • 0 kudos
zgreen
by New Contributor
  • 868 Views
  • 0 replies
  • 0 kudos

jobs.python_wheel_task.enty_point can't find entry points defined in dependency packages

Let's say I have packageA with no entry points, packageA dependents on dependencyA package, which has entry-points.In order to be able to use those entrypoints, i.e.```yamlpython_wheel_task:  package_name: packageA  entry_point:dependencyA_entry```I ...

  • 868 Views
  • 0 replies
  • 0 kudos
Verr
by New Contributor II
  • 2529 Views
  • 2 replies
  • 0 kudos

child notebook is not displaying output.

I have built pipeline to execute databricks notebook having SQL scripts. It is executing notebook but not able to see output for each cell. I am executing child notebook through driver notebook.

  • 2529 Views
  • 2 replies
  • 0 kudos
Latest Reply
koushiknpvs
New Contributor III
  • 0 kudos

Hi Verr,In short it depends on how your child notebook is configured. But I would start with the following points -Output Logging Settings: Check the logging settings for your notebook cells. Ensure that the cells are configured to display output. In...

  • 0 kudos
1 More Replies
GeKo
by Contributor
  • 5326 Views
  • 6 replies
  • 2 kudos

Resolved! column "storage_sub_directory" is now always NULL in system.information_schema.tables

Hello,I am running a job that depends on the information provided in column storage_sub_directory in system.information_schema.tables .... and it worked until 1-2 weeks ago.Now I discovered in the doc that this column is deprecated and always null , ...

Get Started Discussions
Unity Catalog
unitycatalog
  • 5326 Views
  • 6 replies
  • 2 kudos
Latest Reply
GeKo
Contributor
  • 2 kudos

Many thanks for the update @NandiniN 

  • 2 kudos
5 More Replies
Lucifer
by New Contributor
  • 1554 Views
  • 1 replies
  • 0 kudos

displaying unity catalog metadata and other information in sharePoint

Is there any connectors or api which we can use to display metadata information stored in Unity catalog to business users using SharePoint.

  • 1554 Views
  • 1 replies
  • 0 kudos
Latest Reply
Ajay-Pandey
Esteemed Contributor III
  • 0 kudos

Hi @Lucifer As metadata are stored in system schema as table that means you can use databricks to extract the data from databricks and display it to Sharepoint Docs - Statement Execution API: Run SQL on warehouses | Databricks on AWS

  • 0 kudos
Shawn_Eary
by Contributor
  • 9498 Views
  • 2 replies
  • 0 kudos

Create Team Git Repo in Azure Databricks

When I attach a Git repo to Databricks, it always puts the repo under my username/domain name:How can I create a "team" repo at the top level, so teammates don't have to drill into my username?

Shawn_Eary_0-1702321413271.png
  • 9498 Views
  • 2 replies
  • 0 kudos
Latest Reply
DonatienTessier
Contributor
  • 0 kudos

Hi,Interest of using a repo is to have a dedicated area for each of developers.If you want to have only a folder with the last version of the code, you should a CI/CD pipeline that will package the code and then delivered into a folder inside Workspa...

  • 0 kudos
1 More Replies
BLPedersen
by New Contributor II
  • 8380 Views
  • 3 replies
  • 1 kudos

Error when setting up OAuth profile for Databricks CLI

Hello.I'm currently trying to migrate a project from dbx to Databricks Asset Bundles. I have successfully created the required profile using U2M authentication with the command```databricks auth login --host <host-name>```I'm able to see the new prof...

  • 8380 Views
  • 3 replies
  • 1 kudos
Latest Reply
kunalmishra9
New Contributor III
  • 1 kudos

I ran into a similar error just now, and in my case, Pycharm was running some iPython startup scripts each time it opened a console. There was, for some reason, a file at `~/.ipython/profile_default/startup/00-databricks-init-a5acf3baa440a896fa364d18...

  • 1 kudos
2 More Replies
Ahmad_A
by New Contributor II
  • 1345 Views
  • 1 replies
  • 0 kudos

Using Community Edition instance for Customer Academy’s DE Learning Plan

Hi!So I’ve been wondering since I started with the Data Engineering Learning Plan on the Customer Academy, should I go with my Community Edition Databricks, or I should go with creating a premium edition on either a cloud provider or the website.Than...

Get Started Discussions
Customer Academy
  • 1345 Views
  • 1 replies
  • 0 kudos
Latest Reply
Ahmad_A
New Contributor II
  • 0 kudos

Hey, @Kaniz. May I recommend deleting the post if it is not in the right place or changing the forum so I get the proper response?

  • 0 kudos
_databreaks
by New Contributor II
  • 1733 Views
  • 0 replies
  • 0 kudos

DLT to push data instead of a pull

I am relatively new to Databricks, and from my recent experience it appears that every step in a DLT Pipeline, we define each LIVE TABLES (be it streaming or not) to pull data upstream.I have yet to see an implementation where data from upstream woul...

Get Started Discussions
dlt
DLT pipeline
  • 1733 Views
  • 0 replies
  • 0 kudos
NarenderKumar
by New Contributor III
  • 4891 Views
  • 1 replies
  • 1 kudos

Resolved! Unable to provide access in unity catalog using SQL commands

I am trying to provide access in unity catalog using the SQL commands.I am following the below documentation:https://learn.microsoft.com/en-us/azure/databricks/data-governance/unity-catalog/manage-privileges/It suggests to create SQL commands in belo...

NarenderKumar_2-1715253228379.png NarenderKumar_1-1715253206943.png NarenderKumar_3-1715253369442.png
  • 4891 Views
  • 1 replies
  • 1 kudos
Latest Reply
mhiltner
Databricks Employee
  • 1 kudos

Try this. For some reason the quotes are crazy when using these commands.  GRANT USAGE ON CATALOG `uda_dev` TO `your-group`   GRANT SELECT ON SCHEMA uda_dev.default TO `your-group` (without quotes for the schema) 

  • 1 kudos
Kuchnhi
by New Contributor III
  • 1191 Views
  • 1 replies
  • 0 kudos

Databricks workflow jobs run is taking Double time in EU Region

We have a scheduled job in Databricks workflow, This Job run is taking aroud 5 hours Previously before 1 month it was tasking on 2.5 hours. Can any one tell what may be the reason behind this. Note: There is no change has been made in this period of ...

  • 1191 Views
  • 1 replies
  • 0 kudos
Latest Reply
Wojciech_BUK
Valued Contributor III
  • 0 kudos

You can check if you are using spot instances on your Job Cluster.btw. if you are using Azure West Europe is on very high demand and sometimes it takes time to provision compute.But it should be matter of minutes, not hours.Check maybe if your data v...

  • 0 kudos
ducng
by New Contributor II
  • 12152 Views
  • 3 replies
  • 0 kudos

Resolved! How to pass variables to a python file job

Hi everyone,It's relatively straight forward to pass a value to a key-value pair in notebook job. But for the python file job however, I couldn't figure out how to do it. Does anyone have any idea?Have been tried out different variations for a job wi...

ducng_0-1696256847837.png ducng_1-1696256883016.png
  • 12152 Views
  • 3 replies
  • 0 kudos
Latest Reply
Bruno-Castro
New Contributor II
  • 0 kudos

Thanks so much for this! By the way, is there a way to do it with the JSON interface? I am struggling to get the parameters if entered in this way  

  • 0 kudos
2 More Replies
Nagrjuna
by New Contributor II
  • 1507 Views
  • 1 replies
  • 0 kudos

Chat Bot with Azure blob and databricks

Hi Team, I am thinking to start a chat bot application for teams to query data from Azure blob and data bricks tables in python programming language.Please help me out on how i can start and which tools i can use for this requirement.Thanks in advanc...

  • 1507 Views
  • 1 replies
  • 0 kudos
Latest Reply
Yeshwanth
Databricks Employee
  • 0 kudos

@Nagrjuna , that's a great idea! Although we do not know about your use case completely, I am sure you would definitely fall in love with our AI/ML Products. To create a Python chat bot application that can pull data from Azure Blob Storage and Datab...

  • 0 kudos
brian999
by Contributor
  • 2494 Views
  • 1 replies
  • 1 kudos

How to configure github credentials for a service principal NOT using Azure

I want to have a service principal run a job that uses a notebook in our github. We are AWS not Azure. How do I configure git credentials for the service principal? Does this use deploy keys?

  • 2494 Views
  • 1 replies
  • 1 kudos
Latest Reply
brian999
Contributor
  • 1 kudos

Awesome details thank you for your help.

  • 1 kudos

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels