cancel
Showing results for 
Search instead for 
Did you mean: 
Get Started Discussions
Start your journey with Databricks by joining discussions on getting started guides, tutorials, and introductory topics. Connect with beginners and experts alike to kickstart your Databricks experience.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

pjv
by New Contributor III
  • 1347 Views
  • 1 replies
  • 0 kudos

VSCode Databricks Extension

Hi all,I've been trying to sync my VSCode IDE with our Databricks GCP workspace using the Databricks extension. I am able to connect authenticate my account and workspace and find our clusters. However, when I try to sync a destination it throws a st...

  • 1347 Views
  • 1 replies
  • 0 kudos
Latest Reply
pjv
New Contributor III
  • 0 kudos

@Retired_mod thanks for you response.I am not running through a proxy. At least, not on purpose. How do I know if I am running through a proxy? And where can I find the values of <proxy_url> and <port> so that I can try restarting my VSCode.I have tr...

  • 0 kudos
rachit-prodigal
by New Contributor
  • 24587 Views
  • 0 replies
  • 0 kudos

Identity column has null values

I created a table in databricks using a dbt model pre hook  CREATE TABLE IF NOT EXISTS accounts ( account_id BIGINT GENERATED ALWAYS AS IDENTITY, description STRING other columns)I use the same dbt model to merge values into this table in the post...

  • 24587 Views
  • 0 replies
  • 0 kudos
unity_Catalog
by New Contributor III
  • 1712 Views
  • 0 replies
  • 0 kudos

UCX Installation Error

While downloading and installing ucx from a shell code,I am facing the below error. Can anyone provide a solution[i] Creating isolated Virtualenv with Python: /c/Program Files/Python312/pythonActual environment location may have moved due to redirect...

  • 1712 Views
  • 0 replies
  • 0 kudos
leelee3000
by Databricks Employee
  • 6155 Views
  • 3 replies
  • 1 kudos

Setting up Unity Catalog in Azure

Trying to create a metastore that will be connected to an external storage (ADLS) but we don't have the option to create a new metastore in 'Catalog' tab in the UI. Based on some research, we see that we'll have to go into "Manage Account" and then c...

  • 6155 Views
  • 3 replies
  • 1 kudos
Latest Reply
bsadler
New Contributor II
  • 1 kudos

I have been wrestling with this question for days now.  I seem to be the only one with this question so I am sure I am doing something wrong.  I am trying to create a UC metastore but there is not an option in "Catalog" to create a metastore.  This s...

  • 1 kudos
2 More Replies
nikhilprajapati
by New Contributor
  • 2277 Views
  • 2 replies
  • 1 kudos

Data in dataframe is also getting deleted when we are trying to delete records from underlying table

  Hi , We are trying to load data from a delta table to a dataframe(a copy of original table) . Initially delta table has count 911 . The dataframe in which the data is loaded also has the same count .Now,  we are deleting some records from the delta...

nikhilprajapati_1-1701930598953.png nikhilprajapati_2-1701930598960.png nikhilprajapati_3-1701930598967.png nikhilprajapati_4-1701930598974.png
  • 2277 Views
  • 2 replies
  • 1 kudos
Latest Reply
Hkesharwani
Contributor II
  • 1 kudos

Hi, There is a way to retain the copy of data frame, even if the data in underling table is manipulated but that's a memory expensive operation, be careful while using it.df1 = spark.createDataFrame(df.rdd.map(lambda x: x), schema=df.schema)Here we a...

  • 1 kudos
1 More Replies
Ramakrishnan83
by New Contributor III
  • 20687 Views
  • 6 replies
  • 0 kudos

Renaming the database Name in Databricks

Team,Initially our team created the databases with the environment name appended. Ex: cust_dev, cust_qa, cust_prod.I am looking to standardize the database name as consistent name across environments. I want to rename to "cust". All of my tables are ...

  • 20687 Views
  • 6 replies
  • 0 kudos
Latest Reply
Avvar2022
Contributor
  • 0 kudos

You can also use “CASCADE” to drop schema and tables as well. It is recursive. 

  • 0 kudos
5 More Replies
eheinlein
by New Contributor
  • 1020 Views
  • 0 replies
  • 0 kudos

How to confirm a workspace ID via an api token?

Hello! We are integrating with Databricks and we get the API key, workspace ID, and host from our users in order to connect to Databricks. We need the to validate the workspace ID because we do need it outside of the context of the API key (with webh...

  • 1020 Views
  • 0 replies
  • 0 kudos
df_Jreco
by New Contributor II
  • 1163 Views
  • 1 replies
  • 0 kudos

Custom python package iin Notebook task using bundle

Hi mates!I'n my company, we are moving our pipelines to Databricks bundles, our pipelines use a notebook that receives some parameters.This notebook uses a custom python package to apply the business logic based on the parameters that receive.The thi...

Get Started Discussions
databricks-bundles
  • 1163 Views
  • 1 replies
  • 0 kudos
Latest Reply
df_Jreco
New Contributor II
  • 0 kudos

 Solved understanding the databricks.yml configuration!

  • 0 kudos
esi
by New Contributor
  • 11325 Views
  • 2 replies
  • 0 kudos

numpy.ndarray size changed, may indicate binary incompatibility

Hi All,I have installed the following libraries on my cluster (11.3 LTS that includes Apache Spark 3.3.0, Scala 2.12):numpy==1.21.4flair==0.12 ‎on executing `from flair.models import TextClassifier`, I get the following error:"numpy.ndarray size chan...

  • 11325 Views
  • 2 replies
  • 0 kudos
Latest Reply
sean_owen
Databricks Employee
  • 0 kudos

You have changed the numpy version, and presumably that is not compatible with other libraries in the runtime. If flair requires later numpy, then use a later DBR runtime for best results, which already has later numpy versions

  • 0 kudos
1 More Replies

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels