cancel
Showing results for 
Search instead for 
Did you mean: 
Get Started Discussions
Start your journey with Databricks by joining discussions on getting started guides, tutorials, and introductory topics. Connect with beginners and experts alike to kickstart your Databricks experience.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Phani1
by Databricks MVP
  • 1833 Views
  • 0 replies
  • 0 kudos

Billing usage per user

Hi Team ,Unity catalog is not enabled in our workspace, We would like to know the billing usage information per user ,could you please help us how to get these details( by using notebook level script).Regards,Phanindra

  • 1833 Views
  • 0 replies
  • 0 kudos
alexiswl
by Contributor
  • 10045 Views
  • 1 replies
  • 0 kudos

Resolved! 'Unity Catalog Volumes is not enabled on this instance' error

Hi all,tl;dr I ran the following on a docker-backed personal compute instance (running 13.3-LTS)```%sqlUSE CATALOG hail;USE SCHEMA volumes_testing;CREATE VOLUME 1kg    COMMENT 'Testing 1000 Genomes volume';```But this gives```ParseException: [UC_VOLU...

  • 10045 Views
  • 1 replies
  • 0 kudos
Latest Reply
alexiswl
Contributor
  • 0 kudos

Resolved with the setting "spark.databricks.unityCatalog.volumes.enabled" = "true"

  • 0 kudos
nyck33
by New Contributor II
  • 2852 Views
  • 1 replies
  • 0 kudos

Databricks learning festival, but my trial is over

I just emailed the onboarding-help email account to ask for an extension for 2 weeks as I want to complete the Data Engineer course to prepare for my new position. I have 2 accounts where the trial expired, one community account which cannot be used ...

  • 2852 Views
  • 1 replies
  • 0 kudos
Latest Reply
nyck33
New Contributor II
  • 0 kudos

is what happened when trying to sign up with another email.

  • 0 kudos
hukel
by Contributor
  • 4958 Views
  • 1 replies
  • 0 kudos

Resolved! Convert multiple string fields to int or long during streaming

Source data looks like: { "IntegrityLevel": "16384", "ParentProcessId": "10972929104936", "SourceProcessId": "10972929104936", "SHA256Hash": "a26a1ffb81a61281ffa55cb7778cc3fb0ff981704de49f75f51f18b283fba7a2", "ImageFileName": "\\Device\\Harddisk...

  • 4958 Views
  • 1 replies
  • 0 kudos
Latest Reply
hukel
Contributor
  • 0 kudos

Thanks for confirming that the readStream.withColumn() approach is the best available option.  Unfortunately, this will force me to maintain a separate notebook for each of the event types,  but it does work.   I was hoping to create just one paramet...

  • 0 kudos
anandreddy23
by New Contributor III
  • 5494 Views
  • 1 replies
  • 0 kudos

unpersist doesn't clear

from pyspark.sql import SparkSessionfrom pyspark import SparkContext, SparkConffrom pyspark.storagelevel import StorageLevelspark = SparkSession.builder.appName('TEST').config('spark.ui.port','4098').enableHiveSupport().getOrCreate()df4 = spark.sql('...

  • 5494 Views
  • 1 replies
  • 0 kudos
Latest Reply
anandreddy23
New Contributor III
  • 0 kudos

Thank you so much for taking time and explaining the concepts

  • 0 kudos
rpl
by Contributor
  • 3985 Views
  • 2 replies
  • 1 kudos

Bug report: the delimiter option does not work when run on DLT

I have a semicolon separated file in an ADLS container that's been added to Unity Catalog as an External location.When I run the following code on an all-purpose cluster, it runs ok and displays the schema.import dlt @dlt.table def test_data_csv(): ...

  • 3985 Views
  • 2 replies
  • 1 kudos
Latest Reply
rpl
Contributor
  • 1 kudos

@Retired_mod can you confirm that .option("delimiter", ";") is ignored when run in a DLT pipeline? (please see the post above) My colleage confirmed the behavior. 

  • 1 kudos
1 More Replies
MFrandsen
by New Contributor
  • 1105 Views
  • 0 replies
  • 0 kudos

Question for exam project

For my exam i have to do a small project for the company im interning at. I am creating a datawarehouse where i will have to transfer data from another database, and then transforming it to a star schema. would databricks be good for this, or is it t...

  • 1105 Views
  • 0 replies
  • 0 kudos
dvmentalmadess
by Valued Contributor
  • 5872 Views
  • 1 replies
  • 2 kudos

Get exceptionTraceId details

I'm getting the following error:  module.consumer_stage_catalog.databricks_external_location.catalog: Creating... â•· │ Error: cannot create external location: AWS IAM role does not have READ permissions on url s3://[bucket name]/catalogs. Please conta...

  • 5872 Views
  • 1 replies
  • 2 kudos
TrevorC
by New Contributor III
  • 4982 Views
  • 1 replies
  • 0 kudos

Resolved! Can we pass parameters thru SQL UDF's?

Is it possible to pass a parameter to a SQL UDF to another SQL UDF that is called by the first SQL UDF?Below is an example where I would like to call tbl_filter() from tbl_func() by passing the tbl_func.a_val parameter to tbl_filter(). Obviously, I c...

  • 4982 Views
  • 1 replies
  • 0 kudos
inspiredelement
by New Contributor
  • 2331 Views
  • 0 replies
  • 0 kudos

company profile

At Inspired Elements, we redefine living spaces in London, offering bespoke fitted wardrobes and fitted kitchens that seamlessly blend functionality with exquisite design. Our commitment to innovation and quality ensures every piece is a work of art,...

  • 2331 Views
  • 0 replies
  • 0 kudos
elgeo
by Valued Contributor II
  • 7902 Views
  • 1 replies
  • 1 kudos

Resolved! Αdd columns delta table

Hello. Do you know if you can add columns at a specific position (before / after a column) by altering a delta table ?

  • 7902 Views
  • 1 replies
  • 1 kudos
Latest Reply
-werners-
Esteemed Contributor III
  • 1 kudos

yes, using the FIRST or AFTER parameter.https://docs.databricks.com/en/sql/language-manual/sql-ref-syntax-ddl-alter-table-manage-column.html#add-column

  • 1 kudos
memo
by New Contributor II
  • 9579 Views
  • 1 replies
  • 0 kudos

Pivot on multiple columns

I want to pass multiple column as argument to pivot a dataframe  in pyspark pivot likemydf.groupBy("id").pivot("day","city").agg(F.sum("price").alias("price"),F.sum("units").alias("units")).show(). One way I found is to create multiple df with differ...

  • 9579 Views
  • 1 replies
  • 0 kudos
rudyevers
by New Contributor III
  • 8504 Views
  • 4 replies
  • 4 kudos

Unity catalog - external table lastUpdateversion

We are currently upgrading our Lakehouse to use the Unity Catalog benefits. We will mostly use external tables because alle our DETLA tables are already stored in Azure Storage. I try to figure out how to update the table property "delta.lastUpdateve...

  • 8504 Views
  • 4 replies
  • 4 kudos
Latest Reply
-werners-
Esteemed Contributor III
  • 4 kudos

I am in the same boat.That is the reason I opted to use managed tables instead.  OK; it means migrating tables and changing notebooks but besides not having to struggle with external tables, you also get something in return (liquid clustering f.e.).

  • 4 kudos
3 More Replies
Phani1
by Databricks MVP
  • 1494 Views
  • 2 replies
  • 0 kudos

checklist for : process to move and deploy in the prod

Hi Team,Could you please help me with best practices to move and deploy (code, workspace, notebooks, etc) in the prod?Regards,Phanindra

  • 1494 Views
  • 2 replies
  • 0 kudos
Latest Reply
-werners-
Esteemed Contributor III
  • 0 kudos

the most important is to use Repos!Link your workspace with git and use feature branches and pull requests to promote code/notebooks.Check the databricks docs on Repos.  If you have further questions; shoot.

  • 0 kudos
1 More Replies
Phani1
by Databricks MVP
  • 3015 Views
  • 1 replies
  • 0 kudos

Archival Strategy for Delta tables

 Hi Team, We would like to define the archival strategy for data. Could you please share best practices /guide me on the below are the 3 use cases Case-1: On-Prem SQL and Oracle Data which is more than 20 years and they wanted to bring them into clou...

  • 3015 Views
  • 1 replies
  • 0 kudos
Latest Reply
-werners-
Esteemed Contributor III
  • 0 kudos

case 1: I'd extract the data from the db to a data lake (cold storage if that is possible, that is cheaper) using an ETL tool like Data Factory, Glue etc.  Then the archiving can take place.  Perhaps also create a backup of the data on a 2nd data lak...

  • 0 kudos

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels