cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

shreyag
by New Contributor II
  • 3194 Views
  • 2 replies
  • 0 kudos

scheduling tasks through CLI

Is there a way to schedule tasks or jobs through the Databricks CLI instead of the GUI? I want to be able to create a job flow with different notebook through the CLI.

  • 3194 Views
  • 2 replies
  • 0 kudos
Latest Reply
Atanu
Databricks Employee
  • 0 kudos

I agreed with @Kaniz Fatma​  https://docs.databricks.com/dev-tools/cli/jobs-cli.html?_ga=2.101966982.684786035.1646666830-480220406.1638459894 this is the job CLI we currently support @Shreya Gupta​ 

  • 0 kudos
1 More Replies
Alx
by Databricks Partner
  • 4296 Views
  • 1 replies
  • 0 kudos

Problem with network security group (NSG) rules in case of VNet injection

Hi everyone,Our internal company security policy for the Cloud infrastructure requires to have custom outbound NSG rule that denies all traffic. The rules attributes should be as follows: Priority: 4096Port: AnyProtocol: AnySource: AnyDestination: An...

  • 4296 Views
  • 1 replies
  • 0 kudos
Latest Reply
Atanu
Databricks Employee
  • 0 kudos

HELLO @Alexey Tyulyaev​  please check https://docs.microsoft.com/en-us/azure/virtual-network/manage-network-security-group

  • 0 kudos
alejandrofm
by Valued Contributor
  • 6487 Views
  • 3 replies
  • 3 kudos

Resolved! Delta, the specified key does not exist error

Hi, I'm having this error too frequently on a few tables, I check on S3 and the partition exists and the file is there on the partition.error: Spectrum Scan Error: DeltaManifestcode: 15005context: Error fetching Delta Lake manifest delta/product/sub_...

  • 6487 Views
  • 3 replies
  • 3 kudos
Latest Reply
alejandrofm
Valued Contributor
  • 3 kudos

@Hubert Dudek​ , I'll add that sometimes, just running:GENERATE symlink_format_manifest FOR TABLE schema.tablesolves it, but, how can the symlink get broken?Thanks!

  • 3 kudos
2 More Replies
study_community
by New Contributor III
  • 17872 Views
  • 8 replies
  • 3 kudos

Not able to move files from local to dbfs through dbfs CLI

Hi Folks,I have installed and configured databricks CLI in my local machine. I tried to move a local file from my personal computer using dbfs cp to dbfs:/ path. I can see the file is copied from local, and is only visible in local. I am not able to ...

image image
  • 17872 Views
  • 8 replies
  • 3 kudos
Latest Reply
Anonymous
Not applicable
  • 3 kudos

Hi, Could you try to save the file from your local machine to dbfs:/FileStore location?# Put local file test.py to dbfs:/FileStore/test.pydbfs cp test.py dbfs:/FileStore/test.py

  • 3 kudos
7 More Replies
shrewdTurtle
by New Contributor II
  • 4117 Views
  • 2 replies
  • 3 kudos

Cannot open Jobs tab in Databricks Community edition.

Hi,I get the following exception when I try to open jobs tab.Uncaught TypeError: Cannot read properties of undefined (reading 'apply')   Reload the page and try again. If the error persists, contact support. Reference error code: fd9ae37c18c1400cb15...

  • 4117 Views
  • 2 replies
  • 3 kudos
Latest Reply
shrewdTurtle
New Contributor II
  • 3 kudos

@Kaniz Fatma​ , @Werner Stinckens​ thanks for the clarification. I agree with @Werner Stinckens​ , Error message should be more useful.

  • 3 kudos
1 More Replies
Jan_A
by New Contributor III
  • 7390 Views
  • 3 replies
  • 5 kudos

Resolved! Move/Migrate database from dbfs root (s3) to other mounted s3 bucket

Hi,I have a databricks database that has been created in the dbfs root S3 bucket, containing managed tables. I am looking for a way to move/migrate it to a mounted S3 bucket instead, and keep the database name.Any good ideas on how this can be done?T...

  • 7390 Views
  • 3 replies
  • 5 kudos
Latest Reply
User16753724663
Databricks Employee
  • 5 kudos

Hi @Jan Ahlbeck​ we can use below property to set the default location:"spark.sql.warehouse.dir": "S3 URL/dbfs path"Please let me know if this helps.

  • 5 kudos
2 More Replies
databrick_comm
by New Contributor II
  • 6812 Views
  • 3 replies
  • 0 kudos

Not able to connecting Denodo VDP from databricks

I would like connect Denodo VDP from databrick workspace installed ODBC client and Installed denodo Jar in cluster ,not able to understanding other steps.Could you please me

  • 6812 Views
  • 3 replies
  • 0 kudos
Latest Reply
User16753724663
Databricks Employee
  • 0 kudos

Hi @sathyanarayan kokku​ Are you trying to install denodo vdp server in databricks?

  • 0 kudos
2 More Replies
NAS
by New Contributor III
  • 3065 Views
  • 1 replies
  • 1 kudos

Resolved! "import pandas as pd" => [Errno 5]

When I type import pandas as pdfrom a Notebook in a Repo I get:--------------------------------------------------------------------------- AttributeError Traceback (most recent call last) /usr/lib/python3.8/importlib/_boots...

  • 3065 Views
  • 1 replies
  • 1 kudos
Latest Reply
NAS
New Contributor III
  • 1 kudos

Thanks to Elliott Hertz, I found out that the ML Experiments cannot be stored in the repo. After I moved them to my Workspace everything seems to work.

  • 1 kudos
RohanB
by New Contributor III
  • 8073 Views
  • 8 replies
  • 3 kudos

Resolved! Spark Streaming - Checkpoint State EOF Exception

I have a Spark Structured Streaming job which reads from 2 Delta tables in streams , processes the data and then writes to a 3rd Delta table. The job is being run with the Databricks service on GCP.Sometimes the job fails with the following exception...

  • 8073 Views
  • 8 replies
  • 3 kudos
Latest Reply
RohanB
New Contributor III
  • 3 kudos

Hi @Jose Gonzalez​ ,Do you require any more information regarding the code? Any idea what could be cause for the issue?Thanks and Regards,Rohan

  • 3 kudos
7 More Replies
SCOR
by New Contributor II
  • 3382 Views
  • 3 replies
  • 4 kudos

SparkJDBC42.jar Issue ?

Hi there!I am using the SparkJDBC42.jar in my Java application to use my delta lake tables , The connection is made through databricks sql endpoint in where I created a database and store in it my delta tables. I have a simple code to open connection...

  • 3382 Views
  • 3 replies
  • 4 kudos
Latest Reply
jose_gonzalez
Databricks Employee
  • 4 kudos

Hi @Seifeddine SNOUSSI​ ,Are you still having issue or you were able to resolve this issue? please let us know

  • 4 kudos
2 More Replies
Kody_Devl
by New Contributor II
  • 3092 Views
  • 1 replies
  • 0 kudos

HTML Backup Import Into my Account

Hi AllI would like to Import my HTML notebook backup into my databricks account and use it as if it was my master (I am a developer and have many exported HTML backups that I may want to reuse. When you open an .HTML from backup, databricks has, ...

  • 3092 Views
  • 1 replies
  • 0 kudos
Latest Reply
jose_gonzalez
Databricks Employee
  • 0 kudos

Hi @Ross Crill​ ,Did @Kaniz Fatma​ reply helped you to resolve your question? please let us know

  • 0 kudos
Dunken
by New Contributor III
  • 6154 Views
  • 7 replies
  • 3 kudos

Resolved! Databricks and CD4ML

I would like to use Databricks in a CD4ML way (see also https://martinfowler.com/articles/cd4ml.html). Is this possible? I would like to develop and train models in one environment once qualified, I would like to deploy the model with the application...

  • 6154 Views
  • 7 replies
  • 3 kudos
Latest Reply
Atanu
Databricks Employee
  • 3 kudos

something below you are looking for @Armin Galliker​ ?

  • 3 kudos
6 More Replies
brickster_2018
by Databricks Employee
  • 6764 Views
  • 2 replies
  • 2 kudos
  • 6764 Views
  • 2 replies
  • 2 kudos
Latest Reply
MoJaMa
Databricks Employee
  • 2 kudos

Since Workflows (Multi-Task jobs) is now GA, one way to work around the 1000 concurrent jobs limit is to use tasks within a job. Each job can have 100 tasks, and these tasks do not count toward the concurrent job limit.

  • 2 kudos
1 More Replies
alejandrofm
by Valued Contributor
  • 4372 Views
  • 4 replies
  • 5 kudos

Resolved! Show Vacuum operation result (files deleted) without DRY RUN

Hi, I'm runing some scheduled vacuum jobs and would like to know how many files were deleted without making all the computation twice, with and without DRY RUN, is there a way to accomplish this?Thanks!

  • 4372 Views
  • 4 replies
  • 5 kudos
Latest Reply
RKNutalapati
Valued Contributor
  • 5 kudos

We have to enable logging to capture the logs for vacuum.spark.conf.set("spark.databricks.delta.vacuum.logging.enabled","true")

  • 5 kudos
3 More Replies
Oliver_Floyd
by Contributor
  • 3212 Views
  • 2 replies
  • 3 kudos

Resolved! How to update external metastore cluster configuration on the fly ?

Hello,In my use case, my data is pushed to an adls gen2 container called ingestAfter some data processing on a databricks cluster of the ingest workspace, I declare the associated table in an external metastore for this workspaceAt the end of this pr...

  • 3212 Views
  • 2 replies
  • 3 kudos
Latest Reply
Oliver_Floyd
Contributor
  • 3 kudos

Hello @Atanu Sarkar​ ,Thank you for your answer. I have created a feature request. I hope, it will be soon accepted ^^

  • 3 kudos
1 More Replies
Labels