cancel
Showing results for 
Search instead for 
Did you mean: 
Community Platform Discussions
Connect with fellow community members to discuss general topics related to the Databricks platform, industry trends, and best practices. Share experiences, ask questions, and foster collaboration within the community.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

jenshumrich
by Contributor
  • 1834 Views
  • 2 replies
  • 0 kudos

Long running jobs get lost

Hello,I tried to schedule a long running job and surprisingly it does seem to neither terminate (and thus does not let the cluster shut down), nor continue running, even though the state is still "Running":But the truth is that the job has miserably ...

jenshumrich_0-1712742957610.png jenshumrich_2-1712743008070.png jenshumrich_3-1712743098546.png
  • 1834 Views
  • 2 replies
  • 0 kudos
Latest Reply
Lakshay
Esteemed Contributor
  • 0 kudos

Have you looked at the sql plan to see what the  spark job 72 was doing?

  • 0 kudos
1 More Replies
chari
by Contributor
  • 1309 Views
  • 3 replies
  • 0 kudos

Reading csv file with spark throws [insufficient privelage] error

Hello Community,I have some csv files saved in databricks workspace and want to read them with spark. I make use of the commanddf = spark.read.format('csv').load(r'filepath') However, it throws the error.org.apache.spark.SparkSecurityException: [INSU...

  • 1309 Views
  • 3 replies
  • 0 kudos
Latest Reply
Lakshay
Esteemed Contributor
  • 0 kudos

If this a UC enabled workspace, you need to provide the right access.

  • 0 kudos
2 More Replies
Ajay-Pandey
by Esteemed Contributor III
  • 2553 Views
  • 3 replies
  • 2 kudos

Resolved! Update regarding Community Reward Store

Hi Team,Is there any update on the Community Reward Store, as it's been discontinued from the old portal, and we still can't see the new portal for that.Is there any expected date when this will be available for community members?

  • 2553 Views
  • 3 replies
  • 2 kudos
Latest Reply
Ajay-Pandey
Esteemed Contributor III
  • 2 kudos

Thanks for update.

  • 2 kudos
2 More Replies
anonymous_567
by New Contributor II
  • 1270 Views
  • 3 replies
  • 0 kudos

Autoloader update table when new changes are made

Hello,Everyday a new file of the same name gets sent to my storage account with old and new data appended at the end.  Columns may also be added during one of these file updates. This file does a complete overwrite of the previous file. Is it possibl...

  • 1270 Views
  • 3 replies
  • 0 kudos
Latest Reply
data-grassroots
New Contributor III
  • 0 kudos

This may be helpful - the bit on allow overwritehttps://docs.databricks.com/en/ingestion/auto-loader/faq.html

  • 0 kudos
2 More Replies
Alexandru
by New Contributor III
  • 2353 Views
  • 3 replies
  • 0 kudos

Resolved! vscode python project for development

Hi,I'm trying to set up a local development environment using python / vscode / poetry. Also, linting is enabled (Microsoft pylance extension) and the python.analysis.typeCheckingMode is set to strict.We are using python files for our code (.py) whit...

  • 2353 Views
  • 3 replies
  • 0 kudos
Latest Reply
artsheiko
Honored Contributor
  • 0 kudos

Hi Alexandru, Take a look at VSCode extension for Databricks : https://marketplace.visualstudio.com/items?itemName=databricks.databricks 

  • 0 kudos
2 More Replies
Hogan
by New Contributor II
  • 977 Views
  • 1 replies
  • 0 kudos

Can browse external Storage, but can not create a Table from there - VNET, ADLSGen2

Hi there!Hope somebody here can help me. We have created a new Databricks Account on Azure with the ARM template for VNET injection.We have all the subnets etc., unitiy catalog active and the connector for databricks.I want now to create my first tab...

  • 977 Views
  • 1 replies
  • 0 kudos
Latest Reply
Hogan
New Contributor II
  • 0 kudos

Hi,To solve this problem, the following Microsoft documentation can be used to configure the NCC to enable the connection between the private Azure storage and the serverless resources.https://learn.microsoft.com/en-us/azure/databricks/security/netwo...

  • 0 kudos
sai_sathya
by New Contributor III
  • 2170 Views
  • 6 replies
  • 1 kudos

DataFrame to CSV write has issues due to multiple commas inside an row value

Hi alliam working on a data containing JSON fields with embedded commas into CSV format. iam facing challenges due to the commas within the JSON being misinterpreted as column delimiters during the conversion process.i tried several methods to modify...

sai_sathya_0-1712850570456.png sai_sathya_1-1712850991923.png
  • 2170 Views
  • 6 replies
  • 1 kudos
Latest Reply
artsheiko
Honored Contributor
  • 1 kudos

Hi Sai, I assume that the problem comes not from the PySpark, but from Excel. I tried to reproduce the error and didn't find the way - that a good thing, right ? Please try the following :    df.write.format("csv").save("/Volumes/<my_catalog_name>/<m...

  • 1 kudos
5 More Replies
Nithya_r
by New Contributor II
  • 863 Views
  • 1 replies
  • 0 kudos

Access Delta sharing from Azure Data Factory

I recently got access to delta sharing and I am looking to access the data from the tables in share through ADF. I used linked services such as REST API and HTTP and successfully established connection using the credential file token and http path, h...

  • 863 Views
  • 1 replies
  • 0 kudos
Latest Reply
artsheiko
Honored Contributor
  • 0 kudos

Hey, I think you'll need to use a Databricks activity instead of Copy See : https://learn.microsoft.com/en-us/azure/data-factory/connector-overview#integrate-with-more-data-storeshttps://learn.microsoft.com/en-us/azure/data-factory/transform-data-dat...

  • 0 kudos
databird
by New Contributor II
  • 1783 Views
  • 4 replies
  • 1 kudos

Redefine ETL strategy with pypskar approach

Hey everyone!I've some previous experience with Data Engineering, but totally new in Databricks and Delta Tables.Starting this thread hoping to ask some questions and asking for help on how to design a process.So I have essentially 2 delta tables (sa...

  • 1783 Views
  • 4 replies
  • 1 kudos
Latest Reply
artsheiko
Honored Contributor
  • 1 kudos

Hi @databird , You can review the code of each demo by opening the content via "View the Notebooks" or by exploring the following repo : https://github.com/databricks-demos (you can try to search for "merge" to see all the occurrences, for example) T...

  • 1 kudos
3 More Replies
vinay076
by New Contributor III
  • 1229 Views
  • 2 replies
  • 0 kudos

There is no certification number in my Databricks certificate that i had received after passing the

I enrolled myself for the Databricks data engineer certification recently and gave a shot at the exam and i did clear it successfully. I have received the certificate in the form of a pdf file along with a URL in which i can see my certificate and ba...

  • 1229 Views
  • 2 replies
  • 0 kudos
Latest Reply
Cert-Team
Esteemed Contributor
  • 0 kudos

Hi @vinay076 Thanks for asking! Our support team can provide you with a credential ID. Please file a ticket with our support team, give them your email associated with your certification, and they can get you the credential ID.

  • 0 kudos
1 More Replies
VabethRamirez
by New Contributor II
  • 3478 Views
  • 5 replies
  • 4 kudos

Resolved! How obtain a list of workflows in Databricks?

I need to obtain a list of my Databricks workflows with their job IDs in a notebook Databricks

  • 3478 Views
  • 5 replies
  • 4 kudos
Latest Reply
artsheiko
Honored Contributor
  • 4 kudos

Hi @VabethRamirez , Also, instead of using directly the API, you can use databricks Python sdk :  %pip install databricks-sdk --upgrade dbutils.library.restartPython()from databricks.sdk import WorkspaceClient w = WorkspaceClient() job_list = w.jobs...

  • 4 kudos
4 More Replies
RahulChaubey
by New Contributor III
  • 1096 Views
  • 2 replies
  • 0 kudos

Can api for query history /api/2.0/sql/history/queries return data which is older than 30 days?

I am using this api but it is returning the data for only last 30 days. Can this api return data which is older than 30 days?

  • 1096 Views
  • 2 replies
  • 0 kudos
Latest Reply
artsheiko
Honored Contributor
  • 0 kudos

Hi @RahulChaubey, The query history system table was announced during the Q1 roadmap webinar (see the recording, 32:25). There is a chance that it will provide data with a horizon beyond 30 days. Meanwhile, you can enable system tables - I hope some ...

  • 0 kudos
1 More Replies
QPeiran
by New Contributor III
  • 1401 Views
  • 3 replies
  • 0 kudos

Does Delta Table can be the source of streaming/auto loader?

Hi,Since the Auto Loader only accept "append-only" data as the source, I am wondering if the "Delta Table" can also be the source.Does VACCUM(deleting stale files) or _delta_log(creating nested and different file format than parquet) going to break A...

  • 1401 Views
  • 3 replies
  • 0 kudos
Latest Reply
artsheiko
Honored Contributor
  • 0 kudos

Hi @QPeiran, Auto-loader is a feature that allows to integrate files into the Data Platform. Once your data is stored into the Delta Table, you can rely on spark.readStream.table("<my_table_name>") to continuously read from the table. Take a look at ...

  • 0 kudos
2 More Replies
alano
by New Contributor
  • 602 Views
  • 1 replies
  • 0 kudos

Handling large volumes of streamed transactional data using DLT

We have a data stream from event hub with approximately 10 million rows per day (into one table) - these records are insert only (no update). We are trying to find a solution to aggregate / group by the data based on multiple data points and our requ...

  • 602 Views
  • 1 replies
  • 0 kudos
Latest Reply
artsheiko
Honored Contributor
  • 0 kudos

Hi, please find below a set of resources I believe relevant for you. Success stories You can find the success stories of companies leveraging the streaming on Databricks here. Videos Introduction to Data Streaming on the Lakehouse : Structured Stream...

  • 0 kudos
chemajar
by New Contributor III
  • 2534 Views
  • 3 replies
  • 1 kudos

Resolved! Rearrange tasks in databricks workflow

Hello,There is anyway to rearrange tasks in databricks workflow?.I would like that line that join the two marked tasks doesn't pass behind the other tasks.  It is posible that this line by one side?Thanks. 

image.png
  • 2534 Views
  • 3 replies
  • 1 kudos
Latest Reply
artsheiko
Honored Contributor
  • 1 kudos

Hi @chemajar, Take a look at Databricks Asset Bundles. It allows you to streamline the development of complex workflows using a yaml definition. In case you need to change the task dependencies, you can rearrange the flow as you need just change the ...

  • 1 kudos
2 More Replies

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.

If there isn’t a group near you, start one and help create a community that brings people together.

Request a New Group
Top Kudoed Authors