cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 
Data + AI Summit 2024 - Data Engineering & Streaming

Forum Posts

Anonymous
by Not applicable
  • 699 Views
  • 1 replies
  • 2 kudos

Join our Community Social Group and Never Miss a Beat! Are you looking to connect with like-minded individuals and stay on top of the latest news and ...

Join our Community Social Group and Never Miss a Beat!Are you looking to connect with like-minded individuals and stay on top of the latest news and events in your community? Look no further than our special group on Community called the "Community S...

  • 699 Views
  • 1 replies
  • 2 kudos
Latest Reply
Ajay-Pandey
Esteemed Contributor III
  • 2 kudos

@Rishabh Pandey​ 

  • 2 kudos
Data_Analytics1
by Contributor III
  • 9860 Views
  • 5 replies
  • 0 kudos

"Waiting to run" status of the cell

Whenever I tried to execute the cell, it is not executing and it says waiting to run and gets stuck there. I tried this with different clusters but the problem still persist.What should be the standard process if I create a new cluster and wants to a...

  • 9860 Views
  • 5 replies
  • 0 kudos
Latest Reply
NandiniN
Honored Contributor
  • 0 kudos

Hello @Mahesh Chahare​ , The "Waiting to run" message appears usually when the cluster start or library installation is in progress. If you are running the commands on an interactive notebook, and you have clicked on "Run All", the commands will get ...

  • 0 kudos
4 More Replies
YogeshS
by New Contributor II
  • 1824 Views
  • 5 replies
  • 1 kudos

Waiting Waiting Waiting.......Databricks voucher not received

Hello,I have attended the webinar Lakehouse Fundamentals Training in Feb 2023then Completed the Databricks Lakehouse fundamentals accreditation and submitted the survey.As per communication it is expected that I will receive Databricks Certification ...

  • 1824 Views
  • 5 replies
  • 1 kudos
Latest Reply
Anonymous
Not applicable
  • 1 kudos

Hi @YOGESH SINGH​ I'm sorry you could not find a solution to your problem in the answers provided.Our community strives to provide helpful and accurate information, but sometimes an immediate solution may only be available for some issues.I suggest p...

  • 1 kudos
4 More Replies
seboz123
by New Contributor II
  • 2177 Views
  • 3 replies
  • 0 kudos

Display Html from dbfs files

Hi,I want to display some content from dbfs inside my notebook. Let's say I have a image under: /dbfs/mnt/test-bucket/test-custom/sample.pngI want to embed that into my Notebook Html Output like this:displayHTML("""<img src ='/dbfs/mnt/test-bucket/te...

  • 2177 Views
  • 3 replies
  • 0 kudos
Latest Reply
seboz123
New Contributor II
  • 0 kudos

Hi @Vidula Khanna​ ,unfortunately not. I can access the file through the notebook via e.g. !ls /dbfs/mnt/test-bucket/test-custom/ but it cannot be displayed via the displayHTML, with the 401.

  • 0 kudos
2 More Replies
kll
by New Contributor III
  • 4499 Views
  • 2 replies
  • 0 kudos

Access DataBricks file system and transfer files

Is there a solution to access the files in DataBricks file system and transfer them to another directory, local or elsewhere. In other words, is there a FileZilla type solution? where can i find instructions to ssh into the cluster from my mac termin...

  • 4499 Views
  • 2 replies
  • 0 kudos
Latest Reply
Debayan
Esteemed Contributor III
  • 0 kudos

Hi, There are different options around this. You can start with https://docs.databricks.com/files/index.html. Please let us know if this helps. Also, please tag @Debayan​ with your next response which will notify me. Thank you!

  • 0 kudos
1 More Replies
Sorush
by New Contributor II
  • 2474 Views
  • 3 replies
  • 1 kudos

Issue with VS Code extension repo.

I successfully installed the extension and connected it to my databricks account. But when I try to select the repo (which already exists under repos in my databricks account) for syncing , I don't see it. My company uses Azure Devops (Git repo) as s...

  • 2474 Views
  • 3 replies
  • 1 kudos
Latest Reply
Debayan
Esteemed Contributor III
  • 1 kudos

Hi, was the repo already added?You can check on this : https://learn.microsoft.com/en-us/azure/databricks/repos/git-operations-with-reposAlso, please tag @Debayan Mukherjee​ with your next response which will notify me. Thank you!

  • 1 kudos
2 More Replies
alejandrofm
by Valued Contributor
  • 2662 Views
  • 2 replies
  • 1 kudos

Understand if the configs I use to SparkSession.builder still make sense for Databricks 10+

Hi! I currently have this as an old generic template with amends over time to optimize Databricks Spark execution, can you help me to know if this still makes sense for v10-11-12 or if there are new recommendations? Maybe some of this is making my pr...

  • 2662 Views
  • 2 replies
  • 1 kudos
Latest Reply
Anonymous
Not applicable
  • 1 kudos

@Alejandro Martinez​ :Hi! Your template seems to be a good starting point for configuring a SparkSession in Databricks. However, there are some new recommendations that you can consider for Databricks runtime versions v10-11-12. Here are some suggest...

  • 1 kudos
1 More Replies
Gilg
by Contributor II
  • 3117 Views
  • 1 replies
  • 0 kudos

Adding column as StructType

Hi Team,Just wondering, how can I add a column to an existing table.I'd tried the below script but giving me an error:ParseException: [PARSE_SYNTAX_ERROR] Syntax error at or near '<'(line 1, pos 121)ALTER TABLE table_clone ADD COLUMNS col_name1 STRUC...

  • 3117 Views
  • 1 replies
  • 0 kudos
Latest Reply
Anonymous
Not applicable
  • 0 kudos

@Gil Gonong​ :In Databricks, you can add a column to an existing table using the ALTER TABLE statement in SQL. Here is an example:ALTER TABLE table_clone ADD COLUMN col_name1 STRUCT< type: STRING, values: ARRAY<STRING> >Note that you need to ...

  • 0 kudos
Retko
by Contributor
  • 5530 Views
  • 1 replies
  • 1 kudos

Error when using SAS token to connect to azure Storage Account: Unable to load SAS token provider class: java.lang.IllegalArgumentException

Hi, I am trying to connect to the Storage Account using the SAS token, and receive this error: Unable to load SAS token provider class: java.lang.IllegalArgumentException - more on the picture.I couldnt find anything on the web for this error.I also ...

image.png
  • 5530 Views
  • 1 replies
  • 1 kudos
Latest Reply
Anonymous
Not applicable
  • 1 kudos

@Retko Okter​ :It seems that there is an issue with the SAS token provider class. This error can occur when the SAS token is not correctly formatted or is invalid.Here are some steps you can try to resolve the issue:Verify that the SAS token is corre...

  • 1 kudos
kll
by New Contributor III
  • 7235 Views
  • 1 replies
  • 1 kudos

plotly express choropleth map not rendering in jupyter notebook

I have the following code which should render a choropleth map. import plotly.express as px import geopandas as gpd   # Example GeoJSON file with polygon geometries geojson_file = 'example.geojson'   # Read GeoJSON file into GeoDataFrame *** = gpd.re...

  • 7235 Views
  • 1 replies
  • 1 kudos
Latest Reply
Anonymous
Not applicable
  • 1 kudos

@Keval Shah​ :There could be several reasons why the choropleth map is not rendering in your Jupyter notebook. Here are a few things you could try:Check that the GeoJSON file is loaded correctly: Make sure that the GeoDataFrame has been loaded correc...

  • 1 kudos
624398
by New Contributor III
  • 3666 Views
  • 7 replies
  • 0 kudos

is there a read only option in jdbc driver?

Is there a "read only" option when using databricks sql using jdbc driver?I'm looking for an equivalent to this:https://docs.aws.amazon.com/redshift/latest/mgmt/jdbc20-configuration-options.html#jdbc20-readonly-optionThanks!

  • 3666 Views
  • 7 replies
  • 0 kudos
Latest Reply
Vartika
Moderator
  • 0 kudos

Hi @Nativ Issac​ Hope everything is going great.Just wanted to check in if you were able to resolve your issue. If yes, would you be happy to mark an answer as best so that other members can find the solution more quickly? If not, please tell us so w...

  • 0 kudos
6 More Replies
shelly
by New Contributor
  • 1191 Views
  • 2 replies
  • 0 kudos

take() ooperation is throwing error

Traceback (most recent call last): File "/usr/local/spark/python/pyspark/serializers.py", line 458, in dumps return cloudpickle.dumps(obj, pickle_protocol) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/spark/python/pyspa...

  • 1191 Views
  • 2 replies
  • 0 kudos
Latest Reply
Anonymous
Not applicable
  • 0 kudos

@Shelly Bhardwaj​ :The error message you provided seems to be incomplete, as it only shows the traceback of a serialization error. Can you provide the full error message or describe the issue in more detail?Regarding the code you provided, it looks c...

  • 0 kudos
1 More Replies
KayCon86
by New Contributor
  • 2266 Views
  • 3 replies
  • 0 kudos

Creating a Api links by url & list from a saved df

I have 106,000 + api's I need to call, so instead of calling them one by one I would like to create a loop as I have the list of location Id's which I've called from there api locations list and these will sit at the end of the url to get more info o...

  • 2266 Views
  • 3 replies
  • 0 kudos
Latest Reply
Anonymous
Not applicable
  • 0 kudos

@Kay Connolly​ :It looks like you are trying to concatenate a string with a column object, which is causing the error. You need to convert the column object to a string first before concatenating it to the URL. Here's a modified code snippet that sho...

  • 0 kudos
2 More Replies
Anonymous
by Not applicable
  • 1497 Views
  • 2 replies
  • 2 kudos
  • 1497 Views
  • 2 replies
  • 2 kudos
Latest Reply
Anonymous
Not applicable
  • 2 kudos

@ppatel:If you are using insertInto with overwrite=True on a Hive external table in PySpark, it might not work as expected. This is because Hive external tables are not managed by Hive and the table data is stored externally. When you use overwrite=T...

  • 2 kudos
1 More Replies
maartenvr
by New Contributor III
  • 21931 Views
  • 9 replies
  • 2 kudos

Resolved! Unable to clear cache using a pyspark session

Hi all,I am using a persist call on a spark dataframe inside an application to speed-up computations. The dataframe is used throughout my application and at the end of the application I am trying to clear the cache of the whole spark session by calli...

  • 21931 Views
  • 9 replies
  • 2 kudos
Latest Reply
maartenvr
New Contributor III
  • 2 kudos

No solution yet:Hi @Suteja Kanuri​ ,Thank you for thinking along and replying!Unfortunately, I have not found a solution yet.I am getting an error that there exists no ```.getCache()``` method on a spark context. Also note that I have tried to do som...

  • 2 kudos
8 More Replies

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.

If there isn’t a group near you, start one and help create a community that brings people together.

Request a New Group
Labels