databricks billing
Is there a way to use a business checking account to pay for Databricks services?
- 97 Views
- 0 replies
- 0 kudos
Is there a way to use a business checking account to pay for Databricks services?
We are loading a data source that contains XML. I am translating their queries to create views in Databricks. They use 'XMLNAMESPACES' to construct/parse XML. Below is an example. What is best practice for translating 'XMLNAMESPACES' in Databricks?...
Hi @TinaN, To handle XMLNAMESPACES in Databricks, use the from_xml function for parsing XML data, where you can define namespaces within your parsing logic. Start by reading the XML data using spark.read.format("xml"), then apply the from_xml functio...
Hi,I have created this table which contains the data that I need for my source path and target table. source_path: /data/customer/sid={sid}/abc=1/attr_provider={attr_prov}/source_data_provider_code={src_prov}/So basically, the value of each row are c...
Hi @zll_0091, To efficiently load only the necessary files without manually iterating through each row of your table, you can use Spark's DataFrame operations. First, read your table into a DataFrame and determine the maximum key value. Then, filter ...
Hello Team, @Cert-Team , @Cert-TeamOPS I faced a very bad experience while attempting my 1st DataBricks certification.I was asked to exit the exam multiple times by the support team saying technical issues. My test got rescheduled multiple times with...
Hi @ozbieG, I'm sorry to hear your exam was suspended. Thank you for filing a ticket with our support team. Please allow the support team 24-48 hours to resolve. In the meantime, you can review the following documentation: Room requirements Behaviour...
Hi Databricks!I am a relatively new developer that's looking for a solid API testing tool. I am interested in hearing about other developers, new or experienced, about their experiences with API testing tools, regardless if they are good or bad. I've...
Hi @bytetogo,In my daily work I use Postman. It has user-friendly interface, supports automated testing and has support for popular patterns and libraries. It is also compatible with Linux, MacOs, Windows.
I am trying to install the stanza library and try to create a udf function to create NER tags for my chunk_text in the dataframe.Cluster Config: DBR 14.3 LTS SPARK 3.5.0 SCALA 2.12below code:def extract_entities(text import stanza nlp = stanza....
Was this resolved?I installed openai and tried to import and faced the below error. I also tried upgrading the libraries. However I ended up with the same error.Code: import openaiError: ImportError: cannot import name 'override' from 'typing_extensi...
I have been using visualization for a lot of different usecases and has been working for instead of using 3rd party libraries. Recently I had a need to customize the data labels but I haven't seen anything in the documentation that how to do that. If...
Hi,I'm trying to get the certain value of my variable in the for loop but it's returning the syntax instead of the value. Also, is it possible to covert this value to an integer? Thanks
Hi @zll_0091 ,Could you provide more code? What's inside dsfd variable? What's your expected outcome?
Hi!I am using bundles to deploy various workflows, and as part of a CI pipeline, I want to run integration tests. They are in notebooks that I deploy and run with databricks CLI bundle with Azure DevOps.This allows me to run only one job at a time as...
Thank you so much for the information.
Hi,I have registered for Community Edition and can access it with no problems trough: https://community.cloud.databricks.com/login.htmlNow, I'm interested in completing the free "lakehouse fundamentals" training here and taking the quiz to get the ba...
Hi @slechtd, To expedite your request, please list your concerns on our ticketing portal. Our support staff would be able to act faster on the resolution (our standard resolution time is 24-48 hours).
Hi,Is there a "Find and replace" option to edit SQL code? I am not referring to the "replace" function but something similar to Control + shift + F in Snowflake or Control + F in MS Excel.
Hi @abueno , Thanks for reaching out! Please review the response and let us know if it answers your question. Your feedback is valuable to us and the community. If the response resolves your issue, kindly mark it as the accepted solution. This will h...
I need to get the notebook owner using api or some other way by passing notebook path as parameter.
@Ayushi_Suthar Can you advise whether it's possible to make a request to add this feature? It would be incredibly useful for maintaining and auditing our notebooks
I use the following option to write from multiple tasks to the same table with overwrite (in Pyspark).option("partitionOverwriteMode", "dynamic")The table was created with partition by so it works as expected.I read about liquid clustering and it's b...
No it doesn't support partition overwrite.
I created a databricks JDBC connection class following the code ##https://docs.databricks.com/en/integrations/jdbc/authentication.htm I observed that after a number of execute SQL calls , there were 27000 instances of com.databricks.client.jdbc42.int...
Hi @Sameekshaji, Could you please verify that you’re using the latest version of the Databricks JDBC driver? Sometimes, bugs in older versions can cause resource leaks.
I'm getting the following error when trying to install UCX on my local machine using the databricks CLI v0.212.1. I've authenticated the CLI with the required workspace using databricks config with the host and PAT generated. I also have admin privil...
Hi @deane_h_s, Hi, Thank you for reaching out to our community! We're here to help you. To ensure we provide you with the best support, could you please take a moment to review the response and choose the one that best answers your question? Your fee...
Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.
If there isn’t a group near you, start one and help create a community that brings people together.
Request a New Group