Databricks Certification and Badge not received
Hello, I passed DEA-C01 exam on March 25, 2025. I haven't received my certification or credentials yet. Can you please check?
- 57 Views
- 0 replies
- 0 kudos
Hello, I passed DEA-C01 exam on March 25, 2025. I haven't received my certification or credentials yet. Can you please check?
I would like to know if there is a Databricks slack channel? And if not, would anyone be interested in joining one?
I joined with same email and it says try different emails. How do I join?
I am trying to run my spark-dependent application on a Databricks cluster:spark-3.3.2on a cluster with Databricks Runtime 12.2 LTS (spark-3.3.2, scala-2.12)I end up with:NoSuchMethodError: org.apache.spark.sql.execution.datasources.parquet.ParquetToS...
I ran into the same issueDid you happen to find any fix?
Every time I try to create a cluster over the API I get this. Even when I am literally using the example JSON from the API documentation:```{ "cluster_name": "single-node-cluster", "node_type_id": "i3.xlarge", "spark_version": "7.6.x-scala2...
Hi, Matt. Would be super useful if you could provide a more complete code example on how it worked.
Hello,I am trying to use an init script from my repo workspace. The cluster is a 'personal compute' cluster with single user access mode with the following runtime version: '13.1 (includes Apache Spark 3.4.0, Scala 2.12)' I have 'Files in Repos' set ...
Hi, could you please enable init script logging and provide the elaborated logs here? Looks like it is not able to get the package from the repo. https://docs.databricks.com/clusters/init-scripts.html#init-script-loggingAlso, please tag @Debayan Mukh...
Dear support,after 48h I didn't receive the Accreditation Badge for AZURE Databricks Platform Architect. The id is E-02KZJV. Could you check please? I completed the "Azure Databricks Platform Architect Accreditation" learing plan and passed with >80%...
Hi @NIKOLAOS BERDOS I assume you will receive the ticket number since you have initiated the ticket. That means your ticket has been raised.Thanks and Regards
I have one requirement where I need to delete older data from a folder, and folder data can be partitioned or without partition. What approach to follow to build this utility?
Hi @Vidula Khanna @Kaniz Fatma It does not answer my real question. We mount adls path in databricks where we read and write the data. Suppose I have one partitioned (partitioned on audit) folder like below. I want to build an utility to delete the...
What if i delete the delta table in the 10th transaction(Checkpoint) will i be able to recover the table through time travel
Hi @Yatharth Malhotra Thank you for posting your question in our community! We are happy to assist you.To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best ans...
Hi @stephane cochet Thank you for posting your question in our community! We are happy to assist you.To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best answe...
Hi,Running on Databricks-AWS, I have a job running on a cluster with 3 workers, 2-cores each (r6i.large), with autoscaling enabled.The Spark job has two stages: (1) highly parallelizable, cpu-intensive stage. This stage takes 15 minutes.(2) a non-par...
Hi @Yoav Ben Great to meet you, and thanks for your question! Let's see if your peers in the community have an answer to your question. Thanks.
When executing the same Notebook from Repos and Workspace, the current path (`os.getcwd()`) is different:Repos=/Workspace/Repos/USERID/REPONAME.ide/pricingWorkspace=/home/spark-1636b562-0ce7-473b-bb17-56The issue is the Databricks VScode extension re...
Found a work around:``` Pythonimport osnotebook_dir = os.path.dirname(dbutils.notebook.entry_point.getDbutils().notebook().getContext().notebookPath().get())notebook_dir = "/Workspace" + notebook_dircurrent_dir = os.getcwd()if current_dir != notebook...
Ibis now includes pyspark backend. However using the databricks-connect doesn't seem to work with Ibis. Here's the sample code the throws an error.from databricks.connect import DatabricksSessionfrom databricks.sdk.core import Configconfig = Config(p...
Hi Khanna, No solution provided yet. Can escalate with your technical team for a suitable solution please?Thanks,Calvin
Hi all,I am not able to access the delta table from the datbase.While I try to read the table via spark.read.table command, I get the below error:org.apache.spark.sql.AnalysisException: org.apache.hadoop.hive.ql.metadata.HiveException: java.lang.Run...
Hi @Karthikeyan Palanisamy I hope you are doing great!Sorry for the inconvenience. Please specify that the issue is related to Databricks Community or Community -Edition or share the screenshot. Kindly, do let me know on a priority basis.Have a grea...
Hello Everyone!So we have a dev environment in Databricks and want to migrate it to prod. I need it o go over every single table, schema, notebooks, and artifacts in the databricks and make sure nothing is hard-coded for example or that there is noth...
Hi @Parsa Bahraminejad Thank you for posting your question in our community! We are happy to assist you.To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best an...
Hi,I just wanted to connect Databricks with SSAS multidimensional Project/want to use Databricks connection string and provider. Is that possible to do in SSAS multidimensional Project?Which Provider I need to use?Or Is there any other workaround to ...
Hi @Mehala M We haven't heard from you since the last response from @Werner Stinckens , and I was checking back to see if her suggestions helped you.Or else, If you have any solution, please share it with the community, as it can be helpful to oth...
Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!
Sign Up Now