I've triend this code in Databricks SQLcreate table people_db.GLAccount
USING PARQUET
LOCATION "abfss://datamesh@dlseu2dtaedwetldtlak9.dfs.core.windows.net/PricingAnalysis/rdv_60_134.vGLAccount.parquet"But I'm getting a "Invalid configuration v...
you can define 'data access configuration' in the admin panel.go to SQL warehouse settings -> Data Access configurationhttps://learn.microsoft.com/en-us/azure/databricks/sql/admin/data-access-configuration
I've been trying to check, via databricks SQL, which groups a user belongs to. When I run SHOW GROUPS or SHOW USERS, everything is returned correctly.However, when a try to run: SHOW GROUPS WITH USER `user@domain.com`Nothing is returned, but the u...
Hi @Marcelo Castro​ Hope all is well! Just wanted to check in if you were able to resolve your issue and would you be happy to share the solution or mark an answer as best? Else please let us know if you need more help. We'd love to hear from you.Tha...
It was working fine initially.Message: [Simba][SparkJDBCDriver](500618) Error occured while deserializing arrow data: sun.misc.Unsafe or java.nio.DirectByteBuffer.<init>(long, int) not available [SQL State=HY000, DB Errorcode=500618]
@Ehsan Ullah​ :The error message you received indicates that there is an issue with deserializing arrow data in the Spark JDBC driver. This error is caused by the fact that the sun.misc.Unsafe or java.nio.DirectByteBufferclasses are not available.To ...
Refreshing SQL DashboardYou can schedule the dashboard to automatically refresh at an interval.At the top of the page, click Schedule.If the dashboard already has a schedule, you see Scheduled instead of Schedule.Select an interval, such as Every 1 h...
I can see on Databricks SQL warehouse Data tab that clusters, catalogs and schemas have a unique ID. User created tables, views and functions must have and unique ID too, but it is not exposed to the user as far as I can tell.I need to retrieve the...
I am pulling data from Google BigQuery and writing it to a bronze table on an interval. I do this in a separate continuous job because DLT did not like the BigQuery connector calling collect on a dataframe inside of DLT.In Python, I would like to rea...
The below code is a solution. I was missing that I could read from a table with `spark.readStream.format("delta").table("...")`. Simple. Just missed it. This is different than `dlt.read_stream()` which appears in the examples a lot.This is referenced...
We've been using the DQBSQL API to perform CRUD on queries and alerts.Part of that process added a slack channel as alert destination using the /subscriptions element on an alert post as below.As of today I am getting a 410 'Gone' error from the API ...
Hi @Suteja Kanuri​ , @Vidula Khanna​ thanks for getting back with a solution.The suggested solution looks fine, but for a number of reasons I went with another option to use the Jobs API. This allowed me to preserve more of the automation I had alrea...
IPL Prediction is very challenging task. Due to the nature of the game, it is challenging to correctly predict the winner because even one player's performance can significantly affect the outcome. It is also challenging to forecast which club woul...
Hi FerArribas,I struggle with PrivateLink connection between Databricks account and my AWS account. I have seen that you had a similar problem. I can create a workspace, but when I try to create a cluster I get an error: "Cluster terminated. Reason: ...
Hi @Marcin Sieradzan​ Thank you for posting your question in our community! We are happy to assist you.To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best answ...
Hi ,I am trying to load power bi report in databricks,but i am getting..empty while printing report.. i have already installed pip install powerbiclient,and device_auth is also successfull,but report = Report(group_id=group_id, report_id=report_id, ...
Hi @Adarsh Singh​ Thank you for posting your question in our community! We are happy to assist you.To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best answers ...