Here, we're trying to use the Python UDF inside the query.taking the table as function input converting the table into dataframe performing modification converting the dataframe into table returning the table How can we create spark context inside U...
Hi @doremon11, Unfortunately, creating a Spark context inside a UDF directly in the query is not possible. The Spark context is a global object and cannot be created within a UDF. UDFs are designed to operate on data within a DataFrame, not to create...
Hello. Please forgive me if this is not the right place to ask, but I'm having issues with databricks' statement execution api. I'm developing Haskell client for this api.I managed to implement most of it, but I'm running into issues with using named...
Ok, I thing I found what's behind this.This seems to be a bug in spark itself.Here's how I was able to reproduce it in spark-shell running in docker container:https://gist.github.com/jhrcek/49386004a9a47172649158af288106f4I think it's best if I just ...
Hello, We have a workspace in West Europe, which we were using with multiple Serverless SQL Warehouses. Recently it suddenly stopped working and presented with the following error when trying to restart an existing serverless warehouse. I am an adm...
Raised a ticket with Azure support but they haven't been very helpful so far to be honest. No uncleared bills and serverless seems disabled for all workspaces in the subscription. Someone mentioned accepting terms and conditions for serverless comp...
Hi Community!I have the requirement that business users shall be able to edit/update tables in Unity Catalog, e.g. master data records, mapping tables. I also want thes actions to be logged for auditing/troubleshooting.Is there any simple solution to...
We do have a few partners that offer solutions in this space (e.g. Retool). Recently, Sigma added support for their InputTable feature which was designed for this use case: https://www.sigmacomputing.com/blog/bring-your-own-data-to-databricks-with-s...
Hi All!Has anyone encountered a situation where we need to setup data access for Unity Catalog tables for read access such as external data marts, dashboard tools and etc.We are currently using Databricks to serve data to people in our organisation t...
Hi @dwfchu, Setting up data access for Unity Catalog tables in Databricks involves several considerations.
Let’s explore your options and weigh their pros and cons:
Personal Access Tokens (PATs):
Pros:
Quick and easy to set up.Developers can gen...
I see there’s a “test” capability within a DAB, but I’d like to know more about how this should/could be used. Does anyone know of any documentation or examples which might provide insights into its intended use?
Hi @EWhitley
You can check and validate whether the Asset Bundle configuration is valid or not by using the below command
databricks bundle validate
If a JSON representation of the bundle configuration is returned, then the validation succeed...
While setting up metastore in GCP Databricks, I added the bucket name and then service account permissons as well. Still my catalog dont have base root location. This deters me from creating table in my catalog. Root storage credential for metastore...
Hi @manish05485 , Good Day!
Error:Root storage credential for metastore XXXXXX does not exist. Please contact your Databricks representative or consider updating the metastore with a valid storage
Error states that the data access configuration for ...
Hi,I cannot see the query execution time in the response to the "api/2.0/sql/history/queries" request.Basically, I get only the following fields:{"next_page_token":...,"has_next_page":...,"res":[ { "query_id":..., "status":.., "query_tex...
I am attempting to recreate a legacy dashboard in Lakeview. The bar graph in no way resembles what I created in the SQL visualization editor. Lakeview has far fewer formatting options for one thing.How do I recreate the graph so that it resembles the...
Hi All,I've been using notebooks to run patch requests to an external API using the Python requests library. Often times certain notebooks will randomly start to fail throughout the day and will raise a ImportHookFinder.find_spec() error when attempt...
Hey there! Thanks a bunch for being part of our awesome community! We love having you around and appreciate all your questions. Take a moment to check out the responses – you'll find some great info. Your input is valuable, so pick the best solution...
I completed the Data Engineering Lakehouse course and I am familiar with different objects and concepts of databricks and lakehouse but I cant tie them together in my mind.Where do you typically use managed and non-managed tables? Bronze layer? Or no...
Hey there! Thanks a bunch for being part of our awesome community! We love having you around and appreciate all your questions. Take a moment to check out the responses – you'll find some great info. Your input is valuable, so pick the best solution...
Hi Team,I have a delta table in databricks which contains a encrypted column. For encrypting I am using databricks "aes_encrypt" function. For reference: https://docs.databricks.com/en/sql/language-manual/functions/aes_encrypt.html#aes_encrypt-functi...
you can create two cloumns , and display base on userDisplay different columns in Power BI based on logged in user | Paige Liu’s Posts (liupeirong.github.io)
We found a post on LinkedIn that revealed if "this table is deprecated" is added to a table comment, the table will appear with a strikethrough in notebooks and SQL editor windows. Is this functionality GA? Is there any documentation on the use of ...
Using workflows, is there a way to obtain the task name from within a task?EX: I have a workflow with a notebook task. From within that notebook task I would like to retrieve the task name so I can use it for a variety of purposes.Currently, we're re...
@EWhitley - could you please try using jobs API - /api/2.1/jobs/get and look for task_key fields to obtain all the task name within a given job.
Reference - https://docs.databricks.com/api/workspace/jobs/get
Databricks and Snowflake are both powerful platforms designed to address different aspects of data processing and analytics. Databricks shines in big data processing, machine learning, and AI workloads, while Snowflake excels in data warehousing, sto...