- 6919 Views
- 6 replies
- 1 kudos
I know we have 4 keys max on cluster by () for both z-order and partition keys. I got some issues when adding 4 keys and 1 specific key triggers that error (I was not expecting as this is about Create Table) . Stats makes sense if you need to optimiz...
- 6919 Views
- 6 replies
- 1 kudos
Latest Reply
Is Liquid cluster also helpful for high cardinality columns for a table Join?
5 More Replies
- 1597 Views
- 2 replies
- 1 kudos
Hello Community,I’m encountering the following error while working on my project:Error: ${var.FILE_NAME} defined in 'include' section does not match any files.Has anyone faced this issue before? I'm using variables to include specific files, but it s...
- 1597 Views
- 2 replies
- 1 kudos
Latest Reply
Hi @AkhilVydyula ,You can start by updating your cli. I've seen some really weird errors with older cli versions.
1 More Replies
- 399 Views
- 0 replies
- 0 kudos
HelloI am bit surprised - but I don't see any default possibility of creating continous pipeline for fetching updates from MSSQL, and inserting them into delta lake - is it true - or am I missing something? To be clear - I know we can do that semi-li...
- 399 Views
- 0 replies
- 0 kudos
- 2279 Views
- 5 replies
- 5 kudos
Hello,I’m looking for study materials or resources to help me learn Databricks more effectively. Any recommendations would be greatly appreciated!
- 2279 Views
- 5 replies
- 5 kudos
Latest Reply
@igor779, @Rishabh-Pandey and @szymon_dybczak .Thank you all for the suggestions! Unfortunately, I don’t have access to content from partner companies and mainly rely on free resources for my studies.
4 More Replies
- 649 Views
- 0 replies
- 0 kudos
Hi!I have an problem with incrementally processing data within my Delta Live Tables (DLT) pipeline. I have a raw file (in Delta format) where new data is added each day. When I run my DLT pipeline I only want the new data to be processed. As an examp...
- 649 Views
- 0 replies
- 0 kudos
- 2578 Views
- 4 replies
- 0 kudos
Hello Databricks Team,I understand that the mask function can be used to mask columns, but I have a few questions:When users with access use a masked TABLE to create a downstream TABLE, the downstream TABLE does not inherit the mask function directly...
- 2578 Views
- 4 replies
- 0 kudos
Latest Reply
Hi @weilin0323, How are you doing today?As per my understanding, When creating a downstream table from a masked table, you’ll need to reapply the mask function to the necessary columns in the new table if you want to maintain the same level of data p...
3 More Replies
- 750 Views
- 1 replies
- 1 kudos
Hi There,We are planning to move to Unity Catalog soon so started replacing workspace API with Account Level APIs. One use case I have is getting a list of users, pretty straight forward with workspace:{workspaceUrl}api/2.0/preview/scim/v2/Users?coun...
- 750 Views
- 1 replies
- 1 kudos
Latest Reply
Hi @unseen007 ,To get a list of all users in a given workspace, you should use api/2.0/preview/scim/v2/Users api endpoint. Why do you assume this API is inappropriat? Unity Catalog has nothing to do with this. One API let's you list user at account l...
- 1081 Views
- 1 replies
- 1 kudos
I have a workflow with 11 tasks (each task executes one notebook) that run in sequence. The task was run on 9/1 and again today (9/10). I am working reporting task history and status using system table `system.lakeflow.job_task_run_timeline`. The...
- 1081 Views
- 1 replies
- 1 kudos
Latest Reply
Hi @DavidKxx ,There are chances that the table has not been updated today.Check tomorrow.Here what the documentation says:
by
abueno
• New Contributor III
- 2922 Views
- 2 replies
- 0 kudos
I am trying to exclude rows with a specific variable when querying using pyspark but the filter is not working. Similar to the "Not like" function in SQL. e.g. not like '%var4%'. The part of the code that is not working is: (col('col4').rlike('var...
- 2922 Views
- 2 replies
- 0 kudos
- 1192 Views
- 3 replies
- 2 kudos
Hello, one my teammate is trying to put some obversability on our Databricks flows.When he tries to contact our Open Telemetry server, he gets a timeout.I had a look, and the same command (on the same Databricks workspace) works well with runtime 12....
- 1192 Views
- 3 replies
- 2 kudos
Latest Reply
Hello,thank you for your answer.I tried to update library version and it was almost the solution.It realized it neither works in 12.2 but the output was pretty different and it misled me.Probably a network config to set up in the target server.
2 More Replies
by
lux13
• New Contributor
- 366 Views
- 0 replies
- 0 kudos
Hi everyone!I have successfully deployed the Review App of my AI agent following those instructions: Create and log AI agents | Databricks on AWSHowever, one question came up regarding the sources (names). To be precise, is there a possibility to sho...
- 366 Views
- 0 replies
- 0 kudos
- 6751 Views
- 10 replies
- 10 kudos
Hi has anyone else tried to include checkboxes in markdown cells in databricks notebook?I believe I followed the correct way for checkbox: - [ ] and - [x]but the result I got is still no checkboxes.Please help! Thanks! %md
#to do:
- [x] static vari...
- 6751 Views
- 10 replies
- 10 kudos
Latest Reply
Hi @teaholicI faced the same problem and found the ✓ and ✗ notation that did work for me. Hope that helps.
9 More Replies
by
prad18
• New Contributor III
- 3867 Views
- 11 replies
- 3 kudos
Hello Databricks Community,We are in the process of planning a Unity Catalog implementation for our organization, and I'd like to seek input on some architectural decisions. We're considering various approaches to workspace separation, storage accoun...
- 3867 Views
- 11 replies
- 3 kudos
Latest Reply
Hi @prad18 ,I'm glad the previous response was helpful! Let's address your remaining questions:Cost Differences Between Single vs. Multiple Azure Storage Accounts: The cost difference between using a single Azure storage account for both Unity Catalo...
10 More Replies
- 405 Views
- 0 replies
- 0 kudos
Hi, I'm downloading data in chunks using API api/2.0/preview/sql/queries Today I've realized hat, the chunks have different date.. For example results_2024-09-10T10_30_32Z_acf798b5-2a5c-474f-a3b0-b83a2a8eb35a.csv I'm in the UTC timezone, I know that...
- 405 Views
- 0 replies
- 0 kudos
- 7054 Views
- 9 replies
- 1 kudos
hi,I cannot install geopandas in my notebook, ive tried all different forms of generic fix, pip installs etc but always get this error:CalledProcessError: Command 'pip --disable-pip-version-check install geopandas' returned non-zero exit status 1.---...
- 7054 Views
- 9 replies
- 1 kudos
Latest Reply
@brian999 - Conda is subjected to commercial licensing. Referenced here: https://docs.databricks.com/en/archive/legacy/conda.html
8 More Replies