- 44 Views
- 1 replies
- 0 kudos
Hello,When using /api/2.0/preview/sql/queries to list out all available queries, I noticed that certain queries were being shown while others were not. I did a small test on my home workspace, and it was able to recognize certain queries when I defin...
- 44 Views
- 1 replies
- 0 kudos
Latest Reply
Hi,How many queries were returned in the API call in question? The List Queries documentation describes this endpoint as supporting pagination with a default page size of 25, is that how many you saw returned?
Query parameters
page_size integer <= 10...
- 43 Views
- 1 replies
- 0 kudos
I'm tring to build a ETL pipeline in which I'm reading the jsonl files from the azure blob storage, then trying to transform and load it to delta tables in databricks. I have created the below schema for loading my data : schema = StructType([
S...
- 43 Views
- 1 replies
- 0 kudos
Latest Reply
Try this.Add option("multiline","true")
by
mh_db
• New Contributor II
- 44 Views
- 0 replies
- 0 kudos
I'm trying to connect to oracle server hosted in azure from AWS databricks notebook but seems the connection keeps timing out. I tested the connection IP using telnet <hostIP> 1521 command from another EC2 instance and that seems to reach the oracle ...
- 44 Views
- 0 replies
- 0 kudos
- 19 Views
- 0 replies
- 0 kudos
Hi,I have implemented the UNITY CATALOG with multinode cluster in databricks. The workspace instance profile with EC2 access is also created in IAM. but still having a challenge in sending emails from databricks using SES service.The same is working ...
- 19 Views
- 0 replies
- 0 kudos
by
MarkD
• New Contributor II
- 377 Views
- 8 replies
- 0 kudos
Hi,I'm trying to set a dynamic value to use in a DLT query, and the code from the example documentation does not work.SET startDate='2020-01-01';
CREATE OR REFRESH LIVE TABLE filtered
AS SELECT * FROM my_table
WHERE created_at > ${startDate};It is g...
- 377 Views
- 8 replies
- 0 kudos
Latest Reply
Hi @MarkD ,You may use set variable_name.var= '1900-01-01'to set the value of variable and in order to use the value of variable use ${automated_date.var} Example: set automated_date.var= '1800-01-01'
select * from my table where date = CAST(${autom...
7 More Replies
by
pshuk
• New Contributor III
- 104 Views
- 2 replies
- 1 kudos
Hi,I am using CLI to transfer local files to Databricks Volume. At the end of my upload, I want to create a meta table (storing file name, location, and some other information) and have it as a table on databricks Volume. I am not sure how to create ...
- 104 Views
- 2 replies
- 1 kudos
Latest Reply
Hi @pshuk , Greetings!
We understand that you are looking for a CLI command to create a Table but at this moment Databricks doesn't support CLI command to create the table but you can use SQL Execution API -https://docs.databricks.com/api/workspace/...
1 More Replies
- 42 Views
- 1 replies
- 0 kudos
Hi All,I am trying to create an external table from a Azure Blob storage container. I recieve no errors, but there is no data in the table. The Blob Storage contains 4 csv files with the same columns and about 10k rows of data. Am I missing someth...
- 42 Views
- 1 replies
- 0 kudos
Latest Reply
Hi, The code looks completely fine. please check if you have any other delimiter other than , .If your CSV files use a different delimiter, you can specify it in the table definition using the OPTIONS clause.Just to confirm I created a sample table a...
- 1645 Views
- 2 replies
- 0 kudos
Hi data experts.I currently have an OLTP (Azure SQL DB) that keeps data only for the past 14 days. We use Partition switching to achieve that and have an ETL (Azure data factory) process that feeds the Datawarehouse (Azure Synapse Analytics). My requ...
- 1645 Views
- 2 replies
- 0 kudos
Latest Reply
Hi @Kaniz I have looked at this topic extensively and have even tried to implement it.I am a champion of databricks at my organization, but I do not think that it currently enables the OLTP scenarios.The closest I have gotten to it is by using the St...
1 More Replies
by
dbal
• New Contributor III
- 518 Views
- 2 replies
- 0 kudos
- 518 Views
- 2 replies
- 0 kudos
Latest Reply
@dbal - can you please try withColumnsRenamed() instead
Reference: https://docs.databricks.com/en/release-notes/dbconnect/index.html#databricks-connect-1430-python
1 More Replies
- 620 Views
- 1 replies
- 0 kudos
Hi there is requirements to create a pipeline that calls api and store that data in datawarehouse. Can you suggest me the best way to do this
- 620 Views
- 1 replies
- 0 kudos
Latest Reply
Hi @Sushmg, Please refer to the Databricks documentation and resources for more detailed instructions and examples.
- 27 Views
- 0 replies
- 0 kudos
I've posted the same question on stackoverflow (link) as well. I will post any solution I get there.I was trying to understand using many columns in the IN query and came across this statement. SELECT (1, 2) IN (SELECT c1, c2 FROM VALUES(1, 2), (3, 4...
- 27 Views
- 0 replies
- 0 kudos
- 33 Views
- 1 replies
- 1 kudos
Hi all, Is there a way to pass a column name(not a value) in a parametrized Spark SQL query?I am trying to do it like so, however it does not work as I think column name get expanded like 'value' i.e. surrounded by single quotes: def count_nulls(df:D...
- 33 Views
- 1 replies
- 1 kudos
Latest Reply
Hi @StephanKnox , You can use string interpolation (f-strings) to dynamically insert the column name into your query.
- 209 Views
- 2 replies
- 0 kudos
I was reading the data type rules and found about least common type.I have a doubt. What is the least common type of STRING and INT? The referred link gives the following example saying the least common type is BIGINT.-- The least common type between...
- 209 Views
- 2 replies
- 0 kudos
Latest Reply
Hi @Dhruv-22, The concept of the least common type can indeed be a bit tricky, especially when dealing with different data types like STRING and INT.
Let’s dive into this and clarify the behaviour in Apache Spark™ and Databricks.
Coalesce Functi...
1 More Replies
- 3674 Views
- 10 replies
- 1 kudos
My Databricks Experiments is cluttered with a whole lot of experiments. Many of them are notebooks which are showing there for some reason (even though they didn't have an MLflow run associated with it). I would like to delete the experiments, but it...
- 3674 Views
- 10 replies
- 1 kudos
Latest Reply
Hey @Debayan @SparkMaster
A bit late here, but I believe this is being caused by a click on the right side experiments icon. This may look like a meaningless click but it actually triggers a run.
9 More Replies
by
210227
• New Contributor III
- 66 Views
- 1 replies
- 0 kudos
Hi, I'm creating external table from existing external location and am a bit puzzled as to what permissions I need for it or what is the correct way of defining the S3 path with wildcards. This:create external table if not exists test_catalogue_dev.b...
- 66 Views
- 1 replies
- 0 kudos
Latest Reply
Just for the reference, the wildcard is not needed in this case, just a misleading error message. In this case 's3://test-data/full/2023/01/' instead of 's3://test-data/full/2023/01/*/' was the correct PATH