cancel
Showing results for 
Search instead for 
Did you mean: 
Warehousing & Analytics
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

SivaPK
by New Contributor II
  • 1469 Views
  • 3 replies
  • 0 kudos

How to find the distinct count of the below listed result from the table?

Hi,How to get the distinct count from the below listed image,keywords = column nametable = appCatalogkeywords (column)"[""data"",""cis"",""mining"",""financial"",""pso"",""value""]""[""bzo"",""employee news"",""news""]""[""core.store"",""fbi""]""[""d...

  • 1469 Views
  • 3 replies
  • 0 kudos
Latest Reply
Kaniz
Community Manager
  • 0 kudos

Hi @SivaPK , Try this :- # Assuming you have a DataFrame named 'df' with a column 'keywords'from pyspark.sql.functions import explode, col # Explode the array column 'keywords'exploded_df = df.select(explode(col("keywords")).alias("keyword")) # Count...

  • 0 kudos
2 More Replies
Amy
by New Contributor II
  • 2415 Views
  • 5 replies
  • 2 kudos

Gantt chart in dashboard

I'd like to create Gantt charts using the dashboard function. It seems like this could be possible by adding some additional parameters in the bar plot functionality, but I don't see how to do it currently (if there is a way, would love an example!)....

Screen Shot 2022-06-23 at 1.48.05 PM
  • 2415 Views
  • 5 replies
  • 2 kudos
Latest Reply
alexiswl
Contributor
  • 2 kudos

Hi @Vidula,I don't think this has been resolved.  I think gantt charts would look fantastic in a Lakeview Dashboard.  

  • 2 kudos
4 More Replies
bradleyjamrozik
by New Contributor III
  • 503 Views
  • 1 replies
  • 0 kudos

Server ODBC Connection

Is there a preferred method for hosting an odbc connection to a warehouse on a server for use by a report server (SSRS/PBIRS)? I know the odbc driver doesn't support pass-through authentication, so is there a way to configure it with an unattended ac...

  • 503 Views
  • 1 replies
  • 0 kudos
Latest Reply
Kaniz
Community Manager
  • 0 kudos

Hi @bradleyjamrozik , Certainly! Configuring an ODBC connection for a report server (such as SSRS or PBIRS) involves several considerations.    Let’s explore some options:   Direct ODBC Connection: You can create an ODBC data source on the server whe...

  • 0 kudos
Avin_Kohale
by New Contributor
  • 11590 Views
  • 5 replies
  • 2 kudos

Import python files as modules in workspace

I'm deploying a new workspace for testing the deployed notebooks. But when trying to import the python files as module in the newly deployed workspace, I'm getting an error saying "function not found".Two points to note here:1. If I append absolute p...

  • 11590 Views
  • 5 replies
  • 2 kudos
Latest Reply
TimReddick
New Contributor III
  • 2 kudos

Hi @Kaniz, I see your suggestion to append the necessary path to the sys.path. I'm curious if this is the recommendation for projects deployed via Databricks Asset Bundles. I want to maintain a project structure that looks something like this:project...

  • 2 kudos
4 More Replies
tranbau
by New Contributor
  • 287 Views
  • 0 replies
  • 0 kudos

Dynamic Spark Structured Streaming: Handling Stream-Stream Joins with Changing

I want to create a simple application using Spark Structured Streaming to alert users (via email, SMS, etc.) when stock price data meets certain requirements.I have a data stream: data_streamHowever, I'm strugging to address the main issue: how users...

Warehousing & Analytics
kafka
spark
spark-structured-streaming
stream-stream join
  • 287 Views
  • 0 replies
  • 0 kudos
Zer
by New Contributor II
  • 1749 Views
  • 5 replies
  • 4 kudos

SQL Editor, Tab Stops Indenting

Greetings,I use Databricks through Azure. Frequently when I'm working in the SQL Editor, the tab button fails to indent-- instead, it forces my cursor to a seemingly random part of the page. It's been doing this since when I started working in the pl...

  • 1749 Views
  • 5 replies
  • 4 kudos
Latest Reply
HannesM
New Contributor II
  • 4 kudos

Same issue here, sometimes it works by selecting a single but complete line and then hitting tab. If it works, then indentation works on multiple lines again as well. However, the single line select doesn't always work either. Pretty inconsistent beh...

  • 4 kudos
4 More Replies
Mat
by New Contributor III
  • 5430 Views
  • 5 replies
  • 3 kudos

Connect to Databricks SQL Endpoint using Programming language

Hi, I would like to know whether there is a feasibility/options available to connect to databricks sql endpoint using a programming language like java/scala/c#. I can see JDBC URL, but would like to whether it can be considered as any other jdbc conn...

  • 5430 Views
  • 5 replies
  • 3 kudos
Latest Reply
StephanieRivera
Valued Contributor II
  • 3 kudos

I found a similar question on Stackoverflow https://stackoverflow.com/questions/77477103/ow-to-properly-connect-to-azure-databricks-warehouse-from-c-sharp-net-using-jdb

  • 3 kudos
4 More Replies
cyong
by New Contributor II
  • 566 Views
  • 2 replies
  • 0 kudos

Dynamic measure calculation based on filter in Excel

Hi, currently we are using Power BI as the semantic layer because it allows us to build custom measures to do aggregates and business logic calculation, and provides native connection to Excel. I am thinking to move these logics to Databricks using S...

  • 566 Views
  • 2 replies
  • 0 kudos
Latest Reply
cyong
New Contributor II
  • 0 kudos

Thanks @Kaniz , I think Power Query can only perform pre-data transformations, not on-the-fly calculations in response to user filters.

  • 0 kudos
1 More Replies
scrimpton
by New Contributor II
  • 783 Views
  • 1 replies
  • 0 kudos

Resolved! Delta Sharing with Power BI

Using Delta Sharing connector with Power BI, does it only works for import and currently no support for direct query?

Warehousing & Analytics
DELTA SHARING
Power BI
  • 783 Views
  • 1 replies
  • 0 kudos
Latest Reply
karthik_p
Esteemed Contributor
  • 0 kudos

@scrimpton currently it only supports import https://learn.microsoft.com/en-us/power-query/connectors/delta-sharing

  • 0 kudos
Datbth
by New Contributor
  • 269 Views
  • 0 replies
  • 0 kudos

Cancel SQL statement using ODBC driver

Hi,I'm implementing a Databricks connector using the ODBC driver and currently working on the functionality to Cancel an ongoing SQL statement.However, I can't seem to find any ODBC function or SQL function to do so.The only other alternative I see i...

  • 269 Views
  • 0 replies
  • 0 kudos
ckwan48
by New Contributor III
  • 3987 Views
  • 5 replies
  • 8 kudos

Trying to connect to DBeaver from Databricks and getting error message

I am trying to connect to DBeaver from Databricks and getting this error message:[Databricks][DatabricksJDBCDriver](500593) Communication link failure. Failed to connect to server. Reason: javax.net.ssl.SSLHandshakeException: PKIX path building faile...

  • 3987 Views
  • 5 replies
  • 8 kudos
Latest Reply
Hardy
New Contributor III
  • 8 kudos

I have the same issue after upgrading cluster to DBR 12.2. Working fine with DBR 10.4

  • 8 kudos
4 More Replies
Yahya24
by New Contributor III
  • 1721 Views
  • 2 replies
  • 1 kudos

Resolved! API Query

Hello,I created a sql warehouse (cluster size = 2X-Small) and I wanted to use it to execute a query using the sql query api:- url : https://databricks-host/api/2.0/preview/sql/statements- params = {'warehouse_id': 'warehouse_id','statement': 'SELECT ...

  • 1721 Views
  • 2 replies
  • 1 kudos
Latest Reply
karthik_p
Esteemed Contributor
  • 1 kudos

@Yahya24 can you please remove preview in query, they are not in preview any more "/api/2.0/sql/statements/", you should see json response, can you please check drop down menu and change to json, some times it may be setted into text, but usual respo...

  • 1 kudos
1 More Replies
gmiguel
by Contributor
  • 2244 Views
  • 3 replies
  • 2 kudos

Resolved! Does "Merge Into" skip files when reading target table to find files to be touched?

I've been doing some testing with Partitions vs Z-Ordering to optimize the merge process.As the documentation says, tables smaller than 1TB should not be partitioned and can benefit from the Z-Ordering process to optimize the reading process.Analyzin...

  • 2244 Views
  • 3 replies
  • 2 kudos
Latest Reply
Kaniz
Community Manager
  • 2 kudos

Hi @gmiguel,  I've been testing with Partitions vs. Z-Ordering to optimize the merge process. As the documentation says, tables smaller than 1TB should not be partitioned and can benefit from the Z-Ordering process to optimize the reading process.Ana...

  • 2 kudos
2 More Replies
Mswedorske
by New Contributor II
  • 1216 Views
  • 1 replies
  • 2 kudos

Resolved! Historical Reporting

How do you handle reporting monthly trends within a data lakehouse?  Can this be done with timetravel to get the table state at the end of each month or is it better practice to build a data warehouse with SCD types?  We are new to databricks and lak...

  • 1216 Views
  • 1 replies
  • 2 kudos
Latest Reply
daniel_sahal
Esteemed Contributor
  • 2 kudos

@Mswedorske IMO it would be better to use SCD.When you do VACUUM on a table, it removes the data files that are necessary for Time Travel, so it's not a best choice to rely on Time Travel.

  • 2 kudos
Labels
Top Kudoed Authors