cancel
Showing results for 
Search instead for 
Did you mean: 
Warehousing & Analytics
Engage in discussions on data warehousing, analytics, and BI solutions within the Databricks Community. Share insights, tips, and best practices for leveraging data for informed decision-making.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

MadelynM
by New Contributor III
  • 68 Views
  • 0 replies
  • 0 kudos

[Recap] Data + AI Summit 2024 - Warehousing & Analytics | Improve performance and increase insights

Here's your Data + AI Summit 2024 - Warehousing & Analytics recap as you use intelligent data warehousing to improve performance and increase your organization’s productivity with analytics, dashboards and insights.  Keynote: Data Warehouse presente...

Screenshot 2024-07-03 at 10.15.26 AM.png
Warehousing & Analytics
AI BI Dashboards
AI BI Genie
Databricks SQL
  • 68 Views
  • 0 replies
  • 0 kudos
Ismail1
by New Contributor III
  • 1650 Views
  • 2 replies
  • 0 kudos

Migrating from Databases Postgres MySQL to Databricks.

Hi all, working on this project, my team plans to migrate some data from some databases to Databricks. We plan to run this migration by submitting queries to a warehouse through python on a local machine.Now I was wondering what would be the best app...

  • 1650 Views
  • 2 replies
  • 0 kudos
Latest Reply
Rom
New Contributor III
  • 0 kudos

Hi,Your solution it good.but if i'm in charge of this migration i will:create the architecture of all tables with ist constraints in databrick warehouseexport all data in tables of mysql database as csv of txt fileswrite notebook with pyspark code to...

  • 0 kudos
1 More Replies
EminBoz
by New Contributor II
  • 2064 Views
  • 4 replies
  • 1 kudos

"Revoke" permissions for SQL-Warehouse with API

Hello dear community,i am trying to revoke permissions with the API for SQL-Warehouse. Granting permissions isn't a problem and works like a charme. But revoking won't function. I tried "NO_PERMISSIONS", "NO PERMISSIONS", "DENY", "REVOKE" But i alway...

image.png
  • 2064 Views
  • 4 replies
  • 1 kudos
Latest Reply
Haris12
New Contributor II
  • 1 kudos

Cuphead APK is your go-to destination for the latest versions of the beloved game, Cuphead, on Android. We provide a curated selection of up-to-date APK files, ensuring that you can enjoy the thrilling adventures of Cuphead on your Android device has...

  • 1 kudos
3 More Replies
SaugatMukherjee
by New Contributor III
  • 1443 Views
  • 0 replies
  • 1 kudos

SQL Alert Email with QUERY_RESULTS_TABLE results in table with no border

Hi,I am sending databricks sql alerts to an email. I am trying to get the query results table in the body of the email.I have used a custom template with {{QUERY_RESULT_TABLE}}and this works fine for a teams alert. In Teams, I can see the table prope...

  • 1443 Views
  • 0 replies
  • 1 kudos
RobinK
by Contributor
  • 1527 Views
  • 2 replies
  • 1 kudos

Resolved! Use SQL Command LIST Volume for Alerts

Hi,we have implemented a Databricks Workflow that saves an Excel Sheet to a Databricks Volume. Now we want to notify users with an Alert, when new data arrives in the volume.In the docs I found the SQL command LIST which returns the columns path, nam...

  • 1527 Views
  • 2 replies
  • 1 kudos
Latest Reply
gabsylvain
New Contributor III
  • 1 kudos

Hi @RobinK , I've tested your code and I was able to reproduce the error. Unfortunately, I haven't found a pure SQL alternative to selecting the results of the LIST command as part of a subquery or CTE, and create an alert based on that.  Fortunately...

  • 1 kudos
1 More Replies
heymiky
by New Contributor
  • 778 Views
  • 0 replies
  • 0 kudos

Enabling HTML content in Dashboard Visuals

Hii'm seeking some help creating visuals using HTML in SQL queries similar to those in the Retail Revenue & Supply Chain sample dashboards.  When I create my queries based on these my results display the HTML code instead of the HTML formatted result...

  • 778 Views
  • 0 replies
  • 0 kudos
adisalj
by New Contributor II
  • 4847 Views
  • 4 replies
  • 2 kudos

TABLE_OR_VIEW_NOT_FOUND of deep clones

Hello community,We're cloning (deep clones) data objects of the production catalog to our non-production catalog weekly. The non-production catalog is used to run our DBT transformation to ensure we're not breaking any production models. Lately, we h...

  • 4847 Views
  • 4 replies
  • 2 kudos
Latest Reply
karthik_p
Esteemed Contributor
  • 2 kudos

@adisalj have a small question how you are handling deep cloned data in target, are you created managed table with data that is being clone into target. can you please post sample query that you are using between your catalogs to do deep clone.i am f...

  • 2 kudos
3 More Replies
Jennifer
by New Contributor III
  • 3193 Views
  • 2 replies
  • 0 kudos

Resolved! Why does readStream filter go through all records?

Hello,I am running spark structured streaming, reading from one table table_1, do some aggregation and then write results to another table. table_1 is partitioned by ["datehour", "customerID"]My code is like this:spark.readStream.format("delta").tabl...

  • 3193 Views
  • 2 replies
  • 0 kudos
Latest Reply
-werners-
Esteemed Contributor III
  • 0 kudos

To define the initial position please check this:https://learn.microsoft.com/en-us/azure/databricks/structured-streaming/delta-lake#specify-initial-position

  • 0 kudos
1 More Replies
nijhuist
by New Contributor II
  • 1743 Views
  • 2 replies
  • 0 kudos

Import warnings when running DBT as a package on Databricks 13.3LTS

Executing dbt as a Python package triggers about 200 import warnings when ran on Databricks Runtime 13.3 but not on 12.2. The warnings are all the same:  <frozen importlib._bootstrap>:914: ImportWarning: ImportHookFinder.find_spec() not found; fallin...

  • 1743 Views
  • 2 replies
  • 0 kudos
Latest Reply
Kaniz_Fatma
Community Manager
  • 0 kudos

Hi @nijhuist , Note: These are potential approaches; please test them in your specific environment. Cluster Version: Create your own cluster with the suggested runtime (if possible). Sometimes using a different cluster version can resolve such issues...

  • 0 kudos
1 More Replies
SivaPK
by New Contributor II
  • 2317 Views
  • 3 replies
  • 0 kudos

How to find the distinct count of the below listed result from the table?

Hi,How to get the distinct count from the below listed image,keywords = column nametable = appCatalogkeywords (column)"[""data"",""cis"",""mining"",""financial"",""pso"",""value""]""[""bzo"",""employee news"",""news""]""[""core.store"",""fbi""]""[""d...

  • 2317 Views
  • 3 replies
  • 0 kudos
Latest Reply
Kaniz_Fatma
Community Manager
  • 0 kudos

Hi @SivaPK , Try this :- # Assuming you have a DataFrame named 'df' with a column 'keywords'from pyspark.sql.functions import explode, col # Explode the array column 'keywords'exploded_df = df.select(explode(col("keywords")).alias("keyword")) # Count...

  • 0 kudos
2 More Replies
Amy
by New Contributor II
  • 3327 Views
  • 5 replies
  • 2 kudos

Gantt chart in dashboard

I'd like to create Gantt charts using the dashboard function. It seems like this could be possible by adding some additional parameters in the bar plot functionality, but I don't see how to do it currently (if there is a way, would love an example!)....

Screen Shot 2022-06-23 at 1.48.05 PM
  • 3327 Views
  • 5 replies
  • 2 kudos
Latest Reply
alexiswl
Contributor
  • 2 kudos

Hi @Vidula,I don't think this has been resolved.  I think gantt charts would look fantastic in a Lakeview Dashboard.  

  • 2 kudos
4 More Replies
bradleyjamrozik
by New Contributor III
  • 659 Views
  • 1 replies
  • 0 kudos

Server ODBC Connection

Is there a preferred method for hosting an odbc connection to a warehouse on a server for use by a report server (SSRS/PBIRS)? I know the odbc driver doesn't support pass-through authentication, so is there a way to configure it with an unattended ac...

  • 659 Views
  • 1 replies
  • 0 kudos
Latest Reply
Kaniz_Fatma
Community Manager
  • 0 kudos

Hi @bradleyjamrozik , Certainly! Configuring an ODBC connection for a report server (such as SSRS or PBIRS) involves several considerations.    Let’s explore some options:   Direct ODBC Connection: You can create an ODBC data source on the server whe...

  • 0 kudos
Avin_Kohale
by New Contributor
  • 17210 Views
  • 5 replies
  • 2 kudos

Import python files as modules in workspace

I'm deploying a new workspace for testing the deployed notebooks. But when trying to import the python files as module in the newly deployed workspace, I'm getting an error saying "function not found".Two points to note here:1. If I append absolute p...

  • 17210 Views
  • 5 replies
  • 2 kudos
Latest Reply
TimReddick
New Contributor III
  • 2 kudos

Hi @Kaniz_Fatma, I see your suggestion to append the necessary path to the sys.path. I'm curious if this is the recommendation for projects deployed via Databricks Asset Bundles. I want to maintain a project structure that looks something like this:p...

  • 2 kudos
4 More Replies
tranbau
by New Contributor
  • 477 Views
  • 0 replies
  • 0 kudos

Dynamic Spark Structured Streaming: Handling Stream-Stream Joins with Changing

I want to create a simple application using Spark Structured Streaming to alert users (via email, SMS, etc.) when stock price data meets certain requirements.I have a data stream: data_streamHowever, I'm strugging to address the main issue: how users...

Warehousing & Analytics
kafka
spark
spark-structured-streaming
stream-stream join
  • 477 Views
  • 0 replies
  • 0 kudos
Zer
by New Contributor II
  • 2436 Views
  • 5 replies
  • 4 kudos

SQL Editor, Tab Stops Indenting

Greetings,I use Databricks through Azure. Frequently when I'm working in the SQL Editor, the tab button fails to indent-- instead, it forces my cursor to a seemingly random part of the page. It's been doing this since when I started working in the pl...

  • 2436 Views
  • 5 replies
  • 4 kudos
Latest Reply
HannesM
New Contributor II
  • 4 kudos

Same issue here, sometimes it works by selecting a single but complete line and then hitting tab. If it works, then indentation works on multiple lines again as well. However, the single line select doesn't always work either. Pretty inconsistent beh...

  • 4 kudos
4 More Replies
Mat
by New Contributor III
  • 6790 Views
  • 5 replies
  • 3 kudos

Connect to Databricks SQL Endpoint using Programming language

Hi, I would like to know whether there is a feasibility/options available to connect to databricks sql endpoint using a programming language like java/scala/c#. I can see JDBC URL, but would like to whether it can be considered as any other jdbc conn...

  • 6790 Views
  • 5 replies
  • 3 kudos
Latest Reply
StephanieRivera
Valued Contributor II
  • 3 kudos

I found a similar question on Stackoverflow https://stackoverflow.com/questions/77477103/ow-to-properly-connect-to-azure-databricks-warehouse-from-c-sharp-net-using-jdb

  • 3 kudos
4 More Replies
Join 100K+ Data Experts: Register Now & Grow with Us!

Excited to expand your horizons with us? Click here to Register and begin your journey to success!

Already a member? Login and join your local regional user group! If there isn’t one near you, fill out this form and we’ll create one for you to join!

Labels
Top Kudoed Authors