- 11994 Views
- 9 replies
- 4 kudos
Resolved! How to get usage statistics from Databricks or SQL Databricks?
Hi, I am looking for a way to get usage statistics from Databricks (Data Science & Engineering and SQL persona). For example: I created a table. I want to know how many times a specific user queried that table.How many times a pipeline was triggered?...
- 11994 Views
- 9 replies
- 4 kudos
- 4 kudos
You can use System Tables, now available in Unity Catalog metastore, to create the views you described. https://docs.databricks.com/en/admin/system-tables/index.html
- 4 kudos
- 3549 Views
- 4 replies
- 4 kudos
Resolved! DLT pipeline run cost
Hi team,I am looking for a way to find DBU cost for DLT clusters, does it get stored anywhere I have been looking into event_logs but did not find information related to cost. it does have cluster resource utilization details.here is what I found, co...
- 3549 Views
- 4 replies
- 4 kudos
- 4 kudos
Hi @Chhaya Vishwakarma​ Thank you for posting your question in our community! We are happy to assist you.To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best an...
- 4 kudos
- 2215 Views
- 1 replies
- 1 kudos
Resolved! In Python, Streaming read by DLT from Hive Table
I am pulling data from Google BigQuery and writing it to a bronze table on an interval. I do this in a separate continuous job because DLT did not like the BigQuery connector calling collect on a dataframe inside of DLT.In Python, I would like to rea...
- 2215 Views
- 1 replies
- 1 kudos
- 1 kudos
The below code is a solution. I was missing that I could read from a table with `spark.readStream.format("delta").table("...")`. Simple. Just missed it. This is different than `dlt.read_stream()` which appears in the examples a lot.This is referenced...
- 1 kudos
-
Aad
1 -
Access
1 -
Access control
1 -
ADLS Gen
1 -
API
3 -
AWS
1 -
Azure
3 -
Azure databricks
4 -
Azure Databricks SQL
1 -
Azure SQL DB
1 -
Azure synapse
1 -
Batch Processing
1 -
Best Data Warehouse
1 -
Best practice
1 -
Bi
5 -
Bigquery
1 -
Billing and Cost Management
1 -
Broadcast variable
1 -
Bug
1 -
Business Intelligence
2 -
Cache
1 -
Caching
1 -
Catalyst
1 -
CD Pipeline
1 -
Certification
1 -
Certification Voucher
1 -
Class
1 -
Cloud Fetch
1 -
Cluster
3 -
Cluster config
1 -
Cluster Metrics
1 -
ClusterSize
1 -
Code
1 -
ConcurrentQueries
1 -
Connect
1 -
Credential passthrough
1 -
CSV
1 -
CustomKeyVault
1 -
DAIS2023
1 -
Dashboard
1 -
Dashboards
1 -
Data Engineering
1 -
Data Ingestion & connectivity
2 -
Data Science
2 -
databricks
1 -
Databricks Certification
1 -
Databricks Certification Voucher
1 -
Databricks Cluster
3 -
Databricks JDBC
1 -
Databricks notebook
1 -
Databricks Runtime
1 -
Databricks SQL
21 -
Databricks SQL Alerts
1 -
Databricks SQL Analytics
1 -
Databricks SQL Connector
1 -
Databricks SQL Endpoints
2 -
Databricks Table Usage
1 -
Databricks workspace
1 -
DatabrickSQL
1 -
Dataset
1 -
DBeaver
1 -
DBR
2 -
DBSQL
12 -
DBSQL Queries
1 -
Dbu
1 -
Delta
5 -
Delta Live Table Pipeline
1 -
Delta Live Tables
2 -
Delta Pipeline
1 -
Delta table
1 -
Delta Tables
1 -
Different Types
1 -
DLT
3 -
E2
1 -
Endpoint
7 -
Error
1 -
Error Message
2 -
ETL Process
1 -
External Data Sources
1 -
External Hive
1 -
External Table
1 -
File
1 -
Files
1 -
Global Temp Views
1 -
Glue
1 -
Gpu
1 -
Group
1 -
Hive
1 -
Hive Table
1 -
Import
1 -
Jdbc
6 -
Jdbc connection
2 -
Job Cluster
1 -
Key
1 -
Library
1 -
Limit
1 -
LTS
1 -
LTS ML
1 -
Metadata
1 -
Migration
1 -
Multi Cluster Load Balancer
1 -
Mysql
2 -
NodeJS
1 -
Notebook
2 -
Odbc
3 -
Oracle
1 -
OracleDBPackage
1 -
PARAMETER VALUE
1 -
Parquet
1 -
Party Libraries
1 -
Password
1 -
Performance
2 -
Permissions
1 -
Photon
2 -
Pip
1 -
Possible
1 -
PostgresSQL
1 -
Powerbi
7 -
Prod Workspace
1 -
Programming language
1 -
Pyspark
1 -
Python
6 -
Python Dataframe
1 -
Query
6 -
Query History
1 -
Query Parameters
1 -
Query Snippets
1 -
Row Level Security
1 -
Row Limit
1 -
Schedule
1 -
Schema
1 -
ServiceNow User
1 -
Session
1 -
Simba Odbc Connector
1 -
SKU
1 -
Spark
2 -
Spark sql
1 -
Sparkcontext
1 -
Special Characters
1 -
SQL
40 -
SQL Dashboard
3 -
SQL Databricks
1 -
SQL Endpoint
3 -
SQL Endpoints
5 -
SQL Option
1 -
SQL Queries
3 -
Sql query
3 -
SQL Query Execution Plan
1 -
Sql table
1 -
Sql Warehouse
5 -
Sql Workbench
2 -
SQL Workspace Option
2 -
Sqlanalytics
1 -
Sqlexecutionexception
1 -
Sqlserver
1 -
SRC
1 -
Ssl
1 -
ST
1 -
String Agg
1 -
Structfield
1 -
Structured streaming
1 -
Summit22
1 -
Table
1 -
Table Pipeline
1 -
Temporary View
1 -
Trying
1 -
UI SQL
1 -
Unity Catalogue
1 -
Usage
1 -
Usge Statistics
1 -
Value Pair
1 -
Version
1 -
Version Queries
1 -
Visualization
1 -
Vnet Injection
1 -
Works
1 -
Workspace
1 -
Workspace SKU
1 -
Writing
1 -
Xml
1 -
Yarn
2 -
Zip file
1
- « Previous
- Next »