- 4819 Views
- 3 replies
- 0 kudos
Cannot %run notebook using relative path
In the root of my workspace, I have the following folder structure set up:/foo / ETL / setup / utilitiesFrom a notebook in /foo/ETL or /foo/setup, if I try to %run a notebook in /foo/utilities by giving the relative path ../utilities/<x>, I get...
- 4819 Views
- 3 replies
- 0 kudos
- 0 kudos
Hello @DManowitz-BAH ,Hope you are doing great, I tried to replicate this behavior on 12.2 LTS runtime and didn't get any errors for the same folder/workspace structure..Can you share the cluster configuration details so I can mimic exactly the same ...
- 0 kudos
- 3070 Views
- 2 replies
- 1 kudos
Apache spark proficient
What is the best way to be proficient in Apache spark?
- 3070 Views
- 2 replies
- 1 kudos
- 1974 Views
- 1 replies
- 1 kudos
Delta Live Tables: BAD_REQUEST: Pipeline cluster is not reachable.
Hello community: I don´t know why my Delta Live table workflow fails into this step. This is the configuration I have for the pipeline:{"id": "**","pipeline_type": "WORKSPACE","clusters": [{"label": "default","spark_conf": {***},"num_workers": 3},{"l...
- 1974 Views
- 1 replies
- 1 kudos
- 3713 Views
- 4 replies
- 3 kudos
Databricks repos adds white spaces to notebook.
Hi,Databricks repos is making white spaces in the notebooks, without us opening the notebook. So seems to be automatically. Does someone has a solution how we can turn this off?An example:Can someone helps us with this?
- 3713 Views
- 4 replies
- 3 kudos
- 3 kudos
This bug is causing problems in our CI/CD pipeline too. I've had to manually clear things to keep it running as we get merge conflicts when no changes are made.
- 3 kudos
- 5895 Views
- 0 replies
- 0 kudos
- 5895 Views
- 0 replies
- 0 kudos
- 5836 Views
- 0 replies
- 0 kudos
Connecting to hive metastore as well as glue catalog at the sametime
Hi,Is there any way we can connect glue catalog as well as to hive metastore in the same warehouse?I can create a single instance profile and provide all the required access for buckets or for glue catalog.I tried with the below configuration,spark.s...
- 5836 Views
- 0 replies
- 0 kudos
- 1540 Views
- 0 replies
- 0 kudos
Optuna results change when rerun on Databricks
The best trial results seem to change every time the same study is rerun. On Microsoft Azure this can be fixed by setting the sampler seed. However this solution doesn't seem to work on Databricks. Does anyone know why that is the case and how to mak...
- 1540 Views
- 0 replies
- 0 kudos
- 1115 Views
- 0 replies
- 0 kudos
Using SQL to create delta live tables but column headers from csv have spaces
I know this has probably been asked many times, although I have struggled to find the answer using sql CREATE STREAMING LIVE TABLE bronze_site_usage COMMENT "ingest latest snapshot from api"TBLPROPERTIES ("layer" = "bronze") AS SELECT * FROM cloud_fi...
- 1115 Views
- 0 replies
- 0 kudos
- 872 Views
- 0 replies
- 0 kudos
Structured Streaming from TimescaleDB
I realize that the best practice would be to integrate our service with Kafka as a streaming source for Databricks, but given that the service already stores data into TimescaleDB, how can I stream data from TimescaleDB into DBX? Debezium doesn't wor...
- 872 Views
- 0 replies
- 0 kudos
- 3070 Views
- 1 replies
- 0 kudos
No results were found: the query results may no longer be available or you may not have permissions
When anyone (admins included) click on an alert task in a job run we see the error `No results were found: the query results may no longer be available or you may not have permissions`.Should we be seeing something else or is this a matter of a poor ...
- 3070 Views
- 1 replies
- 0 kudos
- 0 kudos
The error message "No results were found: the query results may no longer be available or you may not have permissions" is designed to address a range of potential situations. This includes instances where data might not be accessible due to reasons ...
- 0 kudos
- 10308 Views
- 1 replies
- 1 kudos
How to properly import spark functions?
I have the following command that runs in my databricks notebook.spark.conf.get("spark.databricks.clusterUsageTags.managedResourceGroup")I have wrapped this command into a function (simplified).def get_info(): return spark.conf.get("spark.databri...
- 10308 Views
- 1 replies
- 1 kudos
- 1234 Views
- 1 replies
- 0 kudos
Cannot create an account to try Community Edition
Hi,Whenever I try to signup for an account, I keep getting the following message in the first step - "an error has occurred. please try again later" Could you please let me know why this could be? I tried multiple emails and seems to be having same i...
- 1234 Views
- 1 replies
- 0 kudos
- 1567 Views
- 0 replies
- 0 kudos
Struct type limitation: possible hidden limit for parquet tables
Recently I discovered an issue when creating a PARQUET table that contains a column of type STRUCT with more than 350 string subfields. Such a table can be successfully created via a standard DDL script nevertheless each subsequent attempt to work wi...
- 1567 Views
- 0 replies
- 0 kudos
- 12432 Views
- 5 replies
- 0 kudos
Response for list job runs API request doesn't have next/prev_page_token field
HI,When I do GET request to obtain list of job runs using `/api/2.1/jobs/runs/list` there is no `prev_page_token`, `next_page_token` fields in the response despite of having `has_more: True`.
- 12432 Views
- 5 replies
- 0 kudos
- 0 kudos
@Hubert-Dudek @Retired_mod could you confirm that behavior is an issue? if yes, can I report it anywhere?Otherwise can I report it as a feature request (to get all runs)? If yes, do you have any service to report it?
- 0 kudos
- 1671 Views
- 0 replies
- 2 kudos
Revolutionizing Data Interaction
The new way of working with Databricks allows users to easily connect their notebook to a SQL serverless warehouse, write code in plain English, and generate output with the press of Space-Ctrl-Shift.
- 1671 Views
- 0 replies
- 2 kudos
Join Us as a Local Community Builder!
Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!
Sign Up Now-
.CSV
1 -
Access Data
2 -
Access Databricks
1 -
Access Delta Tables
2 -
Account reset
1 -
ADF Pipeline
1 -
ADLS Gen2 With ABFSS
1 -
Advanced Data Engineering
1 -
AI
1 -
Analytics
1 -
Apache spark
1 -
Apache Spark 3.0
1 -
API Documentation
3 -
Architecture
1 -
asset bundle
1 -
Asset Bundles
2 -
Auto-loader
1 -
Autoloader
4 -
AWS
3 -
AWS security token
1 -
AWSDatabricksCluster
1 -
Azure
5 -
Azure data disk
1 -
Azure databricks
14 -
Azure Databricks SQL
6 -
Azure databricks workspace
1 -
Azure Delta Lake
1 -
Azure Unity Catalog
5 -
Azure-databricks
1 -
AzureDatabricks
1 -
AzureDevopsRepo
1 -
Big Data Solutions
1 -
Billing
1 -
Billing and Cost Management
1 -
Blackduck
1 -
Bronze Layer
1 -
Certification
3 -
Certification Exam
1 -
Certification Voucher
3 -
CICDForDatabricksWorkflows
1 -
Cloud_files_state
1 -
CloudFiles
1 -
Cluster
3 -
Community Edition
3 -
Community Event
1 -
Community Group
1 -
Community Members
1 -
Compute
3 -
Compute Instances
1 -
conditional tasks
1 -
Connection
1 -
Contest
1 -
Credentials
1 -
CustomLibrary
1 -
Data
1 -
Data + AI Summit
1 -
Data Engineering
3 -
Data Explorer
1 -
Data Ingestion & connectivity
1 -
Databrick add-on for Splunk
1 -
databricks
2 -
Databricks Academy
1 -
Databricks AI + Data Summit
1 -
Databricks Alerts
1 -
Databricks Assistant
1 -
Databricks Certification
1 -
Databricks Cluster
2 -
Databricks Clusters
1 -
Databricks Community
10 -
Databricks community edition
3 -
Databricks Community Edition Account
1 -
Databricks Community Rewards Store
3 -
Databricks connect
1 -
Databricks Dashboard
2 -
Databricks delta
2 -
Databricks Delta Table
2 -
Databricks Demo Center
1 -
Databricks Documentation
2 -
Databricks genAI associate
1 -
Databricks JDBC Driver
1 -
Databricks Job
1 -
Databricks Lakehouse Platform
6 -
Databricks Migration
1 -
Databricks notebook
2 -
Databricks Notebooks
3 -
Databricks Platform
2 -
Databricks Pyspark
1 -
Databricks Python Notebook
1 -
Databricks Repo
1 -
Databricks Runtime
1 -
Databricks SQL
5 -
Databricks SQL Alerts
1 -
Databricks SQL Warehouse
1 -
Databricks Terraform
1 -
Databricks UI
1 -
Databricks Unity Catalog
4 -
Databricks Workflow
2 -
Databricks Workflows
2 -
Databricks workspace
2 -
Databricks-connect
1 -
DatabricksJobCluster
1 -
DataDays
1 -
Datagrip
1 -
DataMasking
2 -
DataVersioning
1 -
dbdemos
2 -
DBFS
1 -
DBRuntime
1 -
DBSQL
1 -
DDL
1 -
Dear Community
1 -
deduplication
1 -
Delt Lake
1 -
Delta
22 -
Delta Live Pipeline
3 -
Delta Live Table
5 -
Delta Live Table Pipeline
5 -
Delta Live Table Pipelines
4 -
Delta Live Tables
7 -
Delta Sharing
2 -
deltaSharing
1 -
Deny assignment
1 -
Development
1 -
Devops
1 -
DLT
10 -
DLT Pipeline
7 -
DLT Pipelines
5 -
Dolly
1 -
Download files
1 -
Dynamic Variables
1 -
Engineering With Databricks
1 -
env
1 -
ETL Pipelines
1 -
External Sources
1 -
External Storage
2 -
FAQ for Databricks Learning Festival
2 -
Feature Store
2 -
Filenotfoundexception
1 -
Free trial
1 -
GCP Databricks
1 -
GenAI
1 -
Getting started
2 -
Google Bigquery
1 -
HIPAA
1 -
import
1 -
Integration
1 -
JDBC Connections
1 -
JDBC Connector
1 -
Job Task
1 -
Lineage
1 -
LLM
1 -
Login
1 -
Login Account
1 -
Machine Learning
2 -
MachineLearning
1 -
Materialized Tables
2 -
Medallion Architecture
1 -
Migration
1 -
ML Model
1 -
MlFlow
2 -
Model Training
1 -
Module
1 -
Networking
1 -
Notebook
1 -
Onboarding Trainings
1 -
Pandas udf
1 -
Permissions
1 -
personalcompute
1 -
Pipeline
2 -
Plotly
1 -
PostgresSQL
1 -
Pricing
1 -
Pyspark
1 -
Python
5 -
Python Code
1 -
Python Wheel
1 -
Quickstart
1 -
Read data
1 -
Repos Support
1 -
Reset
1 -
Rewards Store
2 -
Schedule
1 -
Serverless
3 -
Session
1 -
Sign Up Issues
2 -
Spark
3 -
Spark Connect
1 -
sparkui
2 -
Splunk
2 -
SQL
8 -
Summit23
7 -
Support Tickets
1 -
Sydney
2 -
Table Download
1 -
Tags
1 -
Training
2 -
Troubleshooting
1 -
Unity Catalog
4 -
Unity Catalog Metastore
2 -
Update
1 -
user groups
1 -
Venicold
3 -
Voucher Not Recieved
1 -
Watermark
1 -
Weekly Documentation Update
1 -
Weekly Release Notes
2 -
Women
1 -
Workflow
2 -
Workspace
3
- « Previous
- Next »
User | Count |
---|---|
133 | |
90 | |
42 | |
42 | |
30 |