- 461 Views
- 1 replies
- 1 kudos
Your session has expired, please authenticate again
I get this message every fifteen seconds in the community edition. The platform is practically not usable by any means.
- 461 Views
- 1 replies
- 1 kudos
- 1 kudos
Same issue. Its so annoying. Hope they rectify it asap.
- 1 kudos
- 2069 Views
- 7 replies
- 0 kudos
Running dbt in a Databricks task, the dbt_output from the Databricks jobs api is empty
I'm running a scheduled workflow with a dbt task on Azure Databricks. We want to export the dbt-output from the dbt task to a storage container for our Slim CI setup and data observability. The issue is, that the Databricks API (/api/2.1/jobs/runs/ge...
- 2069 Views
- 7 replies
- 0 kudos
- 0 kudos
I am running dbt on a databricks job. It saves all documentation: manifest.json, run_results.json, etc in "Download Artifacts" in a job. I am not able to find out a way to read those in codes, transform and save on databricks. Tried job API. The arti...
- 0 kudos
- 154 Views
- 0 replies
- 0 kudos
Experiencing Performance Issues When Using Databricks as a Transactional System with .NET App
Hello everyone,We are currently facing performance issues when using Databricks as a transactional system with our .NET application via the Simba ODBC driver. Specifically, queries are taking 30 seconds to 1 minute to insert data into 8 to 10 tables,...
- 154 Views
- 0 replies
- 0 kudos
- 1858 Views
- 3 replies
- 0 kudos
Creating a python package that uses dbutils.secrets
Hello Databricks,I wanted to create python package which has a python script which has a class , this class is where we give scope and key and we het the secret. This package will be used inside a databricks notebook I want to use dbutils.secret for ...
- 1858 Views
- 3 replies
- 0 kudos
- 0 kudos
Same question. Did you ever get an answer for this?I want to use databricks secrets inside my custom Python code. This is only necessary since based on this post (https://community.databricks.com/t5/administration-architecture/aws-secrets-manager-acc...
- 0 kudos
- 269 Views
- 1 replies
- 0 kudos
Databricks certificate exam got suspended, Need help to reschedule the exam
Hi @Cert-Team ,Today Mu Databricks certificate exam was suspended in the midway after approx 1 hour. Proctor asked me to show the room and after sometime he conveyed that my exam is being suspended and i will have to reach out to Databricks team for ...
- 269 Views
- 1 replies
- 0 kudos
- 0 kudos
@Ajaykumar the support team has responded to your ticket.
- 0 kudos
- 470 Views
- 6 replies
- 0 kudos
Getting an error while editing databrics metastore "databricks_metastore" using terraform (Azure)
Hi , I am getting an error (Metastore has reached the limit for metastores in region ) while updating metatsore using terraform , below the script i am using for updating metastore . There is already a metastore available i dont want to create the ne...
- 470 Views
- 6 replies
- 0 kudos
- 0 kudos
Hi, Thank you for the suggestions, But i am still getting same error .
- 0 kudos
- 374 Views
- 4 replies
- 0 kudos
Resolved! Help to reschedule of my exam
Hi team,My Databricks Certified Data Engineer Associate exam got suspended within 40 minutes. I had also shown my exam room to the proctor. My exam got suspended due to watch. I was not using anything. Some questions are so big in the exam so I had t...
- 374 Views
- 4 replies
- 0 kudos
- 778 Views
- 2 replies
- 0 kudos
How to Resolve ConnectTimeoutError When Registering Models with MLflow
Hello everyone,I'm trying to register a model with MLflow in Databricks, but encountering an error with the following command: model_version = mlflow.register_model(f"runs:/{run_id}/random_forest_model", model_name) The error message is as follows:...
- 778 Views
- 2 replies
- 0 kudos
- 0 kudos
@otara_geni if you are still struggling, try this - set the environmental variable in your code just before logging the model with the URL of the regional S3 endpoint (from the error it looks like MLFlow is attempting to use global one, which may not...
- 0 kudos
- 526 Views
- 1 replies
- 0 kudos
Improve query performance of direct query with Databricks
I’m building a dashboard in Power BI’s Pro Workspace, connecting data via Direct Query from Databricks (around 60 million rows from 15 combined tables), using a SQL Serverless (small size and 4 clusters).The problem is that the dashboard is taking to...
- 526 Views
- 1 replies
- 0 kudos
- 0 kudos
Hi @viniciuscini, How are you doing?It sounds like you've already taken a lot of the right steps to optimize your Power BI dashboard, especially given the constraints you mentioned. Since you've ruled out many common optimizations and noticed that th...
- 0 kudos
- 198 Views
- 3 replies
- 0 kudos
What exactly is Vectorized query processing and columnar acceleration
Hey folks! I want to know and understand while using photon acceleration, there is a feature called columnar acceleration which basically is a method of storing data in columns rather than rows, which is particularly advantageous for analytical datab...
- 198 Views
- 3 replies
- 0 kudos
- 0 kudos
Hi @Slash, Thanks for reaching out! Please review the response and let us know if it answers your question. Your feedback is valuable to us and the community. If the response resolves your issue, kindly mark it as the accepted solution. This will hel...
- 0 kudos
- 1021 Views
- 3 replies
- 1 kudos
How do I create spark.sql.session.SparkSession?
When I create a session n Databricks it is defaulting to spark.sql.connect.session.SparkSession. How can I connect to spark with out spark connect?
- 1021 Views
- 3 replies
- 1 kudos
- 1 kudos
Is there any solution to this? Pandera, Evidently and Ydata Profiling break because they don't speak a sql.connect session object. They expect a spark.sql.session.SparkSession it's very frustrating not being to use any of these libraries with the new...
- 1 kudos
- 496 Views
- 3 replies
- 1 kudos
Resolved! Best Approach for Handling ETL Processes in Databricks
I am currently managing nearly 300 tables from a production database and considering moving the entire ETL process away from Azure Data Factory to Databricks.This process, which involves extraction, transformation, testing, and loading, is executed d...
- 496 Views
- 3 replies
- 1 kudos
- 1 kudos
Hi,Instead of 300 individual files or one massive script, try grouping similar tables together. For example, you could have 10 scripts, each handling 30 tables. This way, you get the best of both approches—This way you will have a freedom of easy deb...
- 1 kudos
- 784 Views
- 4 replies
- 2 kudos
Understanding Flight Cancellations and Rescheduling in Airlines Using Databricks and PySpark
In the airline industry, it’s important to manage flights efficiently. Knowing why flights get canceled or rescheduled helps improve customer satisfaction and operational performance. In this article, I’ll show you how to use Databricks and PySpark t...
- 784 Views
- 4 replies
- 2 kudos
- 2 kudos
@Brahmareddy Interesting one , thanks for sharing
- 2 kudos
- 2770 Views
- 10 replies
- 0 kudos
Unable to reactive an inactive user
Hi all,I am facing an issue with reactivating an inactive user i tried the following json with databricks cli run_update = { "schemas": [ "urn:ietf:params:scim:api:messages:2.0:PatchOp" ], "Operations": [ { "op": "replace", "path": "ac...
- 2770 Views
- 10 replies
- 0 kudos
- 0 kudos
@FunkybunchOO Thank you for your response! I will look into other connections, but we are not currently using SCIM. There must be something similar blocking the activation.
- 0 kudos
- 148 Views
- 0 replies
- 0 kudos
UCX
Hey folks! I want to know what are the features that UCX does not provides in UC or specially Hive to UC Migration that can be done manually but not using UCX. As UCX is currently in developing mode so there are so many drawbacks, can someone share t...
- 148 Views
- 0 replies
- 0 kudos
Connect with Databricks Users in Your Area
Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.
If there isn’t a group near you, start one and help create a community that brings people together.
Request a New Group-
AI Summit
4 -
Azure
3 -
Azure databricks
3 -
Bi
1 -
Certification
1 -
Certification Voucher
2 -
Chatgpt
1 -
Community
7 -
Community Edition
3 -
Community Members
2 -
Community Social
1 -
Contest
1 -
Data + AI Summit
1 -
Data Engineering
1 -
Data Processing
1 -
Databricks Certification
1 -
Databricks Cluster
1 -
Databricks Community
11 -
Databricks community edition
3 -
Databricks Community Rewards Store
3 -
Databricks Lakehouse Platform
5 -
Databricks notebook
1 -
Databricks Office Hours
1 -
Databricks Runtime
1 -
Databricks SQL
4 -
Databricks-connect
1 -
DBFS
1 -
Dear Community
3 -
Delta
10 -
Delta Live Tables
1 -
Documentation
1 -
Exam
1 -
Featured Member Interview
1 -
HIPAA
1 -
Integration
1 -
LLM
1 -
Machine Learning
1 -
Notebook
1 -
Onboarding Trainings
1 -
Python
2 -
Rest API
11 -
Rewards Store
2 -
Serverless
1 -
Social Group
1 -
Spark
1 -
SQL
8 -
Summit22
1 -
Summit23
5 -
Training
1 -
Unity Catalog
4 -
Version
1 -
VOUCHER
1 -
WAVICLE
1 -
Weekly Release Notes
2 -
weeklyreleasenotesrecap
2 -
Workspace
1
- « Previous
- Next »