cancel
Showing results for 
Search instead for 
Did you mean: 
Get Started Discussions
Start your journey with Databricks by joining discussions on getting started guides, tutorials, and introductory topics. Connect with beginners and experts alike to kickstart your Databricks experience.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

otara_geni
by New Contributor
  • 2149 Views
  • 1 replies
  • 0 kudos

How to Resolve ConnectTimeoutError When Registering Models with MLflow

Hello everyone,I'm trying to register a model with MLflow in Databricks, but encountering an error with the following command: model_version = mlflow.register_model(f"runs:/{run_id}/random_forest_model", model_name)   The error message is as follows:...

  • 2149 Views
  • 1 replies
  • 0 kudos
Latest Reply
some-rsa
Databricks Employee
  • 0 kudos

@otara_geni if you are still struggling, try this - set the environmental variable in your code just before logging the model with the URL of the regional S3 endpoint (from the error it looks like MLFlow is attempting to use global one, which may not...

  • 0 kudos
viniciuscini
by New Contributor
  • 4748 Views
  • 1 replies
  • 0 kudos

Improve query performance of direct query with Databricks

I’m building a dashboard in Power BI’s Pro Workspace, connecting data via Direct Query from Databricks (around 60 million rows from 15 combined tables), using a SQL Serverless (small size and 4 clusters).The problem is that the dashboard is taking to...

  • 4748 Views
  • 1 replies
  • 0 kudos
Latest Reply
Brahmareddy
Esteemed Contributor
  • 0 kudos

Hi @viniciuscini, How are you doing?It sounds like you've already taken a lot of the right steps to optimize your Power BI dashboard, especially given the constraints you mentioned. Since you've ruled out many common optimizations and noticed that th...

  • 0 kudos
TestuserAva
by New Contributor II
  • 9576 Views
  • 12 replies
  • 2 kudos

Getting HTML sign I page as api response from databricks api with statuscode 200

Response:<!doctype html><html><head>    <meta charset="utf-8" />    <meta http-equiv="Content-Language" content="en" />    <title>Databricks - Sign In</title>    <meta name="viewport" content="width=960" />    <link rel="icon" type="image/png" href="...

TestuserAva_0-1701165195616.png
  • 9576 Views
  • 12 replies
  • 2 kudos
Latest Reply
SJR
New Contributor III
  • 2 kudos

Hey @schunduri, not entirely sure because our SRE did the change, but the machine the pipeline runs on must be within the same vnet as your DBKS workspace. If you need more guidance, I could try and check what we did but our SRE left the company sinc...

  • 2 kudos
11 More Replies
dvl_priyansh
by New Contributor
  • 1635 Views
  • 3 replies
  • 0 kudos

What exactly is Vectorized query processing and columnar acceleration

Hey folks! I want to know and understand while using photon acceleration, there is a feature called columnar acceleration which basically is a method of storing data in columns rather than rows, which is particularly advantageous for analytical datab...

  • 1635 Views
  • 3 replies
  • 0 kudos
Latest Reply
Retired_mod
Esteemed Contributor III
  • 0 kudos

Hi @szymon_dybczak, Thanks for reaching out! Please review the response and let us know if it answers your question. Your feedback is valuable to us and the community. If the response resolves your issue, kindly mark it as the accepted solution. This...

  • 0 kudos
2 More Replies
valjas
by New Contributor III
  • 4068 Views
  • 3 replies
  • 1 kudos

How do I create spark.sql.session.SparkSession?

When I create a session n Databricks it is defaulting to spark.sql.connect.session.SparkSession. How can I connect to spark with out spark connect?

  • 4068 Views
  • 3 replies
  • 1 kudos
Latest Reply
miguel_ortiz
New Contributor II
  • 1 kudos

Is there any solution to this? Pandera, Evidently and Ydata Profiling break because they don't speak a sql.connect session object. They expect a spark.sql.session.SparkSession it's very frustrating not being to use any of these libraries with the new...

  • 1 kudos
2 More Replies
liv1
by New Contributor II
  • 3136 Views
  • 2 replies
  • 1 kudos

Structured Streaming from a delta table that is a dump of kafka and get the latest record per key

I'm trying to use Structured Streaming in scala to stream from a delta table that is a dump of a kafka topic where each record/message is an update of attributes for the key and no messages from kafka are dropped from the dump, but the value is flatt...

  • 3136 Views
  • 2 replies
  • 1 kudos
Latest Reply
Maatari
New Contributor III
  • 1 kudos

I am confused about this recommendation. I thought the use of the append output mode in combination with aggregate queries is restricted to queries for which the aggregation is expressed using event-time and it defines a watermark.Could you clarify ?

  • 1 kudos
1 More Replies
itsmejoeyong
by New Contributor II
  • 3240 Views
  • 3 replies
  • 1 kudos

Resolved! Best Approach for Handling ETL Processes in Databricks

I am currently managing nearly 300 tables from a production database and considering moving the entire ETL process away from Azure Data Factory to Databricks.This process, which involves extraction, transformation, testing, and loading, is executed d...

  • 3240 Views
  • 3 replies
  • 1 kudos
Latest Reply
Brahmareddy
Esteemed Contributor
  • 1 kudos

Hi,Instead of 300 individual files or one massive script, try grouping similar tables together. For example, you could have 10 scripts, each handling 30 tables. This way, you get the best of both approches—This way you will have a freedom of easy deb...

  • 1 kudos
2 More Replies
Brahmareddy
by Esteemed Contributor
  • 1969 Views
  • 4 replies
  • 3 kudos

Understanding Flight Cancellations and Rescheduling in Airlines Using Databricks and PySpark

In the airline industry, it’s important to manage flights efficiently. Knowing why flights get canceled or rescheduled helps improve customer satisfaction and operational performance. In this article, I’ll show you how to use Databricks and PySpark t...

Get Started Discussions
airlines
artificial intelligence
feature engineering
machine learning
  • 1969 Views
  • 4 replies
  • 3 kudos
Latest Reply
Rishabh-Pandey
Esteemed Contributor
  • 3 kudos

@Brahmareddy Interesting one , thanks for sharing

  • 3 kudos
3 More Replies
AyushPandey
by New Contributor II
  • 8259 Views
  • 8 replies
  • 0 kudos

Unable to reactive an inactive user

Hi all,I am facing an issue with reactivating an inactive user i tried the following json with databricks cli run_update = {  "schemas": [ "urn:ietf:params:scim:api:messages:2.0:PatchOp" ],  "Operations": [    {      "op": "replace",      "path": "ac...

  • 8259 Views
  • 8 replies
  • 0 kudos
Latest Reply
bencsik
New Contributor III
  • 0 kudos

@FunkybunchOO Thank you for your response! I will look into other connections, but we are not currently using SCIM. There must be something similar blocking the activation.

  • 0 kudos
7 More Replies
priyansh
by New Contributor III
  • 578 Views
  • 0 replies
  • 0 kudos

UCX

Hey folks! I want to know what are the features that UCX does not provides in UC or specially Hive to UC Migration that can be done manually but not using UCX. As UCX is currently in developing mode so there are so many drawbacks, can someone share t...

  • 578 Views
  • 0 replies
  • 0 kudos
TinaN
by New Contributor III
  • 1197 Views
  • 2 replies
  • 0 kudos

Resolved! Translating XMLNAMESPACE in SQL Databricks

We are loading a data source that contains XML. I am translating their queries to create views in Databricks. They use 'XMLNAMESPACES' to construct/parse XML.  Below is an example.  What is best practice for translating 'XMLNAMESPACES' in Databricks?...

  • 1197 Views
  • 2 replies
  • 0 kudos
Latest Reply
Retired_mod
Esteemed Contributor III
  • 0 kudos

Hi @TinaN, To handle XMLNAMESPACES in Databricks, use the from_xml function for parsing XML data, where you can define namespaces within your parsing logic. Start by reading the XML data using spark.read.format("xml"), then apply the from_xml functio...

  • 0 kudos
1 More Replies
zll_0091
by New Contributor III
  • 623 Views
  • 1 replies
  • 0 kudos

Can I load the files based on the data in my table as variable without iterating through each row?

Hi,I have created this table which contains the data that I need for my source path and target table. source_path: /data/customer/sid={sid}/abc=1/attr_provider={attr_prov}/source_data_provider_code={src_prov}/So basically, the value of each row are c...

zll_0091_2-1722958875477.png zll_0091_1-1722958858553.png zll_0091_3-1722958975973.png
  • 623 Views
  • 1 replies
  • 0 kudos
Latest Reply
Retired_mod
Esteemed Contributor III
  • 0 kudos

Hi @zll_0091, To efficiently load only the necessary files without manually iterating through each row of your table, you can use Spark's DataFrame operations. First, read your table into a DataFrame and determine the maximum key value. Then, filter ...

  • 0 kudos
ozbieG
by New Contributor II
  • 1240 Views
  • 2 replies
  • 0 kudos

Databricks Certification exam got Suspended - Need Support

Hello Team, @Cert-Team , @Cert-TeamOPS I faced a very bad experience while attempting my 1st DataBricks certification.I was asked to exit the exam multiple times by the support team saying technical issues. My test got rescheduled multiple times with...

  • 1240 Views
  • 2 replies
  • 0 kudos
Latest Reply
Retired_mod
Esteemed Contributor III
  • 0 kudos

Hi @ozbieG, I'm sorry to hear your exam was suspended. Thank you for filing a ticket with our support team. Please allow the support team 24-48 hours to resolve. In the meantime, you can review the following documentation: Room requirements Behaviour...

  • 0 kudos
1 More Replies
bytetogo
by New Contributor
  • 1080 Views
  • 1 replies
  • 0 kudos

What API Testing Tool Do You Use?

Hi Databricks!I am a relatively new developer that's looking for a solid API testing tool. I am interested in hearing about other developers, new or experienced, about their experiences with API testing tools, regardless if they are good or bad. I've...

  • 1080 Views
  • 1 replies
  • 0 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 0 kudos

Hi @bytetogo,In my daily work I use Postman. It has user-friendly interface, supports automated testing and has support for popular patterns and libraries. It is also compatible with Linux, MacOs, Windows.

  • 0 kudos

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels