cancel
Showing results for 
Search instead for 
Did you mean: 
Machine Learning
Dive into the world of machine learning on the Databricks platform. Explore discussions on algorithms, model training, deployment, and more. Connect with ML enthusiasts and experts.
cancel
Showing results for 
Search instead for 
Did you mean: 
Data + AI Summit 2024 - Data Science & Machine Learning

Forum Posts

DataBRObin
by New Contributor III
  • 2009 Views
  • 2 replies
  • 0 kudos

Running Keras model training with HorovodRunner works until the training function is exited ("The MPI_Query_thread() function was called after MPI_FINALIZE was invoked.")

I am running training of a Keras/Tensorflow deep learning model on a cluster of (for now) 2 workers and 1 driver (T4 GPU, 28GB, 4 core) using the Databricks provided HorovodRunner. It all seems to go well and the performance scales quite nicely over ...

  • 2009 Views
  • 2 replies
  • 0 kudos
Latest Reply
sean_owen
Databricks Employee
  • 0 kudos

I personally suspect it's your callbacks. Can you remove all those state callbacks and see if that is it?

  • 0 kudos
1 More Replies
DipakBachhav
by New Contributor III
  • 13792 Views
  • 1 replies
  • 4 kudos

Resolved! [Databricks][DatabricksJDBCDriver](500593) Communication link failure. Failed to connect to server. Reason: HTTP Response code: 403

I am trying to connect to databricks using java code. Can someone help me please? Here is the code so far I have got::    import java.sql.Connection;  import java.sql.DriverManager;  import java.sql.SQLException;  import java.util.Properties;     ...

  • 13792 Views
  • 1 replies
  • 4 kudos
Latest Reply
karthik_p
Esteemed Contributor
  • 4 kudos

@Dipak Bachhav​ do you have any restriction in terms if IP to access databricks, in case of that you need to enable particular ip from security groups

  • 4 kudos
Thanapat_S
by Contributor
  • 3104 Views
  • 2 replies
  • 5 kudos

Resolved! Is it possible to use both `Dynamic partition overwrites` and `overwriteSchema` options when writing a DataFrame to a Delta table?"

In my ETL case, I want to be able to adjust the table schema as needed, meaning the number of columns may increase or decrease depending on the ETL script. Additionally, I would like to use dynamic partition overwrite to avoid potential errors when u...

image
  • 3104 Views
  • 2 replies
  • 5 kudos
Latest Reply
Vartika
Databricks Employee
  • 5 kudos

Hi @Thanapat Sontayasara​,Does @Werner Stinckens​'s response answer your question? If yes, would you be happy to mark it as best so that other members can find the solution more quickly? If not, would you be happy to give us more information?Thanks! 

  • 5 kudos
1 More Replies
AleksandraFrolo
by New Contributor III
  • 6729 Views
  • 5 replies
  • 6 kudos

Resolved! Merge 12 CSV files in Databricks.

Hello everybody,I am absolutely new in Databricks, so I need your help.Details:Task: merge 12 CSV files in Databricks with the best way.Location of files: I will describe it in details, because I can not good orientate yet. If i go to Data -> Browse ...

  • 6729 Views
  • 5 replies
  • 6 kudos
Latest Reply
Lakshay
Databricks Employee
  • 6 kudos

It seems that all your csv files are present under one folder and since you are able to union them, all these files must have same schema as well.Given the above conditions, you can simply read all the data by referring the folder name instead of ref...

  • 6 kudos
4 More Replies
ptawil
by New Contributor III
  • 2975 Views
  • 2 replies
  • 4 kudos

Runtime error using MLFlow and Spark on databricks

Here is some model I created:class SomeModel(mlflow.pyfunc.PythonModel): def predict(self, context, input): # do fancy ML stuff # log results pandas_df = pd.DataFrame(...insert predictions here...) spark_df = spark...

  • 2975 Views
  • 2 replies
  • 4 kudos
Latest Reply
Nikhil3107
New Contributor III
  • 4 kudos

Any updates on this? I am running into the same issue@Patrick Tawil​ were you able to solve this problem? If so, do you mind sharing?

  • 4 kudos
1 More Replies
Saurabh707344
by New Contributor III
  • 1171 Views
  • 1 replies
  • 1 kudos

Resolved! Comparative study of Azure Databricks MLOps capabilities in conjuction with Azuredevops, GIT, Jenkins

Looking for Comparative study of capabilities of below tools combination. In what situation I should use which of the below combination for MLOps project?a) Azure Databricks MLb) Azure Databricks ML + Azure Devops + GITc) Azure Databricks ML + Jenkin...

  • 1171 Views
  • 1 replies
  • 1 kudos
Latest Reply
shyam_9
Databricks Employee
  • 1 kudos

Hi @saurabh707344, you can use Azure Databricks ML when you're in the initial stages and developing some POCs. The other tools you mentioned were used based on your usecase when you moved some of the models to production and actively developing and ...

  • 1 kudos
Anjali
by New Contributor II
  • 1037 Views
  • 1 replies
  • 1 kudos

Databricks exam got suspended

today my databricks exam got suspended in the middle, the support team said they cannot reschdule my exam and asked me to contact my test sponsor i mailed tem many times but i haven't receied any reply can u please help with this.

  • 1037 Views
  • 1 replies
  • 1 kudos
Latest Reply
jose_gonzalez
Databricks Employee
  • 1 kudos

Adding @Suteja Kanuri​ , @Kaniz Fatma​ and @Vidula Khanna​ for visibility, to help you with your request

  • 1 kudos
536750
by New Contributor II
  • 1895 Views
  • 2 replies
  • 3 kudos

Resolved! ImportError: cannot import name 'FMIN_CANCELLED_REASON_EARLY_STOPPING' from 'hyperopt.spark'

I am trying to run an automl experiment from notebook. The error occurs when I import automl from databricks. Finally the error points to hyperopt. I have tried using the older version of hyperopt as well but nothings working.

  • 1895 Views
  • 2 replies
  • 3 kudos
Latest Reply
Anonymous
Not applicable
  • 3 kudos

Hi @Salar Ahmad​ Thank you for posting your question in our community! We are happy to assist you.To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best answers y...

  • 3 kudos
1 More Replies
animadurkar
by New Contributor III
  • 5396 Views
  • 2 replies
  • 2 kudos

Resolved! Invalid catalog and schema for table name error when creating a Feature Store

I've created a dataset that I want to create a feature store table for. I created a database in my dev unity catalog to store the feature tables, but each time I try to create the table I get this error:ValueError: Invalid catalog 'dev' or schema 'fe...

  • 5396 Views
  • 2 replies
  • 2 kudos
Latest Reply
Oogh5zai1
New Contributor II
  • 2 kudos

Hello,The error message suggests that there might be an issue with the catalog or schema name you are using when trying to create the feature store table. Here are a few tips to help you troubleshoot this problem:Verify the catalog and schema names: ...

  • 2 kudos
1 More Replies
HT
by New Contributor II
  • 10317 Views
  • 5 replies
  • 2 kudos

Trying out Dolly - how to load pytorch_model.bin so it's not downloading it every time the cluster is restarted

Hi, I am new to LLM and am curious to try it out. I did the following code to test from the databricks website:import torch from transformers import pipeline instruct_pipeline = pipeline(model="databricks/dolly-v2-12b", torch_dtype=torch.bfloat16, tr...

  • 10317 Views
  • 5 replies
  • 2 kudos
Latest Reply
sean_owen
Databricks Employee
  • 2 kudos

Just set the HF cache dir to a persistent path on /dbfs:import os os.environ['TRANSFORMERS_CACHE'] = "/dbfs/..."

  • 2 kudos
4 More Replies
Saurabh707344
by New Contributor III
  • 2650 Views
  • 4 replies
  • 3 kudos

Azure Databricks Capabilities: My objective is to evaluate Azure Databricks capability and Do I need to use Azure Devops or Jenkins or Databricks suffice the need.

hi,We have Real time streaming usecase where we have to build pipeline using Azure Databricks.My objective is to evaluate Azure Databricks capability and Do I need to use Azure Devops or Jenkins or Databricks suffice the need.Can you please provide c...

  • 2650 Views
  • 4 replies
  • 3 kudos
Latest Reply
StephanieAlba
Databricks Employee
  • 3 kudos

I found these youtube videos to be beneficial. CI/CD with Azure Dev Ops Terraform Enablement - Part 1 of 2

  • 3 kudos
3 More Replies
DataBRObin
by New Contributor III
  • 4327 Views
  • 6 replies
  • 1 kudos

FFmpeg frame extraction explodes memory, how to mitigate?

For a computer vision project, my raw data consists of encrypted videos (60fps) stored in Azure Blob Storage. In order to have the data usable for model training, I need to do some preprocessing and for that I need the video split into individual fra...

  • 4327 Views
  • 6 replies
  • 1 kudos
Latest Reply
DataBRObin
New Contributor III
  • 1 kudos

In the end, I decided to change around the workflow so it is as efficient as I could imagine it:Extract frames of video files in a containerized application somewhere running ffmpeg and storing the resulting frames in a parquet file in blob storage (...

  • 1 kudos
5 More Replies
Ismail1
by New Contributor III
  • 1641 Views
  • 2 replies
  • 0 kudos

Can an HMS-managed table be upgraded to Unity Catalog?

As the question states, I am not getting the option to upgrade managed tables on UC. Is that possible, I can't find anything on the documentation?

  • 1641 Views
  • 2 replies
  • 0 kudos
Latest Reply
Ismail1
New Contributor III
  • 0 kudos

In case anyone else ever faced the same issue

  • 0 kudos
1 More Replies
David_K93
by Contributor
  • 4854 Views
  • 3 replies
  • 2 kudos

Resolved! Issues loading .txt files from DBFS into Langchain TextLoader()

Hello,I am working on building a Langchain QA application in Databricks. I currently have 13 .txt files loaded into the DBFS and am trying to read them in iteratively with TextLoader(), load them into the RecursiveCharacterTextSplitter() from Langcha...

guru_error Screenshot 2023-05-19 171751
  • 4854 Views
  • 3 replies
  • 2 kudos
Latest Reply
Anonymous
Not applicable
  • 2 kudos

Hi @David Kersey​ Thank you for posting your question in our community! We are happy to assist you.To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best answers ...

  • 2 kudos
2 More Replies

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.

If there isn’t a group near you, start one and help create a community that brings people together.

Request a New Group
Labels