cancel
Showing results for 
Search instead for 
Did you mean: 
Machine Learning
Dive into the world of machine learning on the Databricks platform. Explore discussions on algorithms, model training, deployment, and more. Connect with ML enthusiasts and experts.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Thanapat_S
by Contributor
  • 3371 Views
  • 2 replies
  • 5 kudos

Resolved! Is it possible to use both `Dynamic partition overwrites` and `overwriteSchema` options when writing a DataFrame to a Delta table?"

In my ETL case, I want to be able to adjust the table schema as needed, meaning the number of columns may increase or decrease depending on the ETL script. Additionally, I would like to use dynamic partition overwrite to avoid potential errors when u...

image
  • 3371 Views
  • 2 replies
  • 5 kudos
Latest Reply
Vartika
Databricks Employee
  • 5 kudos

Hi @Thanapat Sontayasara​,Does @Werner Stinckens​'s response answer your question? If yes, would you be happy to mark it as best so that other members can find the solution more quickly? If not, would you be happy to give us more information?Thanks! 

  • 5 kudos
1 More Replies
AleksandraFrolo
by New Contributor III
  • 7027 Views
  • 5 replies
  • 6 kudos

Resolved! Merge 12 CSV files in Databricks.

Hello everybody,I am absolutely new in Databricks, so I need your help.Details:Task: merge 12 CSV files in Databricks with the best way.Location of files: I will describe it in details, because I can not good orientate yet. If i go to Data -> Browse ...

  • 7027 Views
  • 5 replies
  • 6 kudos
Latest Reply
Lakshay
Databricks Employee
  • 6 kudos

It seems that all your csv files are present under one folder and since you are able to union them, all these files must have same schema as well.Given the above conditions, you can simply read all the data by referring the folder name instead of ref...

  • 6 kudos
4 More Replies
ptawil
by New Contributor III
  • 3063 Views
  • 2 replies
  • 4 kudos

Runtime error using MLFlow and Spark on databricks

Here is some model I created:class SomeModel(mlflow.pyfunc.PythonModel): def predict(self, context, input): # do fancy ML stuff # log results pandas_df = pd.DataFrame(...insert predictions here...) spark_df = spark...

  • 3063 Views
  • 2 replies
  • 4 kudos
Latest Reply
Nikhil3107
New Contributor III
  • 4 kudos

Any updates on this? I am running into the same issue@Patrick Tawil​ were you able to solve this problem? If so, do you mind sharing?

  • 4 kudos
1 More Replies
Saurabh707344
by New Contributor III
  • 1231 Views
  • 1 replies
  • 1 kudos

Resolved! Comparative study of Azure Databricks MLOps capabilities in conjuction with Azuredevops, GIT, Jenkins

Looking for Comparative study of capabilities of below tools combination. In what situation I should use which of the below combination for MLOps project?a) Azure Databricks MLb) Azure Databricks ML + Azure Devops + GITc) Azure Databricks ML + Jenkin...

  • 1231 Views
  • 1 replies
  • 1 kudos
Latest Reply
shyam_9
Databricks Employee
  • 1 kudos

Hi @saurabh707344, you can use Azure Databricks ML when you're in the initial stages and developing some POCs. The other tools you mentioned were used based on your usecase when you moved some of the models to production and actively developing and ...

  • 1 kudos
Anjali
by New Contributor II
  • 1159 Views
  • 1 replies
  • 1 kudos

Databricks exam got suspended

today my databricks exam got suspended in the middle, the support team said they cannot reschdule my exam and asked me to contact my test sponsor i mailed tem many times but i haven't receied any reply can u please help with this.

  • 1159 Views
  • 1 replies
  • 1 kudos
Latest Reply
jose_gonzalez
Databricks Employee
  • 1 kudos

Adding @Suteja Kanuri​ , @Kaniz Fatma​ and @Vidula Khanna​ for visibility, to help you with your request

  • 1 kudos
536750
by New Contributor II
  • 2002 Views
  • 2 replies
  • 3 kudos

Resolved! ImportError: cannot import name 'FMIN_CANCELLED_REASON_EARLY_STOPPING' from 'hyperopt.spark'

I am trying to run an automl experiment from notebook. The error occurs when I import automl from databricks. Finally the error points to hyperopt. I have tried using the older version of hyperopt as well but nothings working.

  • 2002 Views
  • 2 replies
  • 3 kudos
Latest Reply
Anonymous
Not applicable
  • 3 kudos

Hi @Salar Ahmad​ Thank you for posting your question in our community! We are happy to assist you.To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best answers y...

  • 3 kudos
1 More Replies
animadurkar
by New Contributor III
  • 5661 Views
  • 2 replies
  • 2 kudos

Resolved! Invalid catalog and schema for table name error when creating a Feature Store

I've created a dataset that I want to create a feature store table for. I created a database in my dev unity catalog to store the feature tables, but each time I try to create the table I get this error:ValueError: Invalid catalog 'dev' or schema 'fe...

  • 5661 Views
  • 2 replies
  • 2 kudos
Latest Reply
Oogh5zai1
New Contributor II
  • 2 kudos

Hello,The error message suggests that there might be an issue with the catalog or schema name you are using when trying to create the feature store table. Here are a few tips to help you troubleshoot this problem:Verify the catalog and schema names: ...

  • 2 kudos
1 More Replies
HT
by New Contributor II
  • 10506 Views
  • 5 replies
  • 2 kudos

Trying out Dolly - how to load pytorch_model.bin so it's not downloading it every time the cluster is restarted

Hi, I am new to LLM and am curious to try it out. I did the following code to test from the databricks website:import torch from transformers import pipeline instruct_pipeline = pipeline(model="databricks/dolly-v2-12b", torch_dtype=torch.bfloat16, tr...

  • 10506 Views
  • 5 replies
  • 2 kudos
Latest Reply
sean_owen
Databricks Employee
  • 2 kudos

Just set the HF cache dir to a persistent path on /dbfs:import os os.environ['TRANSFORMERS_CACHE'] = "/dbfs/..."

  • 2 kudos
4 More Replies
Saurabh707344
by New Contributor III
  • 2723 Views
  • 4 replies
  • 3 kudos

Azure Databricks Capabilities: My objective is to evaluate Azure Databricks capability and Do I need to use Azure Devops or Jenkins or Databricks suffice the need.

hi,We have Real time streaming usecase where we have to build pipeline using Azure Databricks.My objective is to evaluate Azure Databricks capability and Do I need to use Azure Devops or Jenkins or Databricks suffice the need.Can you please provide c...

  • 2723 Views
  • 4 replies
  • 3 kudos
Latest Reply
StephanieAlba
Databricks Employee
  • 3 kudos

I found these youtube videos to be beneficial. CI/CD with Azure Dev Ops Terraform Enablement - Part 1 of 2

  • 3 kudos
3 More Replies
DataBRObin
by New Contributor III
  • 4544 Views
  • 6 replies
  • 1 kudos

FFmpeg frame extraction explodes memory, how to mitigate?

For a computer vision project, my raw data consists of encrypted videos (60fps) stored in Azure Blob Storage. In order to have the data usable for model training, I need to do some preprocessing and for that I need the video split into individual fra...

  • 4544 Views
  • 6 replies
  • 1 kudos
Latest Reply
DataBRObin
New Contributor III
  • 1 kudos

In the end, I decided to change around the workflow so it is as efficient as I could imagine it:Extract frames of video files in a containerized application somewhere running ffmpeg and storing the resulting frames in a parquet file in blob storage (...

  • 1 kudos
5 More Replies
Ismail1
by New Contributor III
  • 1705 Views
  • 2 replies
  • 0 kudos

Can an HMS-managed table be upgraded to Unity Catalog?

As the question states, I am not getting the option to upgrade managed tables on UC. Is that possible, I can't find anything on the documentation?

  • 1705 Views
  • 2 replies
  • 0 kudos
Latest Reply
Ismail1
New Contributor III
  • 0 kudos

In case anyone else ever faced the same issue

  • 0 kudos
1 More Replies
David_K93
by Contributor
  • 5044 Views
  • 3 replies
  • 2 kudos

Resolved! Issues loading .txt files from DBFS into Langchain TextLoader()

Hello,I am working on building a Langchain QA application in Databricks. I currently have 13 .txt files loaded into the DBFS and am trying to read them in iteratively with TextLoader(), load them into the RecursiveCharacterTextSplitter() from Langcha...

guru_error Screenshot 2023-05-19 171751
  • 5044 Views
  • 3 replies
  • 2 kudos
Latest Reply
Anonymous
Not applicable
  • 2 kudos

Hi @David Kersey​ Thank you for posting your question in our community! We are happy to assist you.To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best answers ...

  • 2 kudos
2 More Replies
Aarsh
by New Contributor II
  • 5524 Views
  • 0 replies
  • 1 kudos

Failure in Databricks Serving endpoint <build log says Pip failed due to conflicting dependency.>

Hello All,We are trying to deploy some models using Databricks Serving endpoint, But while deploying the artifact created during experiment run the serving endpoint build log says Pip failed due to conflicting dependency.The model is logged in experi...

  • 5524 Views
  • 0 replies
  • 1 kudos
JAHNAVI
by Databricks Employee
  • 7012 Views
  • 1 replies
  • 2 kudos

Resolved! error in running a LLM model in pyfunc.spark_udf

error in running a LLM model in pyfunc.spark_udf

  • 7012 Views
  • 1 replies
  • 2 kudos
Latest Reply
JAHNAVI
Databricks Employee
  • 2 kudos

Solution:   Please find below the example. Creating a secret and scope is a one time activity once we create a scope and secret we can access the token using any notebook or cluster in the workspace as shown below. After creating a secret if we want ...

  • 2 kudos

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.

If there isn’t a group near you, start one and help create a community that brings people together.

Request a New Group
Labels