cancel
Showing results for 
Search instead for 
Did you mean: 
Machine Learning
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Abdurrahman
by New Contributor II
  • 712 Views
  • 2 replies
  • 0 kudos

How to download a pytorch model created via notebook and saved in a folder ?

I have created a pytorch model using databricks notebooks and saved it in a folder in workspace. MLFlow is not used.When I try to download the files from the folder it exceeds the download limit. Is there a way to download the model locally into my s...

  • 712 Views
  • 2 replies
  • 0 kudos
Latest Reply
Kaniz
Community Manager
  • 0 kudos

Hi @Abdurrahman,  If you know the direct URL of the pretrained PyTorch model, you can use wget or a Python script to download it directly to your local system.For example, if you want to download the pretrained ResNet-18 model, you can use the follow...

  • 0 kudos
1 More Replies
BogdanV
by New Contributor III
  • 965 Views
  • 3 replies
  • 0 kudos

Resolved! Query ML Endpoint with R and Curl

I am trying to get a prediction by querying the ML Endpoint on Azure Databricks with R. I'm not sure what is the format of the expected data. Is there any other problem with this code? Thanks!!! 

R Code.png
  • 965 Views
  • 3 replies
  • 0 kudos
Latest Reply
BogdanV
New Contributor III
  • 0 kudos

Hi Kaniz, I was able to find the solution. You should post this in the examples when you click "Query Endpoint"You only have code for Browser, Curl, Python, SQL. You should add a tab for RHere is the solution:library(httr)url <- "https://adb-********...

  • 0 kudos
2 More Replies
VJ3
by New Contributor III
  • 487 Views
  • 2 replies
  • 0 kudos

Security Controls to implement on Machine Learning Persona

Hello,Hope everyone are doing well. You may be aware that we are using Table ACL enabled cluster to ensure the adequate security controls on Databricks. You may be also aware that we can not use Table enabled ACL cluster on Machine Learning Persona. ...

  • 487 Views
  • 2 replies
  • 0 kudos
Latest Reply
Kaniz
Community Manager
  • 0 kudos

Hi @VJ3, Databricks is a powerful platform that combines data engineering, machine learning, and business intelligence. When deploying Databricks in an enterprise environment, it’s crucial to establish robust security practices. Let’s focus on best ...

  • 0 kudos
1 More Replies
G-M
by Contributor
  • 504 Views
  • 1 replies
  • 0 kudos

MLflow Experiments in Unity Catalog

Will MLflow Experiments be incorporated into Unity Catalog similar to models and feature tables? I feel like this is the final piece missing in a comprehensive Unity Catalog backed MLOps workflow. Currently it seems they can only be stored in a dbfs ...

  • 504 Views
  • 1 replies
  • 0 kudos
Latest Reply
Kaniz
Community Manager
  • 0 kudos

Hi @G-M,  While Models in Unity Catalog cover model registration and management, MLflow Experiments focus on experiment tracking, versioning, and metrics.Currently, MLflow Experiments are stored in a DBFS-backed location (Databricks File System), whi...

  • 0 kudos
larsr
by New Contributor
  • 420 Views
  • 1 replies
  • 0 kudos

DBR CLI v0.216.0 failed to pass bundle variable for notebook task

After installing the new version of the CLI (v0.216.0) the bundle variable for the notebook task is not parsed correctly, see code below:tasks:        - task_key: notebook_task          job_cluster_key: job_cluster          notebook_task:            ...

Machine Learning
asset bundles
  • 420 Views
  • 1 replies
  • 0 kudos
Latest Reply
Kaniz
Community Manager
  • 0 kudos

Hi @larsr,  Ensure that the variable ${var.notebook_path} is correctly defined and accessible within the context of your bundle configuration. Sometimes, scoping issues can lead to variable references not being resolved properly.

  • 0 kudos
johnp
by New Contributor II
  • 1043 Views
  • 2 replies
  • 0 kudos

Resolved! pdb debugger on databricks

I am new to databricks. and trying to debug my python application with variable-explore by following the instruction from: https://www.databricks.com/blog/new-debugging-features-databricks-notebooks-variable-explorerI added the "import pdb" in the fi...

  • 1043 Views
  • 2 replies
  • 0 kudos
Latest Reply
johnp
New Contributor II
  • 0 kudos

I test with some simple applications, it works as you described.  However, the application I am debugging uses the pyspark structured streaming, which runs continuously. After inserting pdb.set_trace(), the application paused at the breakpoint, but t...

  • 0 kudos
1 More Replies
Octavian1
by Contributor
  • 545 Views
  • 2 replies
  • 0 kudos

port undefined error in SQLDatabase.from_databricks (langchain.sql_database)

The following assignment:from langchain.sql_database import SQLDatabasedbase = SQLDatabase.from_databricks(catalog=catalog, schema=db,host=host, api_token=token,)fails with ValueError: invalid literal for int() with base 10: ''because ofcls._assert_p...

  • 545 Views
  • 2 replies
  • 0 kudos
Latest Reply
Kaniz
Community Manager
  • 0 kudos

Hi @Octavian1, Ensure that the port parameter you’re passing to SQLDatabase.from_databricks is a valid integer. If it’s empty or contains non-numeric characters, that could be the root cause. In a Stack Overflow post, someone faced a similar issue wh...

  • 0 kudos
1 More Replies
kng88
by New Contributor II
  • 1962 Views
  • 6 replies
  • 7 kudos

How to save model produce by distributed training?

I am trying to save model after distributed training via the following codeimport sys   from spark_tensorflow_distributor import MirroredStrategyRunner   import mlflow.keras   mlflow.keras.autolog()   mlflow.log_param("learning_rate", 0.001)   import...

  • 1962 Views
  • 6 replies
  • 7 kudos
Latest Reply
Xiaowei
New Contributor III
  • 7 kudos

I think I finally worked this out.Here is the extra code to save out the model only once and from the 1st node:context = pyspark.BarrierTaskContext.get() if context.partitionId() == 0: mlflow.keras.log_model(model, "mymodel")

  • 7 kudos
5 More Replies
yorabhir
by New Contributor II
  • 592 Views
  • 1 replies
  • 1 kudos

Resolved! 'error_code': 'INVALID_PARAMETER_VALUE', 'message': 'Too many sources. It cannot be more than 100'

I am getting the following error while saving a delta table in the feature storeWARNING databricks.feature_store._catalog_client_helper: Failed to record data sources in the catalog. Exception: {'error_code': 'INVALID_PARAMETER_VALUE', 'message': 'To...

  • 592 Views
  • 1 replies
  • 1 kudos
Latest Reply
Kaniz
Community Manager
  • 1 kudos

Hi @yorabhir,  Verify how many sources you’re trying to record in the catalog. If it exceeds 100, you’ll need to reduce the number of sources.Ensure that the feature table creation process is correctly configured. In your code snippet, you’re creatin...

  • 1 kudos
MaKarenina
by New Contributor
  • 506 Views
  • 1 replies
  • 0 kudos

ML Flow until January 24

Hi! When i was creating a new endpoint a have this alert  CREATE A MODEL SERVING ENDPOINT TO SERVE YOUR MODEL BEHIND A REST API INTERFACE. YOU CAN STILL USE LEGACY ML FLOW MODEL SERVING UNTIL JANUARY 2024 I don't understand if my Legacy MLFlow Model ...

  • 506 Views
  • 1 replies
  • 0 kudos
Latest Reply
Kaniz
Community Manager
  • 0 kudos

Hi @MaKarenina, The alert you received states that you can continue using Legacy MLflow Model Serving until January 2024. However, there are a few important points to consider: Support: After January 2024, Legacy MLflow Model Serving will no lon...

  • 0 kudos
Alessandro
by New Contributor
  • 547 Views
  • 1 replies
  • 0 kudos

using openai Api in Databricks without iterating rows

 Hi to everyone,I have a delta table with a column 'comment' I would like to add a new column 'sentiment', and I would like to calculate it using openai API.I already know how to create a databricks endpoint to an external model and how to use it (us...

  • 547 Views
  • 1 replies
  • 0 kudos
Latest Reply
Kaniz
Community Manager
  • 0 kudos

Hi @Alessandro, Your question is clear, and I appreciate your curiosity about optimizing the process. Let’s explore a couple of approaches: UDF (User-Defined Function): You can create a UDF in Databricks that invokes the OpenAI API for sentiment...

  • 0 kudos
Mirko
by Contributor
  • 754 Views
  • 3 replies
  • 1 kudos

Resolved! AutoMl Dataset too large

Hello community,i have the following problem: I am using automl to solve a regression model, but in the preprocessing my dataset is sampled to ~30% of the original amount.I am using runtime 14.2 ML Driver: Standard_DS4_v2 28GB Memory 8 coresWorker: S...

  • 754 Views
  • 3 replies
  • 1 kudos
Latest Reply
Mirko
Contributor
  • 1 kudos

I am pretty sure that i know what the problem was. I had a timestamp column (with second precision) as a feature. If they get one hot encoded, the dataset can get pretty large.

  • 1 kudos
2 More Replies
Miki
by New Contributor II
  • 442 Views
  • 3 replies
  • 0 kudos

Error: batch scoring with mlflow.keras flavor model

I am logging a trained keras model using the following:  fe.log_model( model=model, artifact_path="wine_quality_prediction", flavor= mlflow.keras, training_set=training_set, registered_model_name=model_name )And when I call the following:predictions_...

Machine Learning
FeatureEngineeringClient
keras
mlflow
  • 442 Views
  • 3 replies
  • 0 kudos
Latest Reply
Kaniz
Community Manager
  • 0 kudos

Hi @Miki, The OSError: [Errno 30] Read-only file system typically occurs when you attempt to write to a directory that is read-only or does not exist. Let’s explore some possible solutions: Check the Path: Ensure that the path you’ve provided fo...

  • 0 kudos
2 More Replies
stanjs
by New Contributor III
  • 426 Views
  • 2 replies
  • 0 kudos

BAD_REQUEST: ExperimentIds cannot be empty when checking ACLs in bulk

I encountered the error when using Databricks CE to log experiments from mlflow. It worked perfectly fine before, but now I cannot open any of my experiments. I tried clean the cookies, change the browser, and create a new account to manually create ...

  • 426 Views
  • 2 replies
  • 0 kudos
Latest Reply
Kaniz
Community Manager
  • 0 kudos

Hi @stanjs, I understand that you’re facing issues with accessing your MLflow experiments in Databricks CE. Let’s troubleshoot this together. Here are some steps you can take to resolve the issue: Check Experiment Permissions: With the extension ...

  • 0 kudos
1 More Replies
prafull
by New Contributor
  • 428 Views
  • 1 replies
  • 0 kudos

How to use mlflow to log a composite estimator (multiple pipes) and then deploy it as rest endpoint

Hello,I am trying to deploy a composite estimator as single model, by logging the run with mlflow and registering the model.Can anyone help with how this can be done? This estimator contains different chains-text: data- tfidf- svm- svm.decision_funct...

Screenshot 2024-01-17 000758.png
Machine Learning
ML
mlflow
model
python
  • 428 Views
  • 1 replies
  • 0 kudos
Latest Reply
Kaniz
Community Manager
  • 0 kudos

Hi @prafull , Deploying a composite estimator with MLflow involves several steps. Let’s break it down: Logging the Run with MLflow: First, you’ll need to train your composite estimator using the different pipelines you’ve mentioned (text and cat...

  • 0 kudos
Labels