cancel
Showing results for 
Search instead for 
Did you mean: 
Machine Learning
Dive into the world of machine learning on the Databricks platform. Explore discussions on algorithms, model training, deployment, and more. Connect with ML enthusiasts and experts.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Vlad96
by New Contributor II
  • 1653 Views
  • 3 replies
  • 0 kudos

My model serving endpoint is never getting created

Hello, Im trying to serve a Pyfunc Model on a databricks endpoint but for some reason is getting stuck on a pending status. It's been 4 hours since the endpoint deployment started. If I check the build logs, no error appears whatsoever #23 0.133 chan...

Vlad96_0-1748304777139.png
  • 1653 Views
  • 3 replies
  • 0 kudos
Latest Reply
ThijsBertramCZ
  • 0 kudos

Have you ever found a fix?I am experiencing the same issue

  • 0 kudos
2 More Replies
jayshan
by New Contributor II
  • 96 Views
  • 2 replies
  • 1 kudos

Generic Spark Connect ML error. The fitted or loaded model size is too big.

When I train models in the serverless environment V4 (Premium Plan), the system occasionally returns the error message listed below, especially after running the model training code multiple times. We have tried creating new serverless sessions, whic...

  • 96 Views
  • 2 replies
  • 1 kudos
Latest Reply
jayshan
New Contributor II
  • 1 kudos

This issue seems to be a bug in the serverless environment v4. The serverless instance does not clean up no-use models timely, which leads to insufficient space for new model training. To force resetting the serveless instance, neither the "New Sessi...

  • 1 kudos
1 More Replies
boskicl
by New Contributor III
  • 41 Views
  • 2 replies
  • 0 kudos

mlflow spark load_model fails with FMRegressor Model error on Unity Catalog

We trained a Spark ML FMRegressor model and registered it to Unity Catalog via MLflow. When attempting to load it back using mlflow.spark.load_model, we get anOSError: [Errno 5] Input/output error: '/dbfs/tmp' regardless of what dfs_tmpdir path is pa...

  • 41 Views
  • 2 replies
  • 0 kudos
Latest Reply
boskicl
New Contributor III
  • 0 kudos

from pyspark.ml import PipelineModel mlflow.set_registry_uri("databricks-uc") local_model_path = "/local_disk0/mlflow_model" volume_path = f"/Volumes/{catalogue}/default/mlflow_tmp/sparkml" # Works fine - downloads to driver mlflow.artifacts.downl...

  • 0 kudos
1 More Replies
excavator-matt
by Contributor III
  • 2900 Views
  • 9 replies
  • 3 kudos

Resolved! What is the most efficient way of running sentence-transformers on a Spark DataFrame column?

We're trying to run the bundled sentence-transformers library from SBert in a notebook running Databricks ML 16.4 on an AWS g4dn.2xlarge [T4] instance.However, we're experiencing out of memory crashes and are wondering what the optimal to run sentenc...

Machine Learning
memory issues
sentence-transformers
vector embeddings
  • 2900 Views
  • 9 replies
  • 3 kudos
Latest Reply
excavator-matt
Contributor III
  • 3 kudos

Also, I forgot to mention the workaround solution for the first approach. If you write to parquet in a volume, you can then convert it back to a Delta table in a later cell.Instead of thisprojects_pdf.to_delta("europe_prod_catalog.ad_hoc.project_reco...

  • 3 kudos
8 More Replies
fede_bia
by New Contributor
  • 32 Views
  • 0 replies
  • 0 kudos

Model Serving Only Shows WARNING/ERROR Logs

Hi everyone,I’m deploying a custom model using mlflow.pyfunc.PythonModel in Databricks Model Serving. Inside my wrapper code, I configured logging as follows:logging.basicConfig( stream=sys.stdout, level=logging.INFO, format='%(asctime)s ...

  • 32 Views
  • 0 replies
  • 0 kudos
Bodevan
by New Contributor
  • 59 Views
  • 0 replies
  • 0 kudos

Import CV2 results in Fatal Error

Hey thereThe setup for getting the error can be very basic:-Start a runtime (e.g. 17.3 LTS ML with an Standard_NV36ads_A10_v5 [A10] 440 GB memory, 1GPU)- In a notebook, install the cv2 package like this:%pip install opencv-pythonThis seems to install...

  • 59 Views
  • 0 replies
  • 0 kudos
Dali1
by New Contributor II
  • 86 Views
  • 2 replies
  • 2 kudos

Resolved! Python environment DAB

Hello,I am building a pipeline using DAB.The first step of the dab is to deploy my library as a wheel.The pipeline is run on a shared databricks cluster.When I run the job I see that the job is not using exactly the requirements I specified but it us...

  • 86 Views
  • 2 replies
  • 2 kudos
Latest Reply
stbjelcevic
Databricks Employee
  • 2 kudos

Hi @Dali1, +1 to @pradeep_singh, on shared clusters, tasks inherit cluster-installed libraries, so you won’t get a clean, versioned environment. Use a job cluster (new_cluster) or switch to serverless jobs with an environment per task for isolation. ...

  • 2 kudos
1 More Replies
Dali1
by New Contributor II
  • 58 Views
  • 1 replies
  • 0 kudos

Resolved! Install library in notebook

Hello ,I tried installing a custom library in my databricks notebook that is in a git folder of my worskpace.The installation looks successfulI saw the library in the list of libraries but when I want to import it I have : ModuleNotFoundError: No mod...

  • 58 Views
  • 1 replies
  • 0 kudos
Latest Reply
Dali1
New Contributor II
  • 0 kudos

Just found the issue - The installation with editable mode doesnt work you have to install it as a library I don't know why 

  • 0 kudos
Deep_Blue_Whale
by New Contributor
  • 104 Views
  • 1 replies
  • 0 kudos

Error starting or creating custom model serving endpoints - 'For input string: ""'

Hi Databricks Community,I'm having issues starting or creating custom model serving endpoints. When going into Serving endpoints > Selecting the endpoint > Start, I get the error message 'For input string:'This endpoint had worked correctly yesterday...

Deep_Blue_Whale_1-1771333391364.png Deep_Blue_Whale_2-1771333570626.png
  • 104 Views
  • 1 replies
  • 0 kudos
Latest Reply
emma_s
Databricks Employee
  • 0 kudos

Hi, sorry you're having the issue. You mentioned you've tried to recreate the endpoint with this model and other custom models but still having the same issue. Have you tried serving one of the foundation models and seeing if that works or a really s...

  • 0 kudos
Dali1
by New Contributor II
  • 149 Views
  • 1 replies
  • 0 kudos

Resolved! Databricks SDK vs bundles

Hello,In this article: https://www.databricks.com/blog/from-airflow-to-lakeflow-data-first-orchestrationI understand that if I want to create and deploy ml pipeline in production the recommandation is to use databricks asset bundles. But by using it ...

  • 149 Views
  • 1 replies
  • 0 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 0 kudos

Hi @Dali1 ,When you deploy with Asset Bundles, DABk keeps track of what’s already been deployed and what has changed. That means:it only updates what needs updating,detects drift between your desired state and the workspace,lets you generate plans/di...

  • 0 kudos
AlkaSaliss
by New Contributor II
  • 1148 Views
  • 4 replies
  • 2 kudos

Unable to register Scikit-learn or XGBoost model to unity catalog

Hello, I'm following the tutorial provided here https://docs.databricks.com/aws/en/notebooks/source/mlflow/mlflow-classic-ml-e2e-mlflow-3.html for ML model management process using ML FLow, in a unity-catalog enabled workspace, however I'm facing an ...

  • 1148 Views
  • 4 replies
  • 2 kudos
Latest Reply
joelramirezai
Databricks Employee
  • 2 kudos

You need to ensure that your Unity Catalog catalog and schema already exist, that you have the necessary permissions to use them, and that you update the code to reference your own catalog and schema names. You must also run on a classic cluster with...

  • 2 kudos
3 More Replies
venkatkittu
by New Contributor
  • 60 Views
  • 0 replies
  • 0 kudos

Facing 132 error in model serving while using faiss

Hi,I have been trying to deply a rest endpoint for my application using the model serving feature, I have registered my model on the unity catalog and when trying to serve the model it is getting sucess when I removed the code related faiss and when ...

  • 60 Views
  • 0 replies
  • 0 kudos
p4pratikjain
by Contributor
  • 4202 Views
  • 3 replies
  • 0 kudos

DAB - Add/remove task depending on workspace.

I use DAB for deploying Jobs, I want to add a specific Task in dev only but not in staging or prod. Is there any way to achieve this using DAB ?

  • 4202 Views
  • 3 replies
  • 0 kudos
Latest Reply
Pat
Esteemed Contributor
  • 0 kudos

I know it's a bit old, but if someone is looking into a solution, then I was able to resolve the issue where I need to deploy some jobs only into the DEV target:https://github.com/databricks/bundle-examples/tree/main/knowledge_base/target_includes.Us...

  • 0 kudos
2 More Replies
lschneid
by New Contributor II
  • 109 Views
  • 0 replies
  • 1 kudos

Replacing a Monolithic MLflow Serving Pipeline with Composed Models in Databricks

Hi everyone,I’m a senior MLE and recently joined a company where all data science and ML workloads run on Databricks. My background is mostly MLOps on Kubernetes, so I’m currently ramping up on Databricks and trying to improve the architecture of som...

  • 109 Views
  • 0 replies
  • 1 kudos
Labels