cancel
Showing results for 
Search instead for 
Did you mean: 
Machine Learning
Dive into the world of machine learning on the Databricks platform. Explore discussions on algorithms, model training, deployment, and more. Connect with ML enthusiasts and experts.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Vlad96
by Databricks Partner
  • 1934 Views
  • 3 replies
  • 0 kudos

My model serving endpoint is never getting created

Hello, Im trying to serve a Pyfunc Model on a databricks endpoint but for some reason is getting stuck on a pending status. It's been 4 hours since the endpoint deployment started. If I check the build logs, no error appears whatsoever #23 0.133 chan...

Vlad96_0-1748304777139.png
  • 1934 Views
  • 3 replies
  • 0 kudos
Latest Reply
ThijsBertramCZ
New Contributor II
  • 0 kudos

Have you ever found a fix?I am experiencing the same issue

  • 0 kudos
2 More Replies
excavator-matt
by Contributor III
  • 3879 Views
  • 9 replies
  • 3 kudos

Resolved! What is the most efficient way of running sentence-transformers on a Spark DataFrame column?

We're trying to run the bundled sentence-transformers library from SBert in a notebook running Databricks ML 16.4 on an AWS g4dn.2xlarge [T4] instance.However, we're experiencing out of memory crashes and are wondering what the optimal to run sentenc...

Machine Learning
memory issues
sentence-transformers
vector embeddings
  • 3879 Views
  • 9 replies
  • 3 kudos
Latest Reply
excavator-matt
Contributor III
  • 3 kudos

Also, I forgot to mention the workaround solution for the first approach. If you write to parquet in a volume, you can then convert it back to a Delta table in a later cell.Instead of thisprojects_pdf.to_delta("europe_prod_catalog.ad_hoc.project_reco...

  • 3 kudos
8 More Replies
Dali1
by New Contributor III
  • 507 Views
  • 2 replies
  • 2 kudos

Resolved! Python environment DAB

Hello,I am building a pipeline using DAB.The first step of the dab is to deploy my library as a wheel.The pipeline is run on a shared databricks cluster.When I run the job I see that the job is not using exactly the requirements I specified but it us...

  • 507 Views
  • 2 replies
  • 2 kudos
Latest Reply
stbjelcevic
Databricks Employee
  • 2 kudos

Hi @Dali1, +1 to @pradeep_singh, on shared clusters, tasks inherit cluster-installed libraries, so you won’t get a clean, versioned environment. Use a job cluster (new_cluster) or switch to serverless jobs with an environment per task for isolation. ...

  • 2 kudos
1 More Replies
Dali1
by New Contributor III
  • 386 Views
  • 1 replies
  • 0 kudos

Resolved! Install library in notebook

Hello ,I tried installing a custom library in my databricks notebook that is in a git folder of my worskpace.The installation looks successfulI saw the library in the list of libraries but when I want to import it I have : ModuleNotFoundError: No mod...

  • 386 Views
  • 1 replies
  • 0 kudos
Latest Reply
Dali1
New Contributor III
  • 0 kudos

Just found the issue - The installation with editable mode doesnt work you have to install it as a library I don't know why 

  • 0 kudos
Deep_Blue_Whale
by New Contributor
  • 314 Views
  • 1 replies
  • 0 kudos

Error starting or creating custom model serving endpoints - 'For input string: ""'

Hi Databricks Community,I'm having issues starting or creating custom model serving endpoints. When going into Serving endpoints > Selecting the endpoint > Start, I get the error message 'For input string:'This endpoint had worked correctly yesterday...

Deep_Blue_Whale_1-1771333391364.png Deep_Blue_Whale_2-1771333570626.png
  • 314 Views
  • 1 replies
  • 0 kudos
Latest Reply
emma_s
Databricks Employee
  • 0 kudos

Hi, sorry you're having the issue. You mentioned you've tried to recreate the endpoint with this model and other custom models but still having the same issue. Have you tried serving one of the foundation models and seeing if that works or a really s...

  • 0 kudos
Dali1
by New Contributor III
  • 456 Views
  • 1 replies
  • 0 kudos

Resolved! Databricks SDK vs bundles

Hello,In this article: https://www.databricks.com/blog/from-airflow-to-lakeflow-data-first-orchestrationI understand that if I want to create and deploy ml pipeline in production the recommandation is to use databricks asset bundles. But by using it ...

  • 456 Views
  • 1 replies
  • 0 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 0 kudos

Hi @Dali1 ,When you deploy with Asset Bundles, DABk keeps track of what’s already been deployed and what has changed. That means:it only updates what needs updating,detects drift between your desired state and the workspace,lets you generate plans/di...

  • 0 kudos
AlkaSaliss
by New Contributor II
  • 1340 Views
  • 4 replies
  • 2 kudos

Unable to register Scikit-learn or XGBoost model to unity catalog

Hello, I'm following the tutorial provided here https://docs.databricks.com/aws/en/notebooks/source/mlflow/mlflow-classic-ml-e2e-mlflow-3.html for ML model management process using ML FLow, in a unity-catalog enabled workspace, however I'm facing an ...

  • 1340 Views
  • 4 replies
  • 2 kudos
Latest Reply
joelramirezai
Databricks Employee
  • 2 kudos

You need to ensure that your Unity Catalog catalog and schema already exist, that you have the necessary permissions to use them, and that you update the code to reference your own catalog and schema names. You must also run on a classic cluster with...

  • 2 kudos
3 More Replies
p4pratikjain
by Contributor
  • 4272 Views
  • 3 replies
  • 0 kudos

DAB - Add/remove task depending on workspace.

I use DAB for deploying Jobs, I want to add a specific Task in dev only but not in staging or prod. Is there any way to achieve this using DAB ?

  • 4272 Views
  • 3 replies
  • 0 kudos
Latest Reply
Pat
Esteemed Contributor
  • 0 kudos

I know it's a bit old, but if someone is looking into a solution, then I was able to resolve the issue where I need to deploy some jobs only into the DEV target:https://github.com/databricks/bundle-examples/tree/main/knowledge_base/target_includes.Us...

  • 0 kudos
2 More Replies
Danik
by Databricks Partner
  • 1213 Views
  • 2 replies
  • 3 kudos

Resolved! Population stability index (PSI) calculation in Lakehouse monitor

Hi! We are using Lakehouse monitoring for detecting data drift in our metrics. However, the exact calculation of metrics is not documented anywhere (I couldnt find it) and it raises questions on how they are done, in our case especially - PSI. I woul...

  • 1213 Views
  • 2 replies
  • 3 kudos
Latest Reply
iyashk-DB
Databricks Employee
  • 3 kudos

Hi @Danik , I have reviewed this. 1) Is there documentation for PSI and other metrics?Public docs list PSI in the drift table and give thresholds, but don’t detail the exact algorithm.Internally, numeric PSI uses ~1000 quantiles, equal‑height binning...

  • 3 kudos
1 More Replies
naveen_marthala
by Contributor
  • 14846 Views
  • 5 replies
  • 3 kudos

Resolved! How to PREVENT mlflow's autologging from logging ALL runs?

I am logging runs from jupyter notebook. the cells which has `mlflow.sklearn.autlog()` behaves as expected. but, the cells which has .fit() method being called on sklearn's estimators are also being logged as runs without explicitly mentioning `mlflo...

  • 14846 Views
  • 5 replies
  • 3 kudos
Latest Reply
alexsheer9003
New Contributor II
  • 3 kudos

NICE TIP! 

  • 3 kudos
4 More Replies
thomasm
by New Contributor II
  • 537 Views
  • 4 replies
  • 1 kudos

MLFlow Detailed Trace view doesn't work in some workspaces

I've created a Databricks Model Serving Endpoint which serves an MLFlow Pyfunc model. The model uses langchain and I'm using mlflow.langchain.autolog().At my company we have some production(-like) workspaces where users cannot e.g. run Notebooks and ...

thomasm_1-1767785859607.png thomasm_0-1767785737567.png thomasm_2-1767785939124.png
  • 537 Views
  • 4 replies
  • 1 kudos
Latest Reply
thomasm
New Contributor II
  • 1 kudos

Hi Jahnavi,Thanks for your reply. I think the issues you mentioned are not the cause of the discrepancy though. I have attached a screenshot of the same trace ID when displayed in the Experiments UI (where I cannot get a detailed trace view) and in t...

  • 1 kudos
3 More Replies
tonybenzu99
by New Contributor II
  • 1444 Views
  • 2 replies
  • 3 kudos

Resolved! Is Delta Lake deeply tested in Professional Data Engineer Exam?

I wanted to ask people who have already taken the Databricks Certified Professional Data Engineer exam whether Delta Lake is tested in depth or not. While preparing, I’m currently using the Databricks Certified Professional Data Engineer sample quest...

  • 1444 Views
  • 2 replies
  • 3 kudos
Latest Reply
lucafredo
New Contributor III
  • 3 kudos

Yes, Delta Lake concepts are an important part of the Databricks Professional Data Engineer exam, but they aren’t tested in extreme depth compared to core Spark transformations and data pipeline design. The exam mainly focuses on practical understand...

  • 3 kudos
1 More Replies
ryojikn
by New Contributor III
  • 2061 Views
  • 3 replies
  • 2 kudos

Model Serving - Shadow Deployment - Azure

Hey,I'm composing an architecture within the usage of Model Serving Endpoints and one of the needs that we're aiming to resolve is Shadow Deployment.Currently, it seems that the traffic configurations available in model serving do not allow this type...

  • 2061 Views
  • 3 replies
  • 2 kudos
Latest Reply
KaushalVachhani
Databricks Employee
  • 2 kudos

@ryojikn and @irtizak , you’re right. Databricks Model Serving allows splitting traffic between model versions, but it doesn’t have a true shadow deployment where live production traffic is mirrored to a new model for monitoring without affecting use...

  • 2 kudos
2 More Replies
jitenjha11
by Databricks Partner
  • 387 Views
  • 2 replies
  • 3 kudos

Getting error when running databricks deploy bundle command

HI all,I am trying to implement MLOps project using https://github.com/databricks/mlops-stacks repo.I have created azure databricks with Premium (+ Role-based access controls) (Click to change) and following bundle creation and deploy using uRL: http...

  • 387 Views
  • 2 replies
  • 3 kudos
Latest Reply
iyashk-DB
Databricks Employee
  • 3 kudos

This is expected behavior with mlops-stacks and not an issue with your Terraform version or the CLI. The main problem is that your Azure Databricks workspace does not have Unity Catalog enabled or assigned. The mlops-stacks templates assume Unity Cat...

  • 3 kudos
1 More Replies
Suheb
by Contributor
  • 539 Views
  • 2 replies
  • 2 kudos

Why does my MLflow model training job fail on Databricks with an out‑of‑memory error for large datas

I am trying to train a machine learning model using MLflow on Databricks. When my dataset is very large, the training stops and gives an ‘out-of-memory’ error. Why does this happen and how can I fix it?

  • 539 Views
  • 2 replies
  • 2 kudos
Latest Reply
iyashk-DB
Databricks Employee
  • 2 kudos

+1 to what @mukul1409 has told. Please follow the guides below to distribute the training: https://docs.databricks.com/aws/en/machine-learning/train-model/distributed-training/spark-pytorch-d... https://docs.databricks.com/aws/en/notebooks/source/dee...

  • 2 kudos
1 More Replies
Labels