cancel
Showing results for 
Search instead for 
Did you mean: 
Machine Learning
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

venkat09
by New Contributor III
  • 960 Views
  • 1 replies
  • 1 kudos

Resolved! Running into an issue while setting up dbx locally?

Followed the documentation and facing issue while running dbx execute on all-purpose/interactive cluster, which is up and running already. Ran this command  dbx execute --cluster-id=XXXXXX --job=dbx-demo-job --no-rebuild --debug. If anyone faced it ...

image
  • 960 Views
  • 1 replies
  • 1 kudos
Latest Reply
venkat09
New Contributor III
  • 1 kudos

before running package it as wheel before running `dbx execute` fix the issue

  • 1 kudos
Nithin
by New Contributor II
  • 7199 Views
  • 14 replies
  • 4 kudos

Resolved! How to access databricks feature store outside databricks?

We are building the feature store using databricks API. Few of the machine learning engineers are using Jupyter notebooks. Is it possible to use feature store outside databricks?

  • 7199 Views
  • 14 replies
  • 4 kudos
Latest Reply
datariel
New Contributor II
  • 4 kudos

Hi @Kaniz Fatma​  and @Jose Gonzalez​ ,turning back to the original question, and considering that one of the main benefits of the Feature Store is the removal of the online/offline skew, how could I access to the features from a client application l...

  • 4 kudos
13 More Replies
naveen_marthala
by Contributor
  • 2165 Views
  • 4 replies
  • 2 kudos

Resolved! why does the client need to have git installed for auto-logging to an mlflow server running in "--serve-artifacts" mode?

I have an mlflow server with `--serve-artifacts` and with postgres as `--backend-store-uri`. The machine(container with base image python:3.9-bullseye) running the server has git installed which is available on path. I am logging from jupyter-noteboo...

  • 2165 Views
  • 4 replies
  • 2 kudos
Latest Reply
Kaniz
Community Manager
  • 2 kudos

Hi @Naveen Marthala​ â€‹ , Just a friendly follow-up. Do you still need help or the above responses help you to find the solution? Please let us know.

  • 2 kudos
3 More Replies
naveen_marthala
by Contributor
  • 4530 Views
  • 2 replies
  • 3 kudos

Resolved! How to PREVENT mlflow's autologging from logging ALL runs?

I am logging runs from jupyter notebook. the cells which has `mlflow.sklearn.autlog()` behaves as expected. but, the cells which has .fit() method being called on sklearn's estimators are also being logged as runs without explicitly mentioning `mlflo...

  • 4530 Views
  • 2 replies
  • 3 kudos
Latest Reply
Anonymous
Not applicable
  • 3 kudos

https://apkmiz.com/showbox-apk-old-latest-version-android/

  • 3 kudos
1 More Replies
Direo
by Contributor
  • 5262 Views
  • 4 replies
  • 2 kudos

Resolved! xgboost 1.5.1 gives 'XGBModel' object has no attribute 'enable_categorical' error

Should I pip install xgboost==1.4.2. (the last version it worked) or is there a better way to solve it having in mind that this solution might cause problems later if this version of xgboost is not supported on future python versions.

  • 5262 Views
  • 4 replies
  • 2 kudos
Latest Reply
Direo
Contributor
  • 2 kudos

Hi, @Kaniz Fatma​. No, I have found a solution. Needed to retrain models using new version of xgboost.

  • 2 kudos
3 More Replies
mradassaad
by New Contributor III
  • 2008 Views
  • 4 replies
  • 1 kudos

Resolved! Tuning `CrossValidator` spark job performance

I am running a 3-fold cross validation of an ML pipeline that utilizes `GBTClassifier` as the final step. It takes 18 hours to run and I am looking for feedback into how to improve the performance as I expect this to go faster.For context here is the...

Random Forest Job Random Forest Job Summary GBT storage top half
  • 2008 Views
  • 4 replies
  • 1 kudos
Latest Reply
Kaniz
Community Manager
  • 1 kudos

Hi @Assaad Mrad​  , Just a friendly follow-up. Do you still need help, or @Chris Chalcraft​ 's response help you to find the solution? Please let us know.

  • 1 kudos
3 More Replies
Mr__E
by Contributor II
  • 398 Views
  • 0 replies
  • 0 kudos

Custom AutoML evaluation metric for ranking model

I built a model which is used for ranking and I have a notebook that takes that model to generate rankings and then uses a UDF-based metric to evaluate those rankings. Is there any way that I can have this ranking / UDF be used during the AutoML trai...

  • 398 Views
  • 0 replies
  • 0 kudos
findinpath
by Contributor
  • 2780 Views
  • 9 replies
  • 5 kudos

How to mount s3 bucket in community edition cluster?

I'm using Databricks Community Edition for testing purposes on a OSS project.I'm spinning up the cluster automatically through Databricks Clusters API.The automated tests rely on AWS S3 infrastructure, reason why I need to mount the S3 bucket on the ...

  • 2780 Views
  • 9 replies
  • 5 kudos
Latest Reply
findinpath
Contributor
  • 5 kudos

I haven't found any solution.I'm assuming that currently my only option is the usage of Databricks Enterprise to model scenarios involving the mounting of object storage buckets.

  • 5 kudos
8 More Replies
Wayne
by New Contributor III
  • 828 Views
  • 2 replies
  • 3 kudos
  • 828 Views
  • 2 replies
  • 3 kudos
Latest Reply
Wayne
New Contributor III
  • 3 kudos

No error, just seeing the EXPAND DISK in cluster event logs. This is just a regular spark application. I am not sure if the cloud storage matters - a spark application uses it as input and output.

  • 3 kudos
1 More Replies
romanzdk
by New Contributor II
  • 1333 Views
  • 1 replies
  • 0 kudos

Databricks online store - Login to Azure SQL Database with Service Principal

I want to use Databricks Online Store with Azure SQL Database, however I am unable to autenthicate through Databricks Feature Store API. I need to use Service Principal credentials.I tried using Application ID as username and Secret as password, but ...

  • 1333 Views
  • 1 replies
  • 0 kudos
Latest Reply
romanzdk
New Contributor II
  • 0 kudos

no one?

  • 0 kudos
mhansinger
by New Contributor II
  • 11044 Views
  • 6 replies
  • 1 kudos

Resolved! Set default "spark.driver.maxResultSize" from the notebook

Hello,I would like to set the default "spark.driver.maxResultSize" from the notebook on my cluster. I know I can do that in the cluster settings, but is there a way to set it by code?I also know how to do it when I start a spark session, but in my ca...

  • 11044 Views
  • 6 replies
  • 1 kudos
Latest Reply
Anonymous
Not applicable
  • 1 kudos

Hi @Maximilian Hansinger​ Just wanted to check in if you were able to resolve your issue. If yes, would you be happy to mark the answer as best? If not, please tell us so we can help you.Thanks!

  • 1 kudos
5 More Replies
EdoardoVivo
by New Contributor
  • 1010 Views
  • 0 replies
  • 0 kudos

Pymc3 on Databricks: Progress bar

Hello everybody..I am trying to run pymc3 models on Databricks (runtime 9.1) and when I start the sampling process, the progress bar is not showing. It is a bit annoying since this way I do not have any information on when the process is going to end...

  • 1010 Views
  • 0 replies
  • 0 kudos
Vijeth
by New Contributor II
  • 2649 Views
  • 2 replies
  • 2 kudos

Resolved! How to deploy or create mlflow model as docker image with REST api endpoint within databricks?

Is it possible to create mlflow model as a docker image with REST api endpoint and use it for inferencing within databricks or hosting the image in azure container instances?

  • 2649 Views
  • 2 replies
  • 2 kudos
Latest Reply
Kaniz
Community Manager
  • 2 kudos

Hi @Vijeth Moudgalya​ â€‹ , Just a friendly follow-up. Did you follow @Bilal Aslam​ 's suggestion? Please let us know.

  • 2 kudos
1 More Replies
Vik1
by New Contributor II
  • 2260 Views
  • 4 replies
  • 2 kudos

Resolved! Cluster setup for ML work for Pandas in Spark, and vanilla Python.

My setup:Worker type: Standard_D32d_v4, 128 GB Memory, 32 Cores, Min Workers: 2, Max Workers: 8Driver type: Standard_D32ds_v4, 128 GB Memory, 32 CoresDatabricks Runtime Version: 10.2 ML (includes Apache Spark 3.2.0, Scala 2.12)I ran a snowflake quer...

  • 2260 Views
  • 4 replies
  • 2 kudos
Latest Reply
Anonymous
Not applicable
  • 2 kudos

Hey there @Vivek Ranjan​ Checking in. If Joseph's answer helped, would you let us know and mark the answer as best?  It would be really helpful for the other members to find the solution more quickly.Thanks!

  • 2 kudos
3 More Replies
wchen
by New Contributor II
  • 5071 Views
  • 7 replies
  • 2 kudos

Resolved! In Databricks, the Python kafka consumer app in notebook to Confluent Cloud having the issue captured in the Body of question: SASL/PLAIN authentication being used

kafkashaded.org.apache.kafka.common.KafkaException: Failed to construct kafka consumer at kafkashaded.org.apache.kafka.clients.consumer.KafkaConsumer.<init>(KafkaConsumer.java:823) at kafkashaded.org.apache.kafka.clients.consumer.KafkaConsumer.<init>...

  • 5071 Views
  • 7 replies
  • 2 kudos
Latest Reply
bigdata70
New Contributor III
  • 2 kudos

@Kaniz Fatma​ I am having the same issue.%python import pyspark.sql.functions as fn from pyspark.sql.types import StringType   binary_to_string = fn.udf(lambda x: str(int.from_bytes(x, byteorder='big')), StringType())   df = spark.readStream.format("...

  • 2 kudos
6 More Replies
Labels