cancel
Showing results for 
Search instead for 
Did you mean: 
Machine Learning
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Kaan
by New Contributor
  • 886 Views
  • 1 replies
  • 1 kudos

Resolved! Using databricks in multi-cloud, and querying data from the same instance.

I'm looking for a good product to use across two clouds at once for Data Engineering, Data modeling and governance. I currently have a GCP platform, but most of my data and future data goes through Azure, and currently is then transfered to GCS/BQ.Cu...

  • 886 Views
  • 1 replies
  • 1 kudos
Latest Reply
Anonymous
Not applicable
  • 1 kudos

@Karl Andrén​ :Databricks is a great option for data engineering, data modeling, and governance across multiple clouds. It supports integrations with multiple cloud providers, including Azure, AWS, and GCP, and provides a unified interface to access ...

  • 1 kudos
Hubert-Dudek
by Esteemed Contributor III
  • 485 Views
  • 1 replies
  • 7 kudos

Have you heard about databricks latest open-source language model called Dolly? It’s a ChatGPT like model that uses the tatsu-lab/alpaca dataset with ...

Have you heard about databricks latest open-source language model called Dolly? It’s a ChatGPT like model that uses the tatsu-lab/alpaca dataset with examples of questions and answers. To train Dolly, you can combine this dataset (simple solution on ...

Screenshot 2023-03-26 215509
  • 485 Views
  • 1 replies
  • 7 kudos
Latest Reply
Anonymous
Not applicable
  • 7 kudos

Thanks for posting this! I am so excited about the possibilities that this can do for us. It's an exciting development in the natural language processing field, and it has the potential to be a valuable tool for businesses looking to implement chatb...

  • 7 kudos
alisher_pwc
by New Contributor II
  • 1579 Views
  • 2 replies
  • 1 kudos

Model serving with GPU cluster

Hello Databricks community!We are facing a strong need of serving some of public and our private models on GPU clusters and we have several requirements:1) We'd like to be able to start/stop the endpoints (best with scheduling) to avoid excess consum...

  • 1579 Views
  • 2 replies
  • 1 kudos
Latest Reply
Vartika
Moderator
  • 1 kudos

Hi @Alisher Akh​ Does @Debayan Mukherjee​'s answer help? If yes, would you be happy to mark the answer as best so that other members can find the solution more quickly? If not, please tell us so we can help you further. Cheers!

  • 1 kudos
1 More Replies
fuselessmatt
by Contributor
  • 7210 Views
  • 5 replies
  • 6 kudos

Resolved! What does "Command exited with code 50 mean" and how do you solve it?

Hi!We have this dbt model that generates a table with user activity in the previous days, but we get this vague error message in the Databricks SQL Warehouse.Job aborted due to stage failure: Task 3 in stage 4267.0 failed 4 times, most recent failure...

  • 7210 Views
  • 5 replies
  • 6 kudos
Latest Reply
shan_chandra
Honored Contributor III
  • 6 kudos

@Mattias P​  - For the executor lost failure, is it trying to bring in large data volume? can you please reduce the date range and try? or run the workload on a bigger DBSQL warehouse than the current one.

  • 6 kudos
4 More Replies
Ajay-Pandey
by Esteemed Contributor III
  • 838 Views
  • 2 replies
  • 5 kudos

Share information between tasks in a Databricks job  You can use task values to pass arbitrary parameters between tasks in a Databricks job. You pass ...

Share information between tasks in a Databricks jobYou can use task values to pass arbitrary parameters between tasks in a Databricks job. You pass task values using the taskValues subutility in Databricks Utilities. The taskValues subutility provide...

  • 838 Views
  • 2 replies
  • 5 kudos
Latest Reply
newforesee
New Contributor II
  • 5 kudos

We urgently hope for this feature, but to date, we have found that it is only available in Python. Do you have any plans to support Scala?

  • 5 kudos
1 More Replies
apatel
by New Contributor III
  • 5246 Views
  • 2 replies
  • 0 kudos

Resolved! How to resolve this error "Error: cannot create global init script: default auth: cannot configure default credentials"

I'm trying to set the global init script via my Terraform deployment. I did a thorough google search and can't seem to find guidance here.I'm using a very generic call to set these scripts in my TF Deployment.terraform { required_providers { data...

  • 5246 Views
  • 2 replies
  • 0 kudos
Latest Reply
apatel
New Contributor III
  • 0 kudos

Ok in case this helps anyone else, I've managed to resolve.I confirmed in this documentation the databricks CLI is required locally, wherever this is being executed. https://learn.microsoft.com/en-us/azure/databricks/dev-tools/terraform/cluster-note...

  • 0 kudos
1 More Replies
Koliya
by New Contributor II
  • 10696 Views
  • 5 replies
  • 7 kudos

The Python process exited with exit code 137 (SIGKILL: Killed). This may have been caused by an OOM error. Check your command's memory usage.

I am running a hugging face model on a GPU cluster (g4dn.xlarge, 16GB Memory, 4 cores). I run the same model in four different notebooks with different data sources. I created a workflow to run one model after the other. These notebooks run fine indi...

  • 10696 Views
  • 5 replies
  • 7 kudos
Latest Reply
fkemeth
New Contributor II
  • 7 kudos

You might accumulate gradients when running your Huggingface model, which typically leads to out-of-memory errors after some iterations. If you use it for inference only, dowith torch.no_grad(): # The code where you apply the model

  • 7 kudos
4 More Replies
Tilo
by New Contributor
  • 1785 Views
  • 3 replies
  • 3 kudos

Resolved! MLFlow: How to load results from model and continue training

I'd like to continue / finetune training of an existing keras/tensorflow model. We use MLFlow to store the model. How can I load the wieght from an existing model to the model and continue "fit" preferable with a different learning rate.Just loading ...

  • 1785 Views
  • 3 replies
  • 3 kudos
Latest Reply
Anonymous
Not applicable
  • 3 kudos

Hi @Tilo Wünsche​ Hope all is well! Just wanted to check in if you were able to resolve your issue and would you be happy to share the solution or mark an answer as best? Else please let us know if you need more help. We'd love to hear from you.Thank...

  • 3 kudos
2 More Replies
NSRBX
by Contributor
  • 2511 Views
  • 6 replies
  • 6 kudos

Resolved! Error loading model from mlflow: java.io.StreamCorruptedException: invalid type code: 00

Hello,I'm using, in my IDE, Databricks Connect version 9.1LTS ML to connect to a databricks cluster with spark version 3.1 and download a spark model that's been trained and saved using mlflow.So it seems like it's able to find a copy the model, but ...

  • 2511 Views
  • 6 replies
  • 6 kudos
Latest Reply
NSRBX
Contributor
  • 6 kudos

Hi @Kaniz Fatma​ and @Shanmugavel Chandrakasu​,It works after putting hadoop.dll into C:\Windows\System32 folder.I have hadoop version 3.3.1.I already had winutils.exe in the Hadoop bin folder.RegardsNath

  • 6 kudos
5 More Replies
Mike_sb
by New Contributor III
  • 1610 Views
  • 7 replies
  • 4 kudos
  • 1610 Views
  • 7 replies
  • 4 kudos
Latest Reply
Amit_352107
New Contributor III
  • 4 kudos

Hi @Mike M​ Kindly clear cache, and your issue will be resolved

  • 4 kudos
6 More Replies
bruno_valero
by New Contributor II
  • 3263 Views
  • 2 replies
  • 1 kudos

How to download a .csv or .pkl file from databricks?

When I save files on "dbfs:/FileStore/shared_uploads/brunofvn6@gmail.com/", it doesn't appear anywhere in my workspace. I've tried to copy the path of the workspace with the right mouse button, pasted on ("my pandas dataframe").to_csv('path'), but wh...

  • 3263 Views
  • 2 replies
  • 1 kudos
Latest Reply
bruno_valero
New Contributor II
  • 1 kudos

I think I discover how to do this. Is in the label called data in the left menu of the databricks environment, in the top left of the menu there are two labels "Database Tables" and "DBFS" in which "Database Table" is the default label. So it is just...

  • 1 kudos
1 More Replies
Data_Analytics1
by Contributor III
  • 664 Views
  • 1 replies
  • 2 kudos

Resolved! File not found error. Does OPTIMIZE deletes initial versions of the delta table?

df = (spark.readStream.format("delta")\              .option("readChangeFeed", "true")\              .option("startingVersion", 1)\              .table("CatalogName.SchemaName.TableName")             )display(df)A file referenced in the transaction l...

  • 664 Views
  • 1 replies
  • 2 kudos
Latest Reply
swethaNandan
New Contributor III
  • 2 kudos

Had you run vacuum on the table? Vacuum can clean up data files marked for removal and are older than retention period.Optimize compacts files and marks the small files for removal, but does not physically remove the data files

  • 2 kudos
User16461610613
by New Contributor II
  • 1170 Views
  • 1 replies
  • 2 kudos

Free Databricks Training on AWS, Azure, or Google Cloud Good news! You can now access free, in-depth Databricks training on AWS, Azure or Google Cloud...

Free Databricks Training on AWS, Azure, or Google CloudGood news! You can now access free, in-depth Databricks training on AWS, Azure or Google Cloud.  Our on-demand training series walks through how to:Streamline data ingest and management to build ...

image
  • 1170 Views
  • 1 replies
  • 2 kudos
Latest Reply
jose_gonzalez
Moderator
  • 2 kudos

Thank you for sharing this!!

  • 2 kudos
obiamaka
by New Contributor III
  • 2124 Views
  • 6 replies
  • 2 kudos

Resolved! Unable to view jobs in Databricks Workflow

This has been happening for about 2 days now. Any ideas on why this occurs and how it can be resolved. I attached a screenshot.

  • 2124 Views
  • 6 replies
  • 2 kudos
Latest Reply
obiamaka
New Contributor III
  • 2 kudos

This issue got resolved on it's own, but not sure what the problem was, probably a bug from a software update?

  • 2 kudos
5 More Replies
zachclem
by New Contributor III
  • 2172 Views
  • 2 replies
  • 1 kudos

Resolved! Logging model to MLflow using Feature Store API. Getting TypeError: join() argument must be str, bytes, or os.PathLike object, not 'dict'

I'm using databricks. Trying to log a model to MLflow using the Feature Store log_model function. but I have this error: TypeError: join() argument must be str, bytes, or os.PathLike object, not 'dict' I'am using the Databricks runtime ml (10.4 LTS M...

  • 2172 Views
  • 2 replies
  • 1 kudos
Latest Reply
zachclem
New Contributor III
  • 1 kudos

I updated by Databricks Run Time from 10.4 to 12.1 and this solved the issue.

  • 1 kudos
1 More Replies
Labels