Hi,I'm trying to enable inference table for my llama_2_7b_hf serving endpoint, however I'm getting the following error:"Inference tables are currently not available with accelerated inference." Anyone one have an idea on how to overcome this issue? C...
Thank you for posting your question in our community! We are happy to assist you.To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best answers your question?This...
Using `mlflow.pyfunc.log_model` to save a model in the model registry fails with the following error: `PicklingError: Can't pickle <built-in function input>: it's not the same object as builtins.input`I've reproduced it locally and got to the conclus...
Thank you for posting your question in our community! We are happy to assist you.To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best answers your question?This...
I want to set environment variables for serving-endpoints that are not being read as secrets by using the DB API. As an example I want to set the OPENAI_API_BASE for a LangChain model. Is this possible?
Thank you for posting your question in our community! We are happy to assist you.To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best answers your question?This...
I am trying to deploy a Vector Search Index using both the UI and the Python VectorSearchClient. In both cases the command is successful but in the Catalog Explorer the newly created index stalls with the status 'Provisioning Index' for hours. Previo...
I am trying to run an AutoML experiment but the button stays greyed out no matter what I do. I've tried different cluster configurations, different datasets, even blew away the instance in Azure and re-created it across two different Azure accounts s...
Hi,I'm trying to use around 5 numerical features on 3.5 million rows to train and test my model with a spark data frame.My cluster has 60 nodes available but is only using 2. How can I distribute the process or make it for efficient and faster.My cod...
@mohaimen_syed - can you please try using pyspark.ml implementation of randomForestClassifier instead of sklearn and see if it works. Below is an example - https://github.com/apache/spark/blob/master/examples/src/main/python/ml/random_forest_classif...
Hi!Last week I was able to perform some experiments in autoML. Today even after completing all the values in the form, the start button stay greyed. (I try train with the same dataset and parameters that last week train ok, with no luck) have any con...
@Hypatia - can you please detailed out the issue further? What was changed between previous good run and now when entering the details on AutoML experiment page?
Hello,I'm working through the Dolly demo that was released a few days ago. There are a few lines of code that I have had to alter in order to make it work in my environment. However, I've been trying to run similarity_search() on the generated Chroma...
As written in the command output when pip installing the python packages, you should then run a `dbutils.library.restartPython()`. Please try the following commands:%pip install -U transformers langchain chromadb accelerate bitsandbytesdbutils.librar...
Hey everyone,we were trying to start a new AutoML experiment but the 'Start AutoML' button won't activate. We've also noticed that the number of selected trainingset columns does'nt change even if some columns are unselected.Has anybody encountered t...
This seems to be a Databricks-wide problem. Please see related ticket on https://community.databricks.com/t5/machine-learning/can-t-run-an-automl-experiment-because-button-is-greyed-out/td-p/57388
Hi,I am currently transitioning some of our MLops workflows to use models registered in Unity Catalog (UC) instead of the "normal" model registry. I am trying to follow the instructions on https://docs.gcp.databricks.com/en/machine-learning/manage-mo...
Hello,I'm following courses on the Databricks academy and using for that purpose the Databricks Community edition using a runtime 12.2 LTS (includes Apache Spark 3.3.2, Scala 2.12) and I believe it can't be changedI'm following the Data engineering c...
I was facing the same error. This could be resolved by adding the version that you are currently working with in the config function present in '_common' notebook in the "Includes' folder. (This was the case of my folder structure that I downloaded f...
Hi everyone! Has anyone successfully deployed a model saved on Unity Catalog to Model Serving? I get:Event Log: Served model creation failed for served model 'model', config version 15. Error message: Container creation failed. Please see build logs ...
I get this below error while trying to simulate kinesis streams as mentioned in Databricks documentation at https://docs.databricks.com/getting-started/streaming.htmlError:java.nio.file.AccessDeniedException:Amazon S3; Status Code: 403; Error Code: A...
If you do spark.sparkContext._jsc.hadoopConfiguration().set("fs.s3a.access.key", AWS_ACCESS_KEY_ID) + secret with any other secret that has less access than your default one this sometimes happens, so running those commands but with your normal secre...
Hi Team,we have use case to run ML use case using DLT UC, we are facing library issues Error: INVALID ARGUMENT: no module named ‘importlib_metadata’we installed them manually by passing import in notebooks, but we have lot of missing libraries and we...
Hello I am trying to create a text classification model based on this blog https://www.databricks.com/blog/rapid-nlp-development-databricks-delta-and-transformers and the notebook accelerator. I just changed the model to take a french bert but i cann...
Hello Anasse, the LLM landscape has changed drastically since that blog was released in mid-2022. We have new, updated guidance which you can find here (make sure you check out the Next Steps section for the RAG Demo link as well). Additionally, if y...