I saw that Foundation Model API is not HIPAA compliant. Is there a timeline in which we could expect it to be HIPAA compliant? I work for a healthcare company with a BAA with Databricks.
Hi @yhyhy3
Foundation Model API's HIPAA certification:AWS: e.t.a. March 2024Azure: e.t.a. Aug 2024
HIPAA certification is essentially having a third party audit report for HIPAA. That is not the date that a HIPAA product offering may/will necessari...
Hi all,I've been able to create a model serving endpoint through the api, following the docs, but then when trying to update the version, I get the following error:'{"error_code":"RESOURCE_ALREADY_EXISTS","message":"Endpoint with name \'ml-project\' ...
Folks, Alternate way you can also deploy the models in serving layer with different versions. Though I am using mLflow. You may also refer to the below link if it its helpful How to Quickly Deploy, Test & Manage ML Models as REST Endpoints with Datab...
I need to run this kind of code: from pyspark.sql import SparkSession
import pandas as pd
# Create a Spark session
spark = SparkSession.builder.appName("example").getOrCreate()
# Sample data
data = [("Alice", 1), ("Bob", 2), ("Charlie", 3), ("David...
I found a good solution that works both locally and in the cloud. Copy pasting the code in case it helps someone.This is the higher level function in charge of partitioning the data and sending the data and the function fn to each node. def decrypt_d...
Hi! I was able to successfully download and run selenium on both single and no isolation mode clusters. But the shared cluster do not seem to have permission to access the path of drivers needed such as chrome and chromedriver.1) what is the general ...
I am using AutoML and want to split my data to train/validation and test using a dt column (one date for train one different date for validation and a third date for test. The problem that the autoML fails, there are only training metrics (no valiat...
While trying to create a serving endpoint with my custom model, I get a "Failed" state:Model server failed to load the model. Please see service logs for more information.The service logs show the following:Container failed with: failed to create con...
I have faced the similar issue. still didn't find the right solution. In my case, the below is the error trace i found from service logs. Not sure where the issue could be"An error occurred while loading the model. You haven't configured the CLI yet!...
Hello,I am trying to create a model serving endpoint via the API as explained here: https://docs.databricks.com/api/workspace/servingendpoints/createI created a trusted IAM role with access to DynamoDB for the feature store. I try to use this field,"...
Facing challenges in leveraging Databricks for serving, logging, and monitoring OpenAI usage within an Azure environment. Specifically, encountering issues with Inference Tables not being enabled in the UI when creating serving endpoints with externa...
Hi, In addition to the previous message, you can refer to https://learn.microsoft.com/en-us/azure/databricks/large-language-models/ai-generate-text-example.
Hello Team, I am unable to distribute the workload to databricks different worker while using the hugging face GPT2 LLM model. Jobs always use the 1 node even though we have the min and max worker node setting with 2. Appreciate if anyone can share a...
Hello, We have around between 5k to 10K transcript files available in the ADLS gen 2 and we are using hugging face gpt2 model to train and server the model and expecting to pass the workload to different cluster nodes while serving the LLM model and ...
Hi,I'm trying to enable inference table for my llama_2_7b_hf serving endpoint, however I'm getting the following error:"Inference tables are currently not available with accelerated inference." Anyone one have an idea on how to overcome this issue? C...
Thank you for posting your question in our community! We are happy to assist you.To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best answers your question?This...
Using `mlflow.pyfunc.log_model` to save a model in the model registry fails with the following error: `PicklingError: Can't pickle <built-in function input>: it's not the same object as builtins.input`I've reproduced it locally and got to the conclus...
Thank you for posting your question in our community! We are happy to assist you.To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best answers your question?This...
I want to set environment variables for serving-endpoints that are not being read as secrets by using the DB API. As an example I want to set the OPENAI_API_BASE for a LangChain model. Is this possible?
Thank you for posting your question in our community! We are happy to assist you.To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best answers your question?This...
I am trying to deploy a Vector Search Index using both the UI and the Python VectorSearchClient. In both cases the command is successful but in the Catalog Explorer the newly created index stalls with the status 'Provisioning Index' for hours. Previo...
I am trying to run an AutoML experiment but the button stays greyed out no matter what I do. I've tried different cluster configurations, different datasets, even blew away the instance in Azure and re-created it across two different Azure accounts s...