I'm currently experimenting with vector search using Databricks. Everything runs smoothly when I load the model deployed in Unity Catalog into a notebook session and ask questions using Python. However, when I attempt to serve it, I encounter a gener...
i'm using openai text-embedding-3-large as embedding model, dbrx as chat model, and databricks as vectorstore, everything deployed and working fine in the workspace. But for some reason, error trying to serve the model, my unity catalog is aready in ...
Hi @JiviteshM, If you’re interested in breakout sessions related to Revenue Data Analytics and Data Science, here are some recommendations:
Building a data lakehouse for Data Science at DoorDash:
Hien Luu, Senior Engineering Manager at DoorDash, ...
It’s a wonderful summit! I met so many women in tech during this event, it’s a very good opportunity to exchange thought and ideas with such a divers ppl to modernize Data & AI! I actually join here from Japan, and unfortunately not so many women in ...
Hi @Sakura, We're thrilled to hear that you had a wonderful experience at DAIS 2023 and that you were able to connect with so many inspiring women in tech! It's events like these that truly highlight the importance of diversity and collaboration in t...
Thank you so much for organizing this big event. The main purpose of attending this event is to get to know about how the databricks is solving the problems for thousands of customers. The training session I attend for the databricks was amazing and ...
Hi @milinshah, We're thrilled to hear that you had a great experience at DAIS 2023 and found the training session valuable! Your enthusiasm for learning about how Databricks solves problems for customers is exactly what drives us.
We wanted to share ...
Excited to know the dramatic innovations!- Delta Lake 3.0 enabling all the format to one ecosystem- Collaboration with Habu in Clean Room- LakehouseIQ: truly we are at the dawn of democratization of Data and AI!!#UniForm
Hi @Masa, Thank you for sharing your experiences and insights from DAIS 2023!
It's great to hear about the exciting innovations and collaborations that were highlighted during the event, such as Delta Lake 3.0 and the collaboration with Habu in Clean...
Greetings everyone,We are trying to implement a series of visualizations. All these visualizations have queries assigned to them in the form of “ Select * from test table where timestamp between :Days start and :Days end”. The is also a filter applie...
Hi @EDDatabricks, Thank you for sharing your requirements and observations regarding the Lakeview dashboards. Let’s address each of your points:
Dynamic Filters Based on Current Timestamp: To achieve dynamic filtering based on the current timesta...
Can anyone point me towards some resources for achieving this? I already have the token.Trying with: dbsql.WithAccessToken(settings.Token)But I'm getting the following error:Unable to load OAuth Config: request error after 1 attempt(s): unexpected HT...
Hi @Miguel_Grafana, It seems you’re encountering an error related to OAuth configuration when using the dbsql.WithAccessToken method with your token.
Let’s troubleshoot this issue.
Here are a couple of things you can check:
Grant Type Parameter:
...
Hi everyone,I created a service principals called TestServicePrincipal. I tried to grant the catalog access to the service principals, but the error mentioned that it could not find principal with name TestServicePrincipal. If I grant the access to s...
Hi @Kasen, It seems you’re encountering an issue where your newly created service principal isn’t immediately discoverable.
Let’s explore some possible reasons and solutions:
Delayed Propagation:
Sometimes, there can be a delay in the propagation...
Hello everyone,I'm trying to register a model with MLflow in Databricks, but encountering an error with the following command: model_version = mlflow.register_model(f"runs:/{run_id}/random_forest_model", model_name) The error message is as follows:...
Hi @otara_geni, The ConnectTimeoutError you’re encountering when registering a model with MLflow in Databricks is related to a timeout issue while connecting to the specified endpoint URL.
The error message indicates that there’s a timeout when co...
I have a model created with catboost and exported in onnx format in workspace and I want to download that model to my local machine.I tried to use the Export that is in the three points to the right of the model, but the model weighs more than 10 Mb ...
Hey Databricks, Why did you remove the ephemeral notebook links and job Ids from the parallel runs? This has created a huge gap for us. We can no longer view the ephemeral notebooks, and also the Jobids are missing from the output. Waccha doing?...
Hi Kaniz, It's funny you mention these things - we are doing some of those - the problem now is that the JobId is obscured from the output meaning we can't tell which ephemeral notebook goes with which JobId. It looks like the ephemeral notebook ...
Hi folks, I'm working on a project with Databricks using Unity Catalog and a connection to SSIS (SQL Server Integration Services).My team is trying to access data registered in Unity Catalog using Simba ODBC driver version 2.8.0.1002. They mentioned ...
Hi @FelipeRegis, It seems you’re encountering issues with accessing data registered in Unity Catalog using the Simba ODBC driver.
Let’s explore some possible solutions:
Delta Lake Native Connector:
Consider using Delta Lake’s native Delta JDBC/OD...
Hey Databricks, Why did you take away the jobids from the parallel runs? We use those to identify which output goes with which run. Please put them back.Benedetta
Hi @Benedetta,
Thank you for reaching out. I understand your concern regarding the jobids in parallel runs. I will look into this matter and get back to you with more information as soon as possible.
Hi I'm creating a DLT pipeline which uses DLT CDC to implement SCD Type 1 to take the latest record using a datetime column which works with no issues:@dlt.view
def users():
return spark.readStream.table("source_table")
dlt.create_streaming_table(...
Hi @dm7, Thank you for providing the details of your DLT pipeline and the desired outcome!
It looks like you’re trying to implement a Slowly Changing Dimension (SCD) Type 2 behaviour where you want to capture historical changes over time.
Let’s br...
Hi,I'm trying to create a customer docker image with some R packages re-installed. However, when I try to use it in a notebook, it can't seem to find the installed packages. The build runs fine.FROM databricksruntime/rbase:14.3-LTS## update system li...
Hi @BenCCC,
Here are a few things you can check:
Package Installation in Dockerfile:
In your Dockerfile, you’re using the RUN R -e 'install.packages(...)' command to install R packages. While this approach works, there are alternative methods th...