Hi
@Retired_mod and
@cmunteanu , I am having exactly the same problem to create the vector index and it seems that there could be a bug in the demo. What confuses me is that and even when using the Databricks UI, I can not manage to create the vector index.
Well, when running the demo, it stays for a long time repeating:
============
Waiting for index to be ready, this can take a few min... {'detailed_state': 'PROVISIONING_INITIAL_SNAPSHOT', 'message': 'Index is currently is in the process of syncing initial data. Check latest status:
https://adb-393322312342211.5.azuredatabricks.net/explore/data/dev_talk/llm_rag/databricks_documenta...', 'indexed_row_count': 0, 'provisioning_status': {'initial_pipeline_sync_progress': {'latest_version_currently_processing': 1, 'num_synced_rows': 0, 'total_rows_to_sync': 14129, 'sync_progress_completion': 0.0, 'pipeline_metrics': {'total_sync_time_per_row_ms': 0.0, 'ingestion_metrics': {'ingestion_time_per_row_ms': 0.0, 'ingestion_batch_size': 300}, 'embedding_metrics': {'embedding_generation_time_per_row_ms': 0.0, 'embedding_generation_batch_size': 0}}}}, 'ready': False, 'index_url': 'adb-393322312342211.5.azuredatabricks.net/api/2.0/vector-search/endpoints/dbdemos_vs_endpoint/indexes/dev_talk.llm_rag.databricks_documentation_vs_index'} - pipeline url:adb-393322312342211.5.azuredatabricks.net/api/2.0/vector-search/endpoints/dbdemos_vs_endpoint/indexes/dev_talk.llm_rag.databricks_documentation_vs_index
Then after a long time the Cell stops with the following error message:
It seems that the url is wrong (this is the bug I was referring), it has the endpoint and the vector index path interchanged. It should be:
Just like in the output of the Cell that is showing above. There, the URL is showed correctly,
================
@Retired_mod If any specific configuration is required regarding the embedding model, it would be good to have it specified. In your reply you said:
- When creating the Vector Search Index, ensure that you specify the correct parameters:
- embedding_source_column: This should match the column name containing your text data (e.g., ‘content’).
- embedding_model_endpoint_name: Use ‘bge-large-en’ as you’ve set up this model as a serving endpoint.
All these specifications are correctly configured in the demo notebook. So, I am confused on what is left for us to configure.
@cmunteanu I have followed your suggestion of using a self managed embedding to create the vector index. It does work, in the sense that I created the vector index. But, I can not use (easily) the nice features of Databricks vector_search client that converts internally text to vectors and vice-versa. Which make things easier for the RAG - chatbot. Have you got around that?