<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Vector search index creation is incredibly slow in Generative AI</title>
    <link>https://community.databricks.com/t5/generative-ai/vector-search-index-creation-is-incredibly-slow/m-p/111831#M784</link>
    <description>&lt;P&gt;&lt;SPAN&gt;I am trying to create a vector search index for a Delta Table using Azure OpenAI embeddings (text-embedding-3-large). The table contains 5000 chunks with approx. 1000 tokens each. The OpenAI embeddings are generated through a Databricks model serving endpoint which forwards the embedding requests to our Azure deployment.&lt;/SPAN&gt;&lt;BR /&gt;&lt;BR /&gt;&lt;SPAN&gt;The latency of the index creation is incredibly high. To embed just 5000 chunks the initial sync takes over an hour.&amp;nbsp;&lt;/SPAN&gt;&lt;SPAN&gt;If I extrapolate this to 5 Mio. chunks, embedding them would take over a month.&lt;BR /&gt;&lt;BR /&gt;The deployment in Azure can process 350K tokens/min and should not be the limiting factor.&lt;/SPAN&gt;&lt;BR /&gt;&lt;BR /&gt;&lt;SPAN&gt;I am currently creating the index like this:&lt;BR /&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;DIV&gt;&lt;PRE&gt;&lt;FONT face="courier new,courier"&gt;&lt;SPAN&gt;vsc.&lt;/SPAN&gt;&lt;SPAN&gt;create_delta_sync_index&lt;/SPAN&gt;&lt;SPAN&gt;(&lt;/SPAN&gt;&lt;/FONT&gt;&lt;BR /&gt;&lt;FONT face="courier new,courier"&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; &lt;/SPAN&gt;&lt;SPAN&gt;endpoint_name&lt;/SPAN&gt;&lt;SPAN&gt;=&lt;/SPAN&gt;&lt;SPAN&gt;VECTOR_SEARCH_ENDPOINT_NAME,&lt;/SPAN&gt;&lt;/FONT&gt;&lt;BR /&gt;&lt;FONT face="courier new,courier"&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; &lt;/SPAN&gt;&lt;SPAN&gt;index_name&lt;/SPAN&gt;&lt;SPAN&gt;=&lt;/SPAN&gt;&lt;SPAN&gt;vs_index_fullname,&lt;/SPAN&gt;&lt;/FONT&gt;&lt;BR /&gt;&lt;FONT face="courier new,courier"&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; &lt;/SPAN&gt;&lt;SPAN&gt;source_table_name&lt;/SPAN&gt;&lt;SPAN&gt;=&lt;/SPAN&gt;&lt;SPAN&gt;source_table_fullname,&lt;/SPAN&gt;&lt;/FONT&gt;&lt;BR /&gt;&lt;FONT face="courier new,courier"&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; &lt;/SPAN&gt;&lt;SPAN&gt;pipeline_type&lt;/SPAN&gt;&lt;SPAN&gt;=&lt;/SPAN&gt;&lt;SPAN&gt;"TRIGGERED"&lt;/SPAN&gt;&lt;SPAN&gt;,&lt;/SPAN&gt;&lt;/FONT&gt;&lt;BR /&gt;&lt;FONT face="courier new,courier"&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; &lt;/SPAN&gt;&lt;SPAN&gt;primary_key&lt;/SPAN&gt;&lt;SPAN&gt;=&lt;/SPAN&gt;&lt;SPAN&gt;"id"&lt;/SPAN&gt;&lt;SPAN&gt;,&lt;/SPAN&gt;&lt;/FONT&gt;&lt;BR /&gt;&lt;FONT face="courier new,courier"&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; &lt;/SPAN&gt;&lt;SPAN&gt;embedding_source_column&lt;/SPAN&gt;&lt;SPAN&gt;=&lt;/SPAN&gt;&lt;SPAN&gt;'content'&lt;/SPAN&gt;&lt;/FONT&gt;&lt;BR /&gt;&lt;FONT face="courier new,courier"&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; &lt;/SPAN&gt;&lt;SPAN&gt;embedding_model_endpoint_name&lt;/SPAN&gt;&lt;SPAN&gt;=&lt;/SPAN&gt;&lt;SPAN&gt;'text-embedding-3-large'&lt;/SPAN&gt;&lt;/FONT&gt;&lt;BR /&gt;&lt;FONT face="courier new,courier"&gt;&lt;SPAN&gt;&amp;nbsp; )&lt;/SPAN&gt;&lt;/FONT&gt;&lt;/PRE&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;My assumption for the slow speed is that the index calculates the embeddings row by row, without async Azure API calls, and only starts with the next row once the previous embedding is created.&lt;BR /&gt;&lt;BR /&gt;Which options do I have for speeding up the index creation?&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;/DIV&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Wed, 05 Mar 2025 15:15:09 GMT</pubDate>
    <dc:creator>epistoteles</dc:creator>
    <dc:date>2025-03-05T15:15:09Z</dc:date>
    <item>
      <title>Vector search index creation is incredibly slow</title>
      <link>https://community.databricks.com/t5/generative-ai/vector-search-index-creation-is-incredibly-slow/m-p/111831#M784</link>
      <description>&lt;P&gt;&lt;SPAN&gt;I am trying to create a vector search index for a Delta Table using Azure OpenAI embeddings (text-embedding-3-large). The table contains 5000 chunks with approx. 1000 tokens each. The OpenAI embeddings are generated through a Databricks model serving endpoint which forwards the embedding requests to our Azure deployment.&lt;/SPAN&gt;&lt;BR /&gt;&lt;BR /&gt;&lt;SPAN&gt;The latency of the index creation is incredibly high. To embed just 5000 chunks the initial sync takes over an hour.&amp;nbsp;&lt;/SPAN&gt;&lt;SPAN&gt;If I extrapolate this to 5 Mio. chunks, embedding them would take over a month.&lt;BR /&gt;&lt;BR /&gt;The deployment in Azure can process 350K tokens/min and should not be the limiting factor.&lt;/SPAN&gt;&lt;BR /&gt;&lt;BR /&gt;&lt;SPAN&gt;I am currently creating the index like this:&lt;BR /&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;DIV&gt;&lt;PRE&gt;&lt;FONT face="courier new,courier"&gt;&lt;SPAN&gt;vsc.&lt;/SPAN&gt;&lt;SPAN&gt;create_delta_sync_index&lt;/SPAN&gt;&lt;SPAN&gt;(&lt;/SPAN&gt;&lt;/FONT&gt;&lt;BR /&gt;&lt;FONT face="courier new,courier"&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; &lt;/SPAN&gt;&lt;SPAN&gt;endpoint_name&lt;/SPAN&gt;&lt;SPAN&gt;=&lt;/SPAN&gt;&lt;SPAN&gt;VECTOR_SEARCH_ENDPOINT_NAME,&lt;/SPAN&gt;&lt;/FONT&gt;&lt;BR /&gt;&lt;FONT face="courier new,courier"&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; &lt;/SPAN&gt;&lt;SPAN&gt;index_name&lt;/SPAN&gt;&lt;SPAN&gt;=&lt;/SPAN&gt;&lt;SPAN&gt;vs_index_fullname,&lt;/SPAN&gt;&lt;/FONT&gt;&lt;BR /&gt;&lt;FONT face="courier new,courier"&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; &lt;/SPAN&gt;&lt;SPAN&gt;source_table_name&lt;/SPAN&gt;&lt;SPAN&gt;=&lt;/SPAN&gt;&lt;SPAN&gt;source_table_fullname,&lt;/SPAN&gt;&lt;/FONT&gt;&lt;BR /&gt;&lt;FONT face="courier new,courier"&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; &lt;/SPAN&gt;&lt;SPAN&gt;pipeline_type&lt;/SPAN&gt;&lt;SPAN&gt;=&lt;/SPAN&gt;&lt;SPAN&gt;"TRIGGERED"&lt;/SPAN&gt;&lt;SPAN&gt;,&lt;/SPAN&gt;&lt;/FONT&gt;&lt;BR /&gt;&lt;FONT face="courier new,courier"&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; &lt;/SPAN&gt;&lt;SPAN&gt;primary_key&lt;/SPAN&gt;&lt;SPAN&gt;=&lt;/SPAN&gt;&lt;SPAN&gt;"id"&lt;/SPAN&gt;&lt;SPAN&gt;,&lt;/SPAN&gt;&lt;/FONT&gt;&lt;BR /&gt;&lt;FONT face="courier new,courier"&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; &lt;/SPAN&gt;&lt;SPAN&gt;embedding_source_column&lt;/SPAN&gt;&lt;SPAN&gt;=&lt;/SPAN&gt;&lt;SPAN&gt;'content'&lt;/SPAN&gt;&lt;/FONT&gt;&lt;BR /&gt;&lt;FONT face="courier new,courier"&gt;&lt;SPAN&gt;&amp;nbsp; &amp;nbsp; &lt;/SPAN&gt;&lt;SPAN&gt;embedding_model_endpoint_name&lt;/SPAN&gt;&lt;SPAN&gt;=&lt;/SPAN&gt;&lt;SPAN&gt;'text-embedding-3-large'&lt;/SPAN&gt;&lt;/FONT&gt;&lt;BR /&gt;&lt;FONT face="courier new,courier"&gt;&lt;SPAN&gt;&amp;nbsp; )&lt;/SPAN&gt;&lt;/FONT&gt;&lt;/PRE&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;My assumption for the slow speed is that the index calculates the embeddings row by row, without async Azure API calls, and only starts with the next row once the previous embedding is created.&lt;BR /&gt;&lt;BR /&gt;Which options do I have for speeding up the index creation?&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;/DIV&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 05 Mar 2025 15:15:09 GMT</pubDate>
      <guid>https://community.databricks.com/t5/generative-ai/vector-search-index-creation-is-incredibly-slow/m-p/111831#M784</guid>
      <dc:creator>epistoteles</dc:creator>
      <dc:date>2025-03-05T15:15:09Z</dc:date>
    </item>
    <item>
      <title>Re: Vector search index creation is incredibly slow</title>
      <link>https://community.databricks.com/t5/generative-ai/vector-search-index-creation-is-incredibly-slow/m-p/121290#M946</link>
      <description>&lt;P&gt;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/152101"&gt;@epistoteles&lt;/a&gt;&amp;nbsp;I am also facing the same issue while delta table sync with Index. Could you please share if you have any work around or used some different appoarch to decrease latency.&lt;/P&gt;&lt;P&gt;Thanks in advance.&lt;/P&gt;</description>
      <pubDate>Tue, 10 Jun 2025 04:20:59 GMT</pubDate>
      <guid>https://community.databricks.com/t5/generative-ai/vector-search-index-creation-is-incredibly-slow/m-p/121290#M946</guid>
      <dc:creator>amitkumarvish</dc:creator>
      <dc:date>2025-06-10T04:20:59Z</dc:date>
    </item>
    <item>
      <title>Re: Vector search index creation is incredibly slow</title>
      <link>https://community.databricks.com/t5/generative-ai/vector-search-index-creation-is-incredibly-slow/m-p/121299#M947</link>
      <description>&lt;P&gt;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/167812"&gt;@amitkumarvish&lt;/a&gt;&amp;nbsp;My only solution so far has been to increase embedding model capacity on Azure ... still comparably slow.&lt;/P&gt;</description>
      <pubDate>Tue, 10 Jun 2025 07:45:57 GMT</pubDate>
      <guid>https://community.databricks.com/t5/generative-ai/vector-search-index-creation-is-incredibly-slow/m-p/121299#M947</guid>
      <dc:creator>epistoteles</dc:creator>
      <dc:date>2025-06-10T07:45:57Z</dc:date>
    </item>
    <item>
      <title>Re: Vector search index creation is incredibly slow</title>
      <link>https://community.databricks.com/t5/generative-ai/vector-search-index-creation-is-incredibly-slow/m-p/139674#M1426</link>
      <description>&lt;P&gt;Is following an option for you?&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;A href="https://www.databricks.com/blog/announcing-storage-optimized-endpoints-vector-search" target="_blank"&gt;https://www.databricks.com/blog/announcing-storage-optimized-endpoints-vector-search&lt;/A&gt;&lt;/P&gt;
&lt;P&gt;Also, good to have your SA address this further as underlying issue may be somewhere else - model serving performance for instance.&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 19 Nov 2025 12:57:42 GMT</pubDate>
      <guid>https://community.databricks.com/t5/generative-ai/vector-search-index-creation-is-incredibly-slow/m-p/139674#M1426</guid>
      <dc:creator>AbhaySingh</dc:creator>
      <dc:date>2025-11-19T12:57:42Z</dc:date>
    </item>
  </channel>
</rss>

