<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Endpoint deployment is very slow in Machine Learning</title>
    <link>https://community.databricks.com/t5/machine-learning/endpoint-deployment-is-very-slow/m-p/132610#M4321</link>
    <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/154968"&gt;@gbhatia&lt;/a&gt;,&lt;BR /&gt;Let us know if the solution worked for you.&lt;/P&gt;</description>
    <pubDate>Sat, 20 Sep 2025 16:28:09 GMT</pubDate>
    <dc:creator>WiliamRosa</dc:creator>
    <dc:date>2025-09-20T16:28:09Z</dc:date>
    <item>
      <title>Endpoint deployment is very slow</title>
      <link>https://community.databricks.com/t5/machine-learning/endpoint-deployment-is-very-slow/m-p/132030#M4315</link>
      <description>&lt;P&gt;HI team&amp;nbsp;&lt;/P&gt;&lt;P&gt;I am testing some changes on UAT / DEV environment and noticed that the model endpoint are very slow to deploy. Since the environment is just testing and not serving any production traffic, I was wondering if there was a way to expedite this process? I dont need the most stable / secure roll-over in this scenario.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;&lt;P&gt;Gurpreet&lt;/P&gt;</description>
      <pubDate>Mon, 15 Sep 2025 17:41:00 GMT</pubDate>
      <guid>https://community.databricks.com/t5/machine-learning/endpoint-deployment-is-very-slow/m-p/132030#M4315</guid>
      <dc:creator>gbhatia</dc:creator>
      <dc:date>2025-09-15T17:41:00Z</dc:date>
    </item>
    <item>
      <title>Re: Endpoint deployment is very slow</title>
      <link>https://community.databricks.com/t5/machine-learning/endpoint-deployment-is-very-slow/m-p/132056#M4316</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/154968"&gt;@gbhatia&lt;/a&gt;,&lt;BR /&gt;&lt;BR /&gt;I’d need a few more details to fully understand your deployment, but in general, what can help is setting Compute type: CPU (cheaper and sufficient for testing), Compute scale-out: Small (0–4 concurrency, 0–4 DBU) since you don’t need high concurrency in DEV/UAT, and keeping Scale to zero disabled to avoid cold starts and have the endpoint always ready — noting that this increases costs slightly but makes testing much faster; for production, the recommended practice is to use larger instance sizes, more replicas, and only enable scale to zero for truly intermittent workloads.&lt;BR /&gt;&lt;A href="https://docs.databricks.com/aws/en/machine-learning/model-serving/create-manage-serving-endpoints" target="_blank"&gt;https://docs.databricks.com/aws/en/machine-learning/model-serving/create-manage-serving-endpoints&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="WiliamRosa_0-1757995272360.png" style="width: 400px;"&gt;&lt;img src="https://community.databricks.com/t5/image/serverpage/image-id/19997iA87563F70F0EF1ED/image-size/medium?v=v2&amp;amp;px=400" role="button" title="WiliamRosa_0-1757995272360.png" alt="WiliamRosa_0-1757995272360.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 16 Sep 2025 04:01:46 GMT</pubDate>
      <guid>https://community.databricks.com/t5/machine-learning/endpoint-deployment-is-very-slow/m-p/132056#M4316</guid>
      <dc:creator>WiliamRosa</dc:creator>
      <dc:date>2025-09-16T04:01:46Z</dc:date>
    </item>
    <item>
      <title>Re: Endpoint deployment is very slow</title>
      <link>https://community.databricks.com/t5/machine-learning/endpoint-deployment-is-very-slow/m-p/132610#M4321</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/154968"&gt;@gbhatia&lt;/a&gt;,&lt;BR /&gt;Let us know if the solution worked for you.&lt;/P&gt;</description>
      <pubDate>Sat, 20 Sep 2025 16:28:09 GMT</pubDate>
      <guid>https://community.databricks.com/t5/machine-learning/endpoint-deployment-is-very-slow/m-p/132610#M4321</guid>
      <dc:creator>WiliamRosa</dc:creator>
      <dc:date>2025-09-20T16:28:09Z</dc:date>
    </item>
    <item>
      <title>Re: Endpoint deployment is very slow</title>
      <link>https://community.databricks.com/t5/machine-learning/endpoint-deployment-is-very-slow/m-p/132702#M4325</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/179612"&gt;@WiliamRosa&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks for your response on this. I have been using the setting you described aboved, with the exception of `scale_to_zero`. PFA screenshot of the endpoint settings.&amp;nbsp;&lt;BR /&gt;My deployment is a simple Pytorch Deep Learning model wrapped in a `sklearn` Pipeline wrapper. So, something like this:&lt;BR /&gt;&amp;nbsp;```&lt;BR /&gt;Class OnlineModel(&lt;SPAN&gt;mlflow&lt;/SPAN&gt;&lt;SPAN&gt;.&lt;/SPAN&gt;&lt;SPAN&gt;pyfunc&lt;/SPAN&gt;&lt;SPAN&gt;.&lt;/SPAN&gt;&lt;SPAN&gt;PythonModel):&lt;BR /&gt;def __init__(self, deep_learning_model):&lt;BR /&gt;self.model = &lt;SPAN&gt;deep_learning_model&lt;/SPAN&gt; # pytorch model&lt;BR /&gt;```&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="Screenshot 2025-09-22 at 10.50.52 AM.png" style="width: 400px;"&gt;&lt;img src="https://community.databricks.com/t5/image/serverpage/image-id/20142i5996BDC2DF9053D9/image-size/medium?v=v2&amp;amp;px=400" role="button" title="Screenshot 2025-09-22 at 10.50.52 AM.png" alt="Screenshot 2025-09-22 at 10.50.52 AM.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt; &lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;Every small changes to the model takes 20-30 mins of updating the endpoint which causes a significant delay in my testing &amp;amp; development. Wondering if I can decrease this wait time or this is how much databricks endpoint update take and not much i can do with it.&amp;nbsp;&lt;/P&gt;&lt;P&gt;Gurpreet&lt;/P&gt;</description>
      <pubDate>Mon, 22 Sep 2025 14:54:13 GMT</pubDate>
      <guid>https://community.databricks.com/t5/machine-learning/endpoint-deployment-is-very-slow/m-p/132702#M4325</guid>
      <dc:creator>gbhatia</dc:creator>
      <dc:date>2025-09-22T14:54:13Z</dc:date>
    </item>
  </channel>
</rss>

