<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Options sporadic (and cost-efficient) Model Serving on Databricks? in Machine Learning</title>
    <link>https://community.databricks.com/t5/machine-learning/options-sporadic-and-cost-efficient-model-serving-on-databricks/m-p/137756#M4408</link>
    <description>&lt;P&gt;&lt;SPAN&gt;Hi all,&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;I'm new to Databricks so would appreciate some advice.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;I have a ML model deployed using Databricks Model Serving. My use case is very sporadic: I only need to make 5–15 prediction requests per day (industrial application), and there can be long idle periods between requests. I’ve noticed that after a cold start, the serving cluster stays up for at least 30 minutes (the minimum idle timeout), and I am billed for this entire period, even if no further requests are made.&lt;/P&gt;&lt;P&gt;Is there any way to serve models on Databricks where I only pay for actual requests (compute time), and not for idle time? Or are there recommended alternatives, perhaps via integration with other Azure services?&lt;/P&gt;&lt;P&gt;Thanks for any advice!&lt;/P&gt;</description>
    <pubDate>Wed, 05 Nov 2025 12:47:18 GMT</pubDate>
    <dc:creator>cbossi</dc:creator>
    <dc:date>2025-11-05T12:47:18Z</dc:date>
    <item>
      <title>Options sporadic (and cost-efficient) Model Serving on Databricks?</title>
      <link>https://community.databricks.com/t5/machine-learning/options-sporadic-and-cost-efficient-model-serving-on-databricks/m-p/137756#M4408</link>
      <description>&lt;P&gt;&lt;SPAN&gt;Hi all,&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;I'm new to Databricks so would appreciate some advice.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;I have a ML model deployed using Databricks Model Serving. My use case is very sporadic: I only need to make 5–15 prediction requests per day (industrial application), and there can be long idle periods between requests. I’ve noticed that after a cold start, the serving cluster stays up for at least 30 minutes (the minimum idle timeout), and I am billed for this entire period, even if no further requests are made.&lt;/P&gt;&lt;P&gt;Is there any way to serve models on Databricks where I only pay for actual requests (compute time), and not for idle time? Or are there recommended alternatives, perhaps via integration with other Azure services?&lt;/P&gt;&lt;P&gt;Thanks for any advice!&lt;/P&gt;</description>
      <pubDate>Wed, 05 Nov 2025 12:47:18 GMT</pubDate>
      <guid>https://community.databricks.com/t5/machine-learning/options-sporadic-and-cost-efficient-model-serving-on-databricks/m-p/137756#M4408</guid>
      <dc:creator>cbossi</dc:creator>
      <dc:date>2025-11-05T12:47:18Z</dc:date>
    </item>
    <item>
      <title>Re: Options sporadic (and cost-efficient) Model Serving on Databricks?</title>
      <link>https://community.databricks.com/t5/machine-learning/options-sporadic-and-cost-efficient-model-serving-on-databricks/m-p/137777#M4409</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/195220"&gt;@cbossi&lt;/a&gt;&amp;nbsp;, You are right!&lt;/P&gt;
&lt;P&gt;A 30-minute idle period precedes the endpoint's scaling down. You are billed for the compute resources used during this period, plus the actual serving time when requests are made. This is the current expected behaviour.&amp;nbsp;You cannot currently reduce the idle timeout to less than 30 minutes.&lt;/P&gt;
&lt;P&gt;If your use case does not require real-time request prediction, it is better to use a batch prediction by accumulating requests throughout the day and running them all at once. Alternatively, you can explore Azure Functions to host the model.&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 05 Nov 2025 13:29:26 GMT</pubDate>
      <guid>https://community.databricks.com/t5/machine-learning/options-sporadic-and-cost-efficient-model-serving-on-databricks/m-p/137777#M4409</guid>
      <dc:creator>KaushalVachhani</dc:creator>
      <dc:date>2025-11-05T13:29:26Z</dc:date>
    </item>
  </channel>
</rss>

