<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>article Building DBRX class Custom LLMs with Mosaic AI Training in Databricks TV</title>
    <link>https://community.databricks.com/t5/databricks-tv/building-dbrx-class-custom-llms-with-mosaic-ai-training/ba-p/72530</link>
    <description>&lt;P&gt;&lt;IFRAME src="https://www.youtube.com/embed/-5FOZHNwhaE?si=cMbSiTfjOqExEi4T" width="560" height="315" frameborder="0" allowfullscreen="" title="YouTube video player" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share" referrerpolicy="strict-origin-when-cross-origin"&gt;&lt;/IFRAME&gt;&lt;/P&gt;</description>
    <pubDate>Fri, 14 Jun 2024 17:25:16 GMT</pubDate>
    <dc:creator>lara_rachidi</dc:creator>
    <dc:date>2024-06-14T17:25:16Z</dc:date>
    <item>
      <title>Building DBRX class Custom LLMs with Mosaic AI Training</title>
      <link>https://community.databricks.com/t5/databricks-tv/building-dbrx-class-custom-llms-with-mosaic-ai-training/ba-p/72530</link>
      <description>&lt;P&gt;&lt;SPAN&gt;DBRX was trained, fine-tuned, and evaluated using Mosaic AI Training, scaling training to 3072 NVIDIA H100s and processing more than 12 trillion tokens in the process. Mosaic AI Training is available today for Databricks customers to build custom models on their own enterprise data that are tailored to a specific business context, language and domain, and can efficiently power key business use cases. We discuss a blog post that details Mosaic AI Training's core capabilities and how they were critical to the successful training of DBRX. Training LLMs and other large AI models requires the integration of numerous components. To simplify this complexity and to deliver an experience that “just works", Mosaic AI Training offers an optimized training stack that handles all aspects of large-scale distributed training. The stack supports multiple GPU cloud providers (AWS, Azure, OCI, Coreweave, to name a few), is configured with the latest GPU drivers including NVIDIA CUDA and AMD ROCm, and includes core neural network and training libraries (PyTorch, MegaBlocks, Composer, Streaming). Lastly, battle-tested scripts for training, fine-tuning, and evaluating LLMs are available in LLMFoundry, enabling customers to start training their own LLMs immediately. Link to blog: &lt;A href="https://www.databricks.com/blog/mosaic-ai-training-capabilities?utm_source=bambu&amp;amp;utm_medium=social&amp;amp;utm_campaign=advocacy" target="_blank"&gt;https://www.databricks.com/blog/mosaic-ai-training-capabilities?utm_source=bambu&amp;amp;utm_medium=social&amp;amp;utm_campaign=advocacy&lt;/A&gt;&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 14 Jun 2024 17:25:16 GMT</pubDate>
      <guid>https://community.databricks.com/t5/databricks-tv/building-dbrx-class-custom-llms-with-mosaic-ai-training/ba-p/72530</guid>
      <dc:creator>lara_rachidi</dc:creator>
      <dc:date>2024-06-14T17:25:16Z</dc:date>
    </item>
  </channel>
</rss>

