cancel
Showing results forย 
Search instead forย 
Did you mean:ย 
Announcements
cancel
Showing results forย 
Search instead forย 
Did you mean:ย 

Building DBRX-class Custom LLMs with Mosaic AI Training

Sujitha
Community Manager
Community Manager

We recently introduced DBRX: an open, state-of-the-art, general-purpose LLM. DBRX was trained, fine-tuned, and evaluated using Mosaic AI Training, scaling training to 3072 NVIDIA H100s and processing more than 12 trillion tokens in the process.

Training LLMs, and in particular MoE models such as DBRX, is hard. It requires overcoming many infrastructure, performance, and scientific challenges. Mosaic AI Training was purposely built to address these challenges and was battle-tested through the training of DBRX, the MPT series of models, and many other LLMs such as Olaโ€™s Krutrim, AI2โ€™s OLMo, Dynamo AIโ€™s Dynamo 8B, Refuelโ€™s LLM-2, and others. 

Continue to read more here.

0 REPLIES 0