November 10, 2023
Cerebras Software Release 2.0: 50% Faster Training, PyTorch 2.0 Support, Diffusion Transformers, and More
Today we are excited to announce Cerebras software release…
October 12, 2023
How we fine-tuned Llama2-70B to pass the US Medical License Exam in a week
New open-access model by M42 outperforms GPT-3.5 in…
September 5, 2023
Jais: a New Pinnacle in Open Arabic NLP
Introducing a new state-of-the-art bi-lingual…
July 24, 2023
BTLM-3B-8K: 7B Performance in a 3 Billion Parameter Model
Cerebras and Opentensor introduce a new standard for…
July 22, 2023
Accelerating Large Language Model Training with Variable Sparse Pre-training and Dense Fine-tuning
We reduced pre-training FLOPs by 64% using sparsity. To the…
July 22, 2023
Variable Sequence Length Training for Long-Context Large Language Models
We show it is possible to accelerate the training for large…
July 20, 2023
Introducing Condor Galaxy 1: a 4 exaFLOPS Supercomputer for Generative AI
Cerebras, in partnership with G42 unveils CG-1, a 4…
June 9, 2023
SlimPajama: A 627B token, cleaned and deduplicated version of RedPajama
Today we are releasing SlimPajama – the largest…
May 23, 2023
Efficient Large-Scale GPT Training Using a Cerebras Wafer-Scale Cluster
Cerebras has built a platform for push-button training of…
April 17, 2023
Fine-Tuning with Cerebras AI Model Studio Launchpad
Cerebras shares research showing smaller foundation models…