Cerebras Systems visit report

Beyond the Wafer Scale Engine, their challenge doesn't end with just making the biggest chip ever

Wisdom From The Women Leading The AI Industry, With Lakshmi Ramachandran of Cerebras Systems

Beyond the Wafer Scale Engine, their challenge doesn't end with just making the biggest chip ever

Andrew Feldman: Cerebras and AI Hardware

A conversation with Andrew Feldman, co-founder and CEO of Cerebras Systems.

New Cerebras Wafer-Scale Cluster Eliminates Months Of Painstaking Work To Build Massive Intelligence

The hottest trend in AI is the emergence of massive models such as Open AI’s GPT-3. These models are surprising even its developers with capabilities that some claim approach human sentience. Our analyst, Alberto Romero, shows the capabilities of these models in the image below on the right, which he created with a simple prompt using Midjourney, an AI similar to Open AI’s DALL·E. Large Language Models will continue to evolve to become powerful tools in businesses from Pharmaceuticals to Finance, but first they need to become easier and more cost-effective to create.

Cerebras Proposes AI Megacluster with Billions of AI Compute Cores

Chipmaker Cerebras is patching its chips – already considered the world’s largest – to create what could be the largest-ever computing cluster for AI computing.

チップが大きいことは良いことなんです – 世界最大級のAIアクセラレータ

The Cerebras Wafer-Scale Cluster delivers near-perfect linear scaling across hundreds of millions of AI-optimized compute cores while avoiding the pain of the distributed compute.

Radar Trends to Watch: September 2022

Cerebras, the company that released a gigantic (850,000 core) processor, claims their chip will democratize the hardware needed to train and run very large language models by eliminating the need to distribute computation across thousands of smaller GPUs.

So baut Cerebras seinen Wafer-großen Chip

Cerebras' Wafer Scale Engine beeindruckt nicht nur mit der Größe, sondern auch mit der Architektur. Der Chip berechnet 7,5 Petaflops.

Cerebras Systems CS-2 for Long-Read Sequences

Cerebras Systems said that its CS-2 high-performance computing accelerator is now capable of training artificial intelligence models on 20X longer sequences than "traditonal" hardware, including graphics processing units. The company expects this new capability to lead to advances in natural language processing in drug discovery, particularly by allowing researchers to examine individual genes in the context of many thousands of surrounding genes.

AI Leaders Podcast #27: Transformers, Massive Models and the future of AI

Jean-luc Chatelain, Applied Intelligence CTO, talks with Andrew Feldman, Founder and CEO of Cerebras Systems about transformers, massive models and the future of AI. They discuss how we are moving from a world of a large amount of models to fewer more powerful models known as transformers. Hear what they think this means for the future of AI.

Animal Speech, AI Pilot, Wrongful Arrest, Redesigning Streets

Our 106th episode with a summary and discussion of last week's big AI news! Apologies for a lack of consistent episode releases lately, we've been really busy...

Cerebras Wafer Scale Engine WSE-2 and CS-2 at Hot Chips 34

It’s that time of year again—Hot Chips will soon be upon us. Taking place as a virtual event on August 21–23, the conference will once again present the very latest in microprocessor architectures and system innovations.

The Future Is Now: Cindy Orozco Bohorquez Of Cerebras Systems On How Their Technological Innovation Will Shake Up The Tech Scene

In the professional side, it’s very easy to underestimate the power of communication. In my field, when you interact with executives, customers, users…you need to be able to translate and present your ideas and the key points to any audience. It’s very valuable for people — especially those of us in the technical field or mathematics — to understand and execute the power of communication.

Here Comes Hot Chips!

It’s that time of year again—Hot Chips will soon be upon us. Taking place as a virtual event on August 21–23, the conference will once again present the very latest in microprocessor architectures and system innovations.

New Uses For AI In Chips

Artificial intelligence is being deployed across a number of new applications, from improving performance and reducing power in a wide range of end devices to spotting irregularities in data movement for security reasons.

Teach Yourself About Hot Chips: 2022 Preview

We talk a lot about chips on this channel, so it stands to reason that there's a conference about the hottest chips around! Introducing Hot Chips 2022, this year's place-to-be about silicon.


 人工知能(AI)スタートアップのCerebras Systemsは米国時間8月3日、米カリフォルニア州マウンテンビューにあるコンピューター歴史博物館で行われたセレモニーで、その技術の伝統を担っていることを認められた。同博物館は、Cerebrasが開発した史上最大のコンピューターチップである同社の第2世代AIチップ、「Wafer-Scale Engine 2(WSE-2)」の展示を開始した。このチップは2021年に発表されたもので、同社の新型スーパーコンピューター「CS-2」に搭載されている。

Cerebras, le spécialiste de l’IA, remporte un succès dans le domaine des puces

Dans le secteur de la technologie, chaque invention s'appuie sur les succès et les échecs qui l'ont précédée.

Début août, la start-up d'IA Cerebras Systems a été récompensée pour avoir perpétué cette tradition lors d'une cérémonie au Computer History Museum de Mountain View, en Californie.

Last Week in AI #179: AI to save threatened species, ace university math, talk to animals, and more!

How machine learning could help save threatened species from extinction, new algorithm aces university math course questions, a 175B parameter publicly available chatbot

Größer, besser – aber auch schlauer?

Der wohl größte Computerchip der Welt könnte bald ein künstliches Gehirn simulieren, das quantitativ an das menschliche heranreicht. Warum es dennoch kaum wie ein Mensch denken wird.

A.I. Is Not Sentient. Why Do People Say It Is?

The problem is that the people closest to the technology — the people explaining it to the public — live with one foot in the future. They sometimes see what they believe will happen as much as they see what is happening now.
“There are lots of dudes in our industry who struggle to tell the difference between science fiction and real life,” said Andrew Feldman, chief executive and founder of Cerebras, a company building massive computer chips that can help accelerate the progress of A.I.

13 hot chip and semiconductor startups investors are betting on to fill supply-chain gaps and compete with giants like Intel and Nvidia

To get a better sense of why VCs are betting on chip startups, Insider compiled a list of several startups investors say are bringing new technological strides to the field. The firms range from chips designed for artificial intelligence and machine learning to semiconductors embedded in clothes. The firms on this list were picked in consult with VCs and analysts. They were chosen for their funding, leadership, and other criteria to help determine which will stick around.


The silicon chip, or integrated circuit (IC), is one of humankind’s most magnificent, complex, and transformative creations.

Computer History Museum Honors Cerebras Systems – Watch a Replay of the Event

When Cerebras Systems had its coming out at Hot Chips in August 2019, the hardware community wasn’t sure what to think. Attendees were understandably skeptical of the novel “wafer-scale” technology, not to mention an estimated power envelope of ~15 kilowatts for the chip alone. In the intervening three years, the company – under the direction of founder and CEO Andrew Feldman – has won over early critics through a series of impressive milestones.

AI startup Cerebras celebrated for chip triumph where others tried and failed

Company honored by Computer History Museum for cracking the code of making giant chips, with 'stunning' implications.

Die LRZ-Zukunft steht auf Exascale-, KI- und Quantencomputing

Am 14. Juli hat das Leibniz-Rechenzentrum (LRZ) der Bayerischen Akademie der Wissenschaften sein 60jähriges Bestehen gefeiert. Das Hochleistungsrechenzentrum, das unter anderem den Supercomputer „SuperMUC“ beherbergt ist der IT-Dienstleister der Münchner Universitäten und bayerischen Hochschulen sowie Kooperationspartner wissenschaftlicher Einrichtungen in Bayern, Deutschland und Europa. Doch Alter hat in der IT nur bedingt Wert. Wie geht es weiter?

The Microchip Era Is Giving Way to the Megachip Age

To continue making our gadgets more powerful, engineers have worked out a new way to get around the barriers to making microchips faster: Just make them bigger.

Democratizing the hardware side of large language models

There is growing interest in democratizing large language models and making them available to a broader audience. However, the hardware barriers of large language models have gone mostly unaddressed. This is one of the problems that Cerebras, a startup that specializes in AI hardware, aims to solve with its Wafer-Scale processor. In an interview with TechTalks, Cerebras CEO Andrew Feldman discussed the hardware challenges of LLMs and his company’s vision to reduce the costs and complexity of training and running large neural networks.

NETL Researchers Work to Unlock Potential of Artificial Intelligence in Climate Modeling

Researchers at the U.S. National Energy Technology Laboratory (NETL) are helping the National Center for Atmospheric Research (NCAR) unlock the potential of an advanced artificial intelligence (AI) computing resource to perform critical climate modeling that could lead to better climate change predictions.

Meet the nominees for the 2022 VentureBeat Women in AI Awards!

Two Cerebras Systems engineers are finalists for VentureBeat's Women in AI awards!

Age Checks, Theft Prevention, Minecraft AI, Autism, Responsible AI

Discussion of last week's big AI news -- including Cerebras Systems Sets Record for Largest AI Models Ever Trained on a Single Device!

Cerebras trains 20 billion parameter AI model on a single system, sets new record

US semiconductor startup Cerebras claims that it has trained the largest AI model on a single device. The company trained AI models with 20 billion parameters on its Wafer Scale Engine 2 (WSE-2) chip, the world's largest chip.

Training a 20–Billion Parameter AI Model on a Single Processor

Cerebras has shown off the capabilities of its second–generation wafer–scale engine, announcing it has set the record for the largest AI model ever trained on a single device.

For the first time, a natural language processing network with 20 billion parameters, GPT–NeoX 20B, was trained on a single device. Here’s why that matters.

Cerebras breaks record for largest AI models trained on a single device

Cerebras said it can reduce the engineering time to run large NLP models from months to minutes, making it more cost-effective and accessible.

Why The Cerebras CS-2 Machine is a Big Deal

In my first video talking about AI Hardware on this channel, I talk in detail about an exciting Cerebras CS-2 announcement!

Cerebras Systems Thinks Forward on AI Chips as it Claims Performance Win

Cerebras Systems makes the largest chip in the world, but is already thinking about its upcoming AI chips as learning models continue to grow at breakneck speed.

Cerebras Systems sets record for largest AI models ever trained on one device

Cerebras Systems said it has set the record for the largest AI models ever trained on a single device, which in this case is a giant silicon wafer with hundreds of thousands of cores.

Cerebras just built a big chip that could democratize AI

Chip startup Cerebras has developed a foot-wide piece of silicon, compared to average chips measured in millimeters, that makes training AI cheap, and easy.

#77 – VITALIY CHILEY (Cerebras)

Cerebras engineer Vitaliy Chiley joins Machine Learning Street Talk to discuss deep learning, sparsity and different compute architectures for artificial intelligence.

Cerebras HPC Acceleration ISC 2022

At ISC 2022 I had the opportunity to sit down with Andy Hock at Cerebras. This was very cool after I told Cerebras and HPE that STH would not cover the HPE Superdome plus one Cerebras CS-2 being installed at LRZ for ISC, Andy still agreed to meet. As a result, we get a little article about the experience. This was one of the more interesting conversations since we veered fairly far from the Cerebras product. Instead, what I wanted to do was just give a perspective, reinforced by this discussion, on what Cerebras is doing to differentiate itself against a huge competitor (NVIDIA.)

Eine neue Maschine für KI und HPC

Die Anforderungen an die Hardware steigen in KI und beim Hochleistungsrechnen stetig. Die Antwort darauf sind neue Ansätze bei Chip- und Rechnerdesign. Eine neuartige Kombi-Architektur wird jetzt das Leibniz-Rechenzentrum (LRZ) in Garching implementieren.

NCSA Deploys Cerebras CS-2 in New HOLL-I Supercomputer for Large-Scale AI

Cerebras Systems, the pioneer in high performance artificial intelligence (AI) computing, today announced that the National Center for Supercomputing Applications (NCSA) has deployed the Cerebras CS-2 system in their HOLL-I supercomputer.

Leading Supercomputer Sites Choose Cerebras for AI Acceleration

Cerebras Systems, the pioneer in high performance artificial intelligence (AI) computing, shared news about their many supercomputing partners including European Parallel Computing Center (EPCC), Leibniz Supercomputing Centre (LRZ), Lawrence Livermore National Laboratory, Argonne National Laboratory (ANL), the National Center for Supercomputing Applications (NCSA), and the Pittsburgh Supercomputing Center (PSC).

LRZ Adds Mega AI System as It Stacks up on Future Computing Systems

A European supercomputing hub near Munich, called the Leibniz Supercomputing Centre, is deploying Cerebras Systems’ CS-2 AI system as part of an internal initiative called Future Computing to assess alternative computing technologies to inject more speed into the region’s scientific research.

HPE, Cerebras build AI supercomputer for scientific research

Wafer madness hits the LRZ in HPE Superdome supercomputer wrapper

München verbaut riesigen KI-Chip

Als erster Standort in Europa hat das Leibniz-Rechenzentrum (LRZ) ein CS-2-System mit Cerebras' WSE-2 gekauft, welches effizient und schnell ist.

Leibniz Supercomputing Centre to deploy HPE-Cerebras supercomputer

The Leibniz Supercomputing Centre (LRZ) in Germany plans to deploy a new supercomputer featuring the HPE Superdome Flex server and the Cerebras CS-2 system.

HPE is building a rapid AI supercomputer powered by the world’s largest CPU

Hewlett Packard Enterprise (HPE) has announced it is building a powerful new AI supercomputer in collaboration with Cerebras Systems, maker of the world’s largest chip.

Bio-IT World Judges, Community Honor Six Outstanding New Products

The Cerebras CS-2 system was awarded a Best of Show award as an innovative solution to important problems facing the life sciences industry.

Argonne Talks AI Accelerators for COVID Research

As the pandemic swept across the world, virtually every research supercomputer lit up to support Covid-19 investigations. But even as the world transformed, the fairly stable status quo of simulation-based scientific computing was itself beginning to more rapidly change with the burgeoning field of AI and AI-specific accelerators, including the Cerebras CS-2.

Accelerating insights in large scale AI projects

Leading research institutes are choosing HPE Superdome Flex as the basis of new supercomputing systems designed to accelerate AI. Both the University of Edinburgh and Pittsburgh Supercomputing Center (PSC) are combining Superdome Flex with Cerebras CS-1, an AI accelerator based on the largest processor in the industry.

Cerebras WSE-2: Der Pizza-Prozessor

2,6 Billionen (!) Transistoren, 850.000 Rechenkerne, 40 GByte On-Chip-Memory, 20 kW Leistungsaufnahme: Die Wafer Scale Engine 2 von Cerebras ist in jeder Hinsicht extrem. Dabei ist der Name Programm: Der 7-nm-Chip belegt nahezu einen kompletten 300-mm-Wafer. Mittlerweile ist der Chip am LRZ in Garching bei München im Einsatz.

Cerebras, TotalEnergies Announce Stencil Algorithm Leap

Cerebras—in collaboration with French multinational TotalEnergies—has announced the development of a massively scalable stencil algorithm: a development made possible by the use of one of Cerebras’ CS-2 systems.

HPE and Cerebras to Install AI Supercomputer at Leibniz Supercomputing Centre

Cerebras' CSoft software update to version 1.2 gives developers using PyTorch and TensorFlow access to their powerful CS-2 systems through Cerebras Cloud.

The World’s Largest Chip Just Received A Major Machine Learning-Flavored Upgrade

Cerebras Systems, makers of the world’s largest chip, has announced that its CS-2 system now supports PyTorch and TensorFlow which will make it possible for researchers to quickly and easily train models with billions of parameters.

How Cerebras CS-2 update stands up to competitors’ offerings

AI hardware and software vendor Cerebras Systems released an updated version of its platform that includes integrated support for the open source TensorFlow and PyTorch machine learning frameworks.

The world’s largest chip just received a major machine learning-flavored upgrade

The Cerebras CS-2 is the world’s fastest AI system and is powered by its Wafer-Scale Engine 2 (WSE-2) CPU. With the release of version 1.2 of the Cerebras Software Platform (CSoft), the CS-2 now supports additional machine learning (opens in new tab) frameworks which will give developers even more choice when it comes to the types of models they want to run.

‘Bigger is better’ is back for hardware – without any obvious benefits

When I first saw an image of the 'wafer-scale engine' from AI hardware startup Cerebras, my mind rejected it. The company's current product is about the size of an iPad and uses 2.6 trillion transistors that contribute to over 850,000 cores.

Wafer-sized Cerebras AI chips play nicer with PyTorch, TensorFlow

Good news for those who like their AI chips big: Cerebras Systems has expanded support for the popular open-source PyTorch and TensorFlow machine-learning frameworks on the Wafer-Scale Engine 2 processors that power its CS-2 system.

How to Be a Fearless Engineer with Cerebras Systems Founder and CEO Andrew Feldman

Innovators don’t see limitations – they see challenges. And that’s exactly what happened when Andrew Feldman and his team at Cerebras Systems were told that it was impossible to build a computer chip that could deliver the same performance as hundreds of graphics processing units.

AI Chip Startups Pull In Funding as They Navigate Supply Constraints

Investors are funneling billions of dollars into startups like Cerebras Systems that make chips designed for artificial-intelligence applications, which have largely avoided the supply-chain constraints and backlogs faced by larger chip makers, startup executives, investors and industry analysts say.

nference Accelerates Self-Supervised Language Model Training With Cerebras CS-2 System

The ability to harness vast amounts of health data using advanced AI technology will lead to new discoveries and insights needed to improve patient care

Cerebras and nference Launch NLP Collaboration

High performance AI compute company Cerebras Systems and nference, an AI-driven health technology company, today announced a collaboration to accelerate natural language processing (NLP) for biomedical research and development by orders of magnitude with a Cerebras CS-2 system installed at the nference headquarters in Cambridge, Mass.

Startups bag billions to fill gaps left by chip world giants

Venture capitalists funneled billions into semiconductor startups like Cerebras Systems in 2021, we're told, targeting designers of machine-learning technologies that fulfill specific or niche needs.

Cerebras CS-2 System to Accelerate NLP for Biomedical R&D

Cerebras Systems, a pioneer in high performance artificial intelligence (AI) compute, and nference, an AI-driven health technology company, today announced a collaboration to accelerate natural language processing (NLP) for biomedical research and development by orders of magnitude with a Cerebras CS-2 system installed at the nference headquarters in Cambridge, Mass.

Most Expensive Process in the World Explained

The Cerebras CS-2 system for AI is a piece of art

Cambrian-AI 2022 Predictions: Expect More Than Just New Chips

Cerebras should be ready to focus on customer success stories in 2022, having built an accessible, if not expensive, system-level solution that has been adopted in HPC and a few enterprises.

AI in 2022: Here are More AI and Tech Predictions from IT Experts

As 2022 begins, the progress and continuing evolution of AI is in full swing across the worlds of industry, manufacturing, retail, banking and finance, healthcare and medicine and an expanding range of other fields.

10 NLP Predictions for 2022

Natural language processing (NLP) has been one of the hottest sectors in AI over the past two years. Will the string of big data breakthroughs continue into 2022? We checked in with industry experts -- including Natalia Vassilieva from Cerebras Systems -- to find out.

Memory Bottlenecks: Overcoming a Common AI Problem

Wafer-scale AI accelerator company Cerebras Systems has devised a memory bottleneck solution at the far end of the scale. At Hot Chips, the company announced MemoryX, a memory extension system for its CS-2 AI accelerator system aimed at high-performance computing and scientific workloads.


Cerebras Systems said its Wafer Scale Engine 2 chip is the “largest AI processor ever made,” consisting of 2.6 trillion transistors, 850,000 cores and 40 GB of on-chip memory. The startup said those specifications give the WSE-2 chip a massive advantage over GPU competitors.

IEEE Spectrum’s biggest semiconductor headlines of 2021

Last April, Cerebras Systems topped its original, history-making AI processor with a version built using a more advanced chipmaking technology. The result was a more than doubling of the on-chip memory to an impressive 40 gigabytes, an increase in the number of processor cores from the previous 400,000 to a speech-stopping 850,000, and a mind-boggling boost of 1.4 trillion additional transistors.

Cerebras CS-2 Aids in Fight Against SARS-CoV-2

One company contributing to the fight with artificial intelligence and machine learning is Cerebras Systems. The second iteration of the company’s Wafer-Scale Engine machine, the CS-2 system, plays an important role as part of the AI Testbed at Argonne National Laboratory and has contributed to a multi-agency COVID-19 reproduction study that was nominated as a Gordon Bell Special Prize finalist.

LLNL Establishes AI Innovation Incubator to Advance AI for Applied Science

Lawrence Livermore National Laboratory has established the AI Innovation Incubator (AI3), a collaborative hub aimed at uniting experts in artificial intelligence from LLNL, industry and academia including Cerebras Systems to advance AI for large-scale scientific and commercial applications.

EE Times Weekend: Andrew Feldman: ‘Let Your Passion Be Your Guide’

EE Times interviews Cerebras CEO Andrew Feldman on leadership, personal projects and the biggest threat to society posed by technology today

The Most Important Compute Workloads of a Generation: How Cerebras Systems is Accelerating AI’s Potential

It’s a bird, it’s a plane, it’s the largest AI processor ever made! In this week’s Fish Fry podcast, Andy Hock (Cerebras Systems) joins me to chat about the largest AI processor ever made – the 7 nm wafer scale engine 2, the details of their brain-scale AI training, and how Cerebras Systems is democratizing access to high performance AI computation.

The Most Important Compute Workloads of a Generation: How Cerebras Systems is Accelerating AI’s Potential

It’s a bird, it’s a plane, it’s the largest AI processor ever made!  In this week’s Fish Fry podcast, Andy Hock (Cerebras Systems) joins me to chat about the largest AI processor ever made –  the 7 nm wafer scale engine 2, the details of their brain-scale AI training, and how Cerebras Systems is democratizing access to high performance AI computation.

Big Data Industry Predictions for 2022

Cerebras Systems VP Andy Hock predicts that, "In 2022, AI will continue to grow as a valuable and critical workload for enterprise organizations across industries. We will see a larger number of teams investing in world-class AI computing to accelerate their research and business..."

Exascale Ambitions

Simon McIntosh-Smith discusses the role of the ExCALIBUR project in ensuring that UK research is at the forefront of HPC -- including a Cerebras CS-1

Companies Powering the New Economy

This year has been marked by continued uncertainty and upheaval of our consumer habits and supply-and-demand balance. The dynamic business landscape has been challenging for even the most successful corporations, let alone for startups that are earlier in their efforts to establish market traction and growth.

Wisdom From The Women Leading The AI Industry, With Natalia Vassilieva of Cerebras

As part of their series about the women leading the Artificial Intelligence industry, Tyler Gallagher interviewed Natalia Vassilieva, Director of Product, Machine Learning at Cerebras Systems.

How AI Is Aiming at the Bad Math of Drug Development

GlaxoSmithKline Plc hopes to double its drug success rate to 20% by using AI, teaming up with partners including DNA testing provider 23andMe Holding Co. and Cerebras Systems Inc., an upstart chipmaker that provides computer systems to crunch very large data sets.

How the Cloud Powers Moore’s Law, and More

Additional innovations such as three-dimensional chip stacking and Cerebras wafer-scale chips may help extend Moore’s law for another decade or two, at least.

Cerebras Systems & G42 join to bring high-performance AI to Middle East

Cerebras Systems and G42 signed a memorandum of understanding (MOU) at GMIS, committing to offer high-performance AI capabilities to the Middle East. G42 will enhance its technology stack with Cerebras’ CS-2 systems to give unrivaled AI compute capabilities to its partners and the broader ecosystem.

Cerebras Systems, G42 To Bring AI Compute Capabilities To The Region

Artificial intelligence (AI) compute solutions provider Cerebras Systems and G42, the UAE-based AI and cloud computing company, have signed a memorandum of understanding (MOU) at GMIS, under which they will bring high performance AI capabilities to the Middle East.


Semiconductor startups like Cerebras Systems are offering novel ways to improve the performance, efficiency, economics and bandwidth of servers and beyond with new takes on CPUs, accelerators and connectivity solutions.

World’s largest computer chip enters Mideast to push AI adoption further

DUBAI: It seems only fitting that the world’s biggest computer chip, developed by US-based Cerebras, be deployed in the Middle East, home to some of the world’s entertaining superlatives — biggest tower, deepest swimming pool, among many other “firsts.”

Cerebras Systems and G42 sign a MoU for artificial intelligence

Cerebras Systems and G42 today announced at GMIS the signing of a memorandum of understanding (MOU) under which they will bring high performance AI capabilities to the Middle East. G42, who manages the region’s largest cloud computing infrastructure, will upgrade its technology stack with Cerebras’ industry-leading CS-2 systems to deliver unparalleled AI compute capabilities to its partners and the broader ecosystem.

Cerebras Systems Named to Fast Company’s Annual List of the World’s Most Innovative Companies for 2021

Creator of the world’s fastest AI supercomputer among top-ranked companies in the Artificial Intelligence category