$279.72 -31.35 (-10.08%) Closing price 05/15/2026 04:00 PM Eastern
Extended Trading
$278.75 -0.97 (-0.35%) As of 05/15/2026 07:59 PM Eastern
Extended trading is trading that happens on electronic markets outside of regular trading hours. This is a fair market value extended hours price provided by Massive. Learn more. We are building the fastest AI infrastructure in the world. In AI, speed is critical to win. Speed improves user engagement, expands product capabilities, can lower operating costs, and opens new markets. It shortens iteration cycles for engineers, researchers, and professionals across industries, allowing them to be more productive. Speed unlocks new applications and new industries. In technology, “speed unlocking value” is a pattern that has repeated itself over the past 30 years. Faster solutions are used more often and for more demanding tasks. For example, the speed of broadband transformed the internet from static pages into real-time applications, enabling new products and industries. Similarly, in search, Google showed that even short delays in delivering answers significantly reduced usage and engagement. AI repeats this pattern. As AI has moved from novelty to necessity, AI work has grown more demanding, and speed has become a bottleneck. Faster AI does more work in less time, providing better answers sooner. Our solutions are built for speed. Cerebras Inference delivers answers up to 15 times faster than leading GPU-based solutions as benchmarked on leading open-source models. Similarly, many customers have achieved more than 10 times faster training time-to-solution compared to leading GPU systems of the same generation. These performance breakthroughs are the result of our core innovation: the world’s first and only commercialized wafer-scale processor. Called the Wafer-Scale Engine (“WSE”), our processor is 58 times larger than NVIDIA’s B200 chip and has 2,625 times more memory bandwidth than NVIDIA’s B200 package, which contains two individual chips. To build the WSE, we solved the 75-year-old compute industry problem of wafer-scale integration to produce, yield, power, and cool a chip of this size. This size is what enables our incredible AI speeds. By bringing massive compute and memory onto a single piece of silicon and integrating it into a purpose-built system and software stack, we deliver exceptional AI speed for customers on premises and via the cloud. Our strategic partners and customers include hyperscalers, foundation model labs, AI-native and digital-native businesses, enterprises, and Sovereign AI initiatives. OpenAI, the world’s leading foundation model lab, selected us to be its fast inference solution. With Cerebras, OpenAI’s Codex-Spark users turn ideas into working software in seconds. Amazon Web Services (“AWS”), the world’s leading hyperscale cloud, has signed a binding term sheet with us to become the first hyperscaler to deploy Cerebras in its own data centers, providing massive distribution to a broad base of enterprise customers. Our customers use Cerebras solutions to run applications that demand speed, scale, and intelligence. This work includes training and serving large frontier models with near-instant responses, processing massive datasets in real time, and generating full-stack applications in a single step. Once customers adopt fast inference, user expectations for interactivity rise, and engineering teams shift from latency optimizations to other work, making it difficult to return to slower inference. We deliver our solutions to customers in several different ways. Organizations that require full data and infrastructure control can purchase Cerebras AI supercomputers for on-premises deployments. Customers seeking cloud flexibility can access Cerebras compute through consumption-based models on Cerebras Cloud or through partner clouds. For example, our high-speed inference services are available through partners, including AWS Marketplace, Microsoft Marketplace, IBM watsonx Model Gateway, Vercel AI Gateway, OpenRouter, and Hugging Face, enabling seamless adoption within existing workflows. Our ability to deliver differentiated performance has made us a strategic partner to many of our largest customers. Beyond providing compute infrastructure, we provide AI services to our customers to co-develop solutions to address their most complex challenges, from training state-of-the-art models to optimizing deployments for each application’s needs. These partnerships have expanded over time; notably, our top ten customers by year-to-date revenue through December 31, 2025 increased their aggregate spend with us by approximately 80% within 12 months of their initial purchase, often including contracts for co-development. AI is one of the fastest growing technologies in history. We believe that our high-speed AI solutions give us a meaningful competitive advantage in this market. We believe that further adoption of AI, accelerated by increased penetration, more frequent usage, and more complex applications, will continue to rapidly expand the market. According to IDC, investments in AI solutions and services are projected to yield a global cumulative impact of $22.3 trillion by 2030, representing approximately 3.7% of the global gross domestic product (“GDP”). The combined market for AI training infrastructure and our addressable market within AI inference is estimated to be $251 billion in 2025 and is expected to grow to $672 billion by 2029—a 28% CAGR, according to Bloomberg Intelligence. This estimate indicates that AI inference will grow more than twice as fast as AI training infrastructure through 2029. With the fastest inference platform on the market, as benchmarked by Artificial Analysis, and a proven track record in large-scale training, we believe we are well-positioned to capture growth across both parts of the AI infrastructure market. Our growth reflects the broader acceleration of AI adoption. We were incorporated in April 2016 as a Delaware corporation. Our principal executive offices are located in Sunnyvale, California.