OpenAI announced it is partnering with Cerebras to add 750MW of ultra low-latency AI compute to its platform. Terms were not disclosed. The company stated: “Integrating Cerebras into our mix of compute solutions is all about making our AI respond much faster. When you ask a hard question, generate code, create an image, or run an AI agent, there is a loop happening behind the scenes: you send a request, the model thinks, and it sends something back. When AI responds in real time, users do more with it, stay longer, and run higher-value workloads. We will integrate this low-latency capacity into our inference stack in phases, expanding across workloads.” Sachin Katti of OpenAI added: “OpenAI’s compute strategy is to build a resilient portfolio that matches the right systems to the right workloads. Cerebras adds a dedicated low-latency inference solution to our platform. That means faster responses, more natural interactions, and a stronger foundation to scale real-time AI to many more people.”
Claim 70% Off TipRanks Premium
- Unlock hedge fund-level data and powerful investing tools for smarter, sharper decisions
- Stay ahead of the market with the latest news and analysis and maximize your portfolio's potential
Published first on TheFly – the ultimate source for real-time, market-moving breaking financial news. Try Now>>
Read More on MSFT:
- OpenAI in over $10B computing partnership with Cerebras, WSJ says
- Option traders moderately bearish in Microsoft with shares down 2.3%
- Bondholders sue Oracle over losses linked to AI efforts, Reuters reports
- Reports Claim Anthropic and OpenAI Are Preparing to Go Public
- AI Daily: U.S. to review Nvidia, AMD chip exports case by case
