OpenAI partners with Cerebras

Cerebras 专门打造用于加速 AI 模型生成较长输出的定制系统。它的速度优势来自把海量计算、内存和带宽集中到一块巨型芯片上,从而消除了传统硬件在推理时的瓶颈。

把 Cerebras 纳入我们的算力组合,核心就是让我们的 AI 响应更快。无论是提出复杂问题、生成代码、创作图像,还是驱动 AI 代理,后台都有一个循环:你发出请求,模型“思考”,再返回结果。AI 能实现实时响应时,用户会更频繁使用、停留更久,并运行更高价值的工作负载。

我们将分阶段把这套低延迟能力并入推理堆栈,并逐步覆盖更多类型的工作负载。

“ OpenAI 的计算策略是构建一个弹性的组合,将合适的系统匹配到合适的工作负载。 Cerebras 为我们的平台增加了一种专用的低延迟推理解决方案。这意味着更快的响应、更自然的交互,也为把实时 AI 扩展到更多人群打下了更坚实的基础,” OpenAI 的 Sachin Katti 说。

“我们很高兴能与 OpenAI 合作,把世界领先的 AI 模型带到世界上最快的 AI 处理器上。正如宽带改变了互联网,实时推理将改造 AI,开启构建和使用 AI 模型的全新方式,” Cerebras 联合创始人兼首席执行官 Andrew Feldman 表示。

这些算力将分多批上线,直至 2028 年。

----------------------

Cerebras builds purpose-built AI systems to accelerate long outputs from AI models. Its unique speed comes from putting massive compute, memory, and bandwidth together on a single giant chip and eliminating the bottlenecks that slow inference on conventional hardware.

 Integrating Cerebras into our mix of compute solutions is all about making our AI respond much faster. When you ask a hard question, generate code, create an image, or run an AI agent, there is a loop happening behind the scenes: you send a request, the model thinks, and it sends something back. When AI responds in real time, users do more with it, stay longer, and run higher-value workloads.

We will integrate this low-latency capacity into our inference stack in phases, expanding across workloads.

“OpenAI’s compute strategy is to build a resilient portfolio that matches the right systems to the right workloads. Cerebras adds a dedicated low-latency inference solution to our platform. That means faster responses, more natural interactions, and a stronger foundation to scale real-time AI to many more people,” said Sachin Katti of OpenAI.

“We are delighted to partner with OpenAI, bringing the world’s leading AI models to the world’s fastest AI processor. Just as broadband transformed the internet, real-time inference will transform AI, enabling entirely new ways to build and interact with AI models,” said Andrew Feldman, co-founder and CEO of Cerebras.

The capacity will come online in multiple tranches through 2028.

via OpenAI News
 
 
Back to Top
Copyright © 2025 BESTAI. All rights reserved.
BEST AI API中转 - OpenAI DeepSeek Claude Gemini Grok MidJourney API 2.8折起
[email protected]