
When you purchase through links on our site, we may earn an affiliate commission. Here’s how it works .
According to OpenAI, GPT-5.3-Codex-Spark is tuned for interactive development workflows such as editing specific sections of code and running targeted tests, and the model is optimized for high throughput when served on ultra-low latency hardware. The company claims it can exceed 1,000 tokens per second under the right configuration, while also defaulting to minimal edits, and will not automatically execute tests unless instructed.
The hardware behind all this is Cerebras’ third-generation Wafer Scale Engine . Unlike conventional GPU clusters built from many smaller chips connected over high-speed interconnects, Cerebras uses a single wafer-scale processor with hundreds of thousands of AI cores and large pools of on-chip memory. The architecture is designed to minimize data movement and reduce latency, which is often the bottleneck in interactive inference workloads.
OpenAI declares ‘Code Red’ as Google’s Gemini AI outpaces ChatGPT in industry benchmarks, report claims
Microsoft introduces newest in-house AI chip — Maia 200 is faster than other bespoke Nvidia competitors, built on TSMC 3nm with 216GB of HBM3e
Google TPUs garner attention as AI chip alternative, but are only a minor threat to Nvidia's dominance
OpenAI said last month that it had signed a deal to deploy Cerebras hardware for low-latency inference, and that it plans to bring 750 megawatts of Cerebras-backed compute online in phases through 2028. While that capacity will not replace Nvidia’s role in OpenAI’s training infrastructure, it gives the company a dedicated tier optimized for responsiveness rather than training.
Earlier this month, Sam Altman took to X.com to say that OpenAI loves working with Nvidia, and that “they make the best chips in the world,” adding, “We hope to be a gigantic customer for a very long time.” This came following a controversial report from Reuters that claimed OpenAI is unsatisfied with some Nvidia chips.
OpenAI has also described the partnership with Nvidia as “foundational” and said the company is anchored on Nvidia as the core of its training and inference stack, while also expanding the ecosystem around it through partnerships with Cerebras and others. OpenAI’s most powerful models continue to be trained and served on Nvidia systems.
OpenAI has also agreed to deploy 6 gigawatts in chips from AMD over multiple years and has also struck a deal with Broadcom to develop custom AI accelerators and networking components.
Get Tom's Hardware's best news and in-depth reviews, straight to your inbox.
Key considerations
- Investor positioning can change fast
- Volatility remains possible near catalysts
- Macro rates and liquidity can dominate flows
Reference reading
- https://www.tomshardware.com/tech-industry/artificial-intelligence/SPONSORED_LINK_URL
- https://www.tomshardware.com/tech-industry/artificial-intelligence/openai-lauches-gpt-53-codes-spark-on-cerebras-chips#main
- https://www.tomshardware.com
- Get a 4K upgrade for $200 with MSI's 144Hz gaming monitor — save $80 on dual-mode display in Newegg's Valentine's Day sale
- Intel fined $3 million by India’s antitrust regulator over discriminatory CPU warranty policy — says Intel abused its dominant position in the boxed processor m
- GeForce NOW Celebrates Six Years of Streaming With 24 Games in February
- iFixIt calls BMW’s new anti-consumer security screws 'a logo-shaped middle finger to right to repair,' Adafruit 3D prints a solution — BMW's connector reverse e
- From Pilot to Profit: Survey Reveals the Financial Services Industry Is Doubling Down on AI Investment and Open Source
Informational only. No financial advice. Do your own research.