Connect with us

Hi, what are you looking for?

Top Stories

OpenAI Launches Codex-Spark on Cerebras Chips, Achieving 1,000 Tokens per Second

OpenAI launches Codex-Spark, achieving 1,000 tokens per second on Cerebras chips, as it accelerates efforts to outpace competitors like Google and Anthropic.

OpenAI has unveiled its latest coding model, Codex-Spark, which boasts a processing speed of 1,000 tokens per second. While this figure is noteworthy, it is considered modest compared to the company’s previous benchmarks. Cerebras Systems reported speeds of 2,100 tokens per second on the Llama 3.1 70B model and up to 3,000 tokens per second on OpenAI’s own gpt-oss-120B model. The comparatively lower speed of Codex-Spark suggests the complexities associated with larger models.

This year has marked a significant advancement for AI coding agents, with tools like OpenAI’s Codex and Anthropic’s Claude Code demonstrating enhanced capabilities for developing prototypes and boilerplate code efficiently. In a rapidly evolving tech landscape, latency has emerged as a critical differentiator, with faster coding models enabling developers to iterate more swiftly. The competitive atmosphere has pushed OpenAI and its rivals, including Anthropic and Google, to expedite their development cycles.

OpenAI’s Codex line has seen two rapid iterations in recent months: GPT-5.2 was released in December 2025 after CEO Sam Altman issued a “code red” memo in response to mounting competitive pressure from Google, and the latest iteration, GPT-5.3-Codex, was launched just days ago.

The infrastructure underpinning Codex-Spark is significant not only for its performance metrics but also for its hardware implications. The model operates on Cerebras’ Wafer Scale Engine 3, a chip that is notably large and has been central to Cerebras’ business strategy since 2022. The partnership between OpenAI and Cerebras was formalized in January, with Codex-Spark being the inaugural product of this collaboration.

In a calculated move to diversify its technology sources, OpenAI has been reducing its reliance on Nvidia over the past year. Key developments include a substantial multi-year agreement with AMD signed in October 2025, a $38 billion cloud computing deal with Amazon announced in November, and the design of a proprietary AI chip slated for fabrication by TSMC. Although OpenAI had initially sought a $100 billion infrastructure deal with Nvidia, this has yet to materialize, even as Nvidia has committed to a $20 billion investment.

Reports indicate that OpenAI has grown dissatisfied with the performance speed of certain Nvidia chips, particularly for inference tasks—precisely the type of workload that Codex-Spark is intended to address. In the competitive landscape of AI development, the importance of speed cannot be overstated, even if it may come with trade-offs in accuracy. For developers who rely on AI suggestions while coding, a processing speed of 1,000 tokens per second could feel more like a chainsaw than a precision tool, underscoring the necessity for caution in usage.

As AI coding tools continue to evolve, the interplay between speed and complexity will likely shape the future of software development. Companies like OpenAI and Cerebras are poised to play pivotal roles in this transformation, as they seek to refine their models and enhance their hardware capabilities in an increasingly competitive market.

See also
Staff
Written By

The AiPressa Staff team brings you comprehensive coverage of the artificial intelligence industry, including breaking news, research developments, business trends, and policy updates. Our mission is to keep you informed about the rapidly evolving world of AI technology.

You May Also Like

AI Cybersecurity

Google's Threat Intelligence Group reveals cybercriminals are exploiting its Gemini AI models for real-time malware development, complicating detection and raising security alarms.

Top Stories

Global AI leaders, including Sundar Pichai and Sam Altman, will convene at India's AI Impact Summit 2026 to forge strategic partnerships in a $700B...

AI Cybersecurity

State-sponsored hackers from Iran, North Korea, China, and Russia are now leveraging Google's Gemini AI to enhance cyberattacks, complicating enterprise defenses across sectors.

AI Cybersecurity

Sanctioned Chinese hacking group APT31 exploits Google’s Gemini AI for planning cyberattacks on U.S. organizations, raising urgent cybersecurity concerns.

AI Marketing

Google unveils AI-driven advertising innovations, including AI answer ads and Direct Offers, enhancing targeting and consumer engagement in 2023.

AI Regulation

Mrinank Sharma resigns from Anthropic, citing AI safety concerns, ahead of the AI Impact Summit 2026 in New Delhi, amplifying urgent discussions on ethical...

Top Stories

OpenAI tests a new ad model in ChatGPT, enabling real-time, intent-driven advertising that could redefine digital marketing strategies.

Top Stories

Google's AI tools, including Gemini, now block prompts referencing Disney characters after a cease-and-desist over copyright concerns from Disney.

© 2025 AIPressa · Part of Buzzora Media · All rights reserved. This website provides general news and educational content for informational purposes only. While we strive for accuracy, we do not guarantee the completeness or reliability of the information presented. The content should not be considered professional advice of any kind. Readers are encouraged to verify facts and consult appropriate experts when needed. We are not responsible for any loss or inconvenience resulting from the use of information on this site. Some images used on this website are generated with artificial intelligence and are illustrative in nature. They may not accurately represent the products, people, or events described in the articles.