Reports confirm OpenAI is advancing its first custom AI chip, focused on inference (running trained models for predictions and decisions), in collaboration with Broadcom for design and intellectual property (IP) and TSMC for manufacturing on a 3nm process node. Mass production is targeted for 2026, aligning with the details in your query. The project is led by former Google engineer Richard Ho, who heads a team of about 40 specialists, many with experience from Google’s Tensor Processing Units (TPUs). This initiative aims to reduce OpenAI’s heavy reliance on Nvidia GPUs, which dominate the AI hardware market but face shortages and high costs.
Key Developments from Recent Reports (September 2025)
- Partnership Confirmation and $10B Deal: On September 5, 2025, the Financial Times and Reuters reported that OpenAI is finalizing the chip design in the coming months, with Broadcom providing engineering support and TSMC handling fabrication. Broadcom’s CEO Hock Tan disclosed a $10 billion order from a new AI client (widely identified as OpenAI) during an earnings call, boosting Broadcom’s AI revenue projections for fiscal 2026. This deal focuses on custom “XPUs” (AI processors) for internal use, not commercial sale, emphasizing inference workloads with potential for scaled training. OpenAI has scaled back earlier ambitions to build its own foundries due to costs exceeding hundreds of millions per iteration, opting instead for this partnership model.
- Team and Technical Specs: Led by Richard Ho (ex-Google TPU head), the team includes engineers like Thomas Norrie. The chip features a systolic array architecture (similar to Google’s TPUs for efficient matrix computations), high-bandwidth memory (HBM, possibly HBM3E or HBM4), and integrated networking. It’s optimized for OpenAI’s models like GPT-4 and beyond, with initial small-scale deployment for inference to test viability. Analysts note risks, including potential delays or underperformance on the first tape-out (design finalization for production), as seen in other custom chip efforts by Microsoft and Meta.
- Market Impact: Broadcom shares surged over 10% on September 5, reaching a $1.7 trillion market cap, while Nvidia and AMD dipped ~2-3% amid concerns over custom silicon eroding Nvidia’s 80%+ market share. HSBC analysts predict the custom AI chip market could surpass Nvidia’s GPU business by 2026. OpenAI’s move ties into broader AI infrastructure pushes, including the $500B Stargate project (with Oracle) and collaborations like Microsoft’s Maia chips.
Broader Context and Challenges
OpenAI’s compute costs are massive—projected $5B loss in 2024 on $3.7B revenue—driving this diversification. The company is also integrating AMD’s MI300X chips via Azure for training, complementing Nvidia. Geopolitical risks (e.g., TSMC’s Taiwan base) and high development costs (~$500M+ per chip version, plus software) loom, but success could enhance bargaining power and efficiency. No official OpenAI statement yet, but industry sources indicate tape-out soon, with prototypes possible by late 2025.
This positions OpenAI alongside Google, Amazon, and Meta in the custom silicon race, potentially reshaping AI hardware dynamics. Updates could emerge from upcoming tech conferences or earnings.