OpenAI Picks Broadcom for Custom AI Accelerators - Samsung Poised to Gain via HBM Supply
OpenAI has confirmed a partnership with Broadcom to develop custom AI accelerators, targeting a 10GW-class infrastructure buildout. The first chips are slated to hit data centers in the second half of 2026, with rollout completing by 2029. Broadcom will also provide Ethernet, PCIe, and optical connectivity for the racks that host these accelerators.
The move formalizes long-circulating reports and signals a push to reduce cost per token, improve efficiency, and diversify beyond NVIDIA and AMD. Broadcom reportedly edged out Marvell, helped by its track record co-developing Google's TPU hardware stack for cloud-scale training and inference.
Why this matters for engineering and infra teams
This is a bet on vertical integration: optimize silicon, interconnect, and data center design around specific model workloads. Expect tighter coupling between accelerators, memory bandwidth, and Ethernet-based AI fabrics. If you're building or maintaining training clusters, prepare for an environment where ASIC-based accelerators sit alongside GPUs.
Practically, that means prioritizing portability and abstraction: containerized workflows, scheduler-agnostic orchestration (Kubernetes), and model runtimes that can adapt to new compiler backends. Keep an eye on Ethernet-based fabrics (e.g., RoCE-style deployments) as alternatives to proprietary interconnects.
What Broadcom is bringing
- Custom AI accelerators co-developed with OpenAI, optimized for efficiency and throughput.
- Rack-level networking: Ethernet, PCIe, and optical links to tie accelerators together at scale.
- Design-house experience delivering ASICs for hyperscalers, reducing time-to-deployment risk.
For teams planning capacity, this hints at Ethernet-forward AI clusters where congestion management, NIC offloads, and lossless transport tuning move from "nice to have" to required.
Samsung's opening via HBM3E and HBM4
Samsung Electronics is positioned to benefit as a supplier of HBM3E to Broadcom, with discussions reportedly underway for next-gen HBM4. As AI accelerators become more memory-bound, sustained HBM throughput and thermals will define real-world performance.
If you're evaluating future hardware, weight memory bandwidth, capacity per package, and cooling constraints as heavily as raw compute. HBM supply and packaging yields can make or break delivery timelines.
Market reaction (and what it signals)
The announcement lifted semiconductor names across the board: Broadcom jumped nearly 10%, with NVIDIA, TSMC, and Micron also up. AMD gained modestly after previously rallying on OpenAI-related news. Read this as a signal that multi-vendor, multi-accelerator roadmaps are becoming standard for hyperscale AI.
Timeline and planning
- H2 2026: OpenAI/Broadcom accelerators begin deployment.
- Through 2029: Distribution ramps toward the 10GW target.
Translation: Current-gen and next-gen GPUs will remain central for at least the next 24-36 months. If you're responsible for fleet strategy, avoid lock-in to a single toolchain. Build CI/CD for models that can target multiple backends with minimal refactoring.
10GW is huge - and expensive
OpenAI's announced capacity equals roughly 10 nuclear reactors worth of power. Bloomberg estimates peg 1GW AI data centers at around $35 billion to build. With OpenAI's enterprise value near $500 billion and negative cash flow projected through the decade, execution risk is real.
Expect staggered buildouts, location-dependent power constraints, and extended lead times on power and cooling infrastructure. If your workloads depend on new capacity, model for slippage and design for burst usage across multiple providers.
What to watch next
- HBM supply and pricing: Can Samsung, SK hynix, and Micron meet demand as HBM3E ramps and HBM4 approaches?
- Fabric choices: Further validation of Ethernet-based AI fabrics, including RDMA configurations and switch silicon roadmaps.
- Compiler and framework support: Toolchains for Broadcom/OpenAI accelerators, and how quickly they reach parity for training and inference.
- Packaging and cooling: Any signals on advanced packaging capacity and liquid cooling adoption in these racks.
- Capex and power: Siting decisions, grid partnerships, and phased deployment schedules against the 10GW target.
How this affects your roadmap
- Adopt a multi-accelerator strategy: design inference/training stacks that can target GPUs and ASICs.
- Invest in observability for Ethernet-based AI clusters: end-to-end latency, ECN/queue health, and flow control.
- Abstract model execution: use portable runtimes, versioned artifacts, and reproducible builds to swap backends cleanly.
- Secure memory-first telemetry: watch HBM utilization, memory bandwidth bottlenecks, and thermal headroom.
- Negotiate flexible procurement: plan for split fleets and staggered deliveries.
Bottom line
OpenAI's partnership with Broadcom accelerates a shift toward custom silicon at hyperscale. Samsung stands to gain as an HBM supplier, and memory roadmaps will be as consequential as compute. For engineering teams, the winning move is optionality: build portable pipelines, prepare for Ethernet-first AI fabrics, and keep procurement and deployment plans adaptable through 2029.
OpenAI and Broadcom will release more details as deployment nears; align your capacity and tooling plans to accommodate new compilers, drivers, and fabric tuning.
If you're upskilling on AI infrastructure and accelerator tooling, explore curated programs at Complete AI Training.
Your membership also unlocks: