Exclusive: ByteDance in talks with Samsung to manufacture custom AI chip, aiming for 100,000 units this year

ByteDance is reportedly building an inference chip with Samsung, aiming for March samples and about 100k units this year, then a push to 350k. The company disputes it.

Categorized in: AI News IT and Development
Published on: Feb 12, 2026
Exclusive: ByteDance in talks with Samsung to manufacture custom AI chip, aiming for 100,000 units this year

ByteDance's Next Move: An Inference Chip, With Samsung in the Mix

ByteDance is developing an in-house AI inference chip and has held talks with Samsung Electronics for manufacturing and memory supply, according to people familiar with the plans. Sample silicon is targeted by end-March, with at least 100,000 units expected this year and a potential ramp toward 350,000 units.

A company spokesperson said information about an in-house chip project is inaccurate. Samsung declined to comment. Treat timelines and specs as provisional.

What's Reportedly Coming

  • Focus: Inference workloads (serving models at scale), not training.
  • Timeline: Sample chips by end-March.
  • Volume: 100k units this year; plan to scale up to ~350k units.
  • Code name: SeedChip.

Why Samsung Matters

The talks include not just foundry capacity but also memory supply-currently one of the tightest bottlenecks in AI infrastructure. Securing a dependable memory pipeline is as critical as the compute itself for inference throughput and latency.

Context: Diversifying Beyond Nvidia

Global platforms like Google, Amazon, and Microsoft have built custom accelerators to reduce reliance on Nvidia. For Chinese tech firms, U.S. export controls have added pressure to source or develop alternatives. See the U.S. Bureau of Industry and Security for ongoing policy updates here.

ByteDance previously explored an advanced processor with Broadcom, with production tied to TSMC, according to prior reports. Meanwhile, Alibaba unveiled its Zhenwu chip for large AI workloads, and Baidu sells chips to external clients and is preparing to list its Kunlunxin unit.

Budget and Model Strategy

ByteDance plans to spend over 160 billion yuan (~$22B) on AI-related procurement this year, with more than half going to Nvidia hardware, including H200 units (product page), while pushing its own chip effort in parallel.

The chip effort sits alongside a broader model push under the Seed banner. Internally, leadership has acknowledged the company's models trail top-tier players like OpenAI, but backed continued investment across products such as the Doubao chatbot and its overseas sibling, Dola.

Why This Matters for Engineering Teams

  • Cost per token: Purpose-built inference silicon can lower serving costs and reduce dependence on scarce GPUs-if software maturity and memory bandwidth match the claims.
  • Capacity planning: Memory supply is tight. Expect spotty availability and prioritize designs that can swap between accelerators without heavy rewrites.
  • Heterogeneous serving: Prepare your stack (compilers, kernels, runtimes) for mixed fleets. Abstract hardware differences behind common interfaces and strict SLOs.
  • Observability: Make latency budgets, queue depth, and cache hit rates first-class metrics. Inference wins come from memory locality as much as raw TFLOPs.
  • Energy and thermal: More accelerators mean denser racks. Plan for cooling headroom and power-delivery upgrades before the hardware lands.

What to Watch Next

  • Prototype disclosures: Node, memory configuration, interconnect, and supported precisions (FP8/INT8/INT4) for inference.
  • Software stack: Compiler toolchain, runtime, and ecosystem support. Strong kernels and graph optimizations will decide real-world wins.
  • Supply commitments: Any visibility on memory allocations will signal how deployable this is at scale.
  • Competitive moves: Alibaba Zhenwu deployments, Baidu Kunlunxin's listing timeline, and any fresh export-control updates.

The Bottom Line

If ByteDance ships a credible inference part and secures memory, it could ease GPU pressure and improve serving economics across its video, commerce, and cloud products. For now, keep options open, harden portability, and track announcements closely-claims are unconfirmed, and details may change.

Want structured ways to skill up on AI infra and model engineering? Explore focused learning paths by job role here.


Get Daily AI News

Your membership also unlocks:

700+ AI Courses
700+ Certifications
Personalized AI Learning Plan
6500+ AI Tools (no Ads)
Daily AI News by job industry (no Ads)