At Hot Chips 2025, Google, AMD, and NVIDIA each presented new accelerator designs. For data‑center architects, the signal was less about peak FLOPs and more about memory behavior, compiler‑aware execution, and tighter system integration to support long‑context and reasoning‑heavy workloads 123. That simultaneity offers rare, comparable insight into where the next 18–36 months of AI infrastructure may be headed.
Why Hot Chips 2025 signals an accelerator shift
From peak FLOPs to memory‑first, compiler‑aware design
Hot Chips has long been a venue where vendors expose block‑level choices that ripple into rack design. This year’s talks from Google, AMD, and NVIDIA, as reflected in session materials and reporting, converged on the same pressure points: reasoning‑centric AI, memory bandwidth and locality, and system‑level integration 123. Rather than marketing claims, each session discussed concrete microarchitectural and packaging decisions that let architects model perf/W, memory footprints, and cluster‑level effects. Where phrasing here synthesizes across vendors, treat it as analyst interpretation anchored to the sessions and coverage 123.
What each vendor disclosed
Google Ironwood TPU: architecture and software paths
ServeTheHome’s recap describes Ironwood as oriented to long‑context and reasoning behavior, with changes to tensor pipelines and bandwidth provisions (HBM plus near‑memory tactics) intended to sustain utilization on attention‑heavy and sparse/structured workloads 1. According to the session, the design emphasizes practical bandwidth and locality over headline link rates, with interconnect and placement choices meant to keep critical dataflows close to memory while scaling across devices—patterns that align with RAG and tool‑use pipelines 1. Where STH characterizes support for attention variants and conditional or sparse execution, treat that as coverage of the talk rather than a verbatim vendor claim unless confirmed directly in slides 1.
On software, the stack is reported to expose sparsity and quantization paths through compiler/runtime surfaces, aiming to map structured sparsity without manual kernel work 1. The hardware choices suggest the need for attention‑aware operator fusion, near‑memory scheduling, and disciplined activation management for very long contexts; confirm specific compiler features and release timelines with Google materials before planning migrations 1.
AMD CDNA 4/MI350: chiplet fabric and memory subsystem
AMD’s Hot Chips briefing, as summarized by STH, presents CDNA 4 as a major rework relative to CDNA 3: reorganized compute clusters, updated interconnect/fabric, and a memory subsystem rethink to balance bandwidth and latency at scale 2. MI350 packages these choices in a chiplet‑first design that trades reticle‑limited monoliths for yield and modularity, while working to preserve locality essential to attention‑heavy models 2. Performance and scaling targets were framed across multi‑accelerator nodes bound by AMD fabrics with attention to realistic thermals and duty cycles; treat the efficiency and throughput goals as vendor claims pending independent validation 2.
On the software side, AMD outlined a ROCm‑centered roadmap spanning compilers, graph optimizers, and runtime scheduling, with an explicit aim to reduce porting effort for large transformer and retrieval pipelines and improve out‑of‑the‑box experiences compared with prior generations 2. As with any generational transition, verify framework/library support and toolchain maturity on target models before committing migrations 2.
NVIDIA GB10 SoC: on‑die integration and node design
NVIDIA’s GB10 presentation, per STH coverage, emphasized a system‑on‑chip approach that integrates more node functionality onto a single die: on‑die controllers, an internal fabric, and system‑level blocks for telemetry, security, and I/O alongside accelerator cores and CPU‑adjacent resources 3. The memory discussion focused on tighter coupling for predictable bandwidth and latency, with fewer external components and more coordinated resource management inside the SoC 3.
For data‑center design, the bet is that deeper integration reduces board‑level variance and yields more consistent software behavior across nodes, simplifying provisioning envelopes and scaling predictability—especially for inference fleets where latency SLOs dominate 3. Match any “on‑chip networking” claims to NVIDIA’s exact terminology in the slides (e.g., on‑die fabric vs. NIC) to avoid ambiguity, and validate with vendor materials where possible 3.
Convergent themes—and where strategies diverge
Reasoning and long‑context implications
Across sessions, vendors addressed the shift from pure throughput to mixed fleets that combine training, fine‑tuning, retrieval‑augmented generation, and long‑sequence inference. Each talk prioritized mechanisms to stabilize latency and sustain utilization under long contexts and mixed dense/sparse graphs—though the emphasis and terminology differed by vendor 123.
Memory, bandwidth, and locality trade‑offs
All three positioned memory systems as first‑order: HBM capacity/bandwidth and near‑memory scheduling (Google), fabric‑aware chiplets and a reworked memory subsystem (AMD), and SoC‑level coupling to reduce external components and manage resources holistically (NVIDIA) 123. The practical question for architects is how these choices preserve locality for attention blocks while scaling across devices without negating gains on the wire.
Chiplets vs SoC vs TPU programming models
The strategic splits matter. AMD’s chiplets trade packaging complexity for yield and modular scaling; NVIDIA’s SoC path trades component flexibility for integration predictability; Google advances a TPU programming model tuned to targeted model behaviors. Those choices imply different cost curves, scaling patterns, and software lock‑in dynamics that buyers should test against their workloads and operational constraints 123.
Practical guidance for architects and buyers
Metrics that matter for long‑context workloads
- End‑to‑end latency distributions at long sequence lengths (p50–p99.9), including retrieval hops where applicable
- Sustained tokens/s under mixed dense/sparse graphs and structured attention
- Energy per token (inference) or per step (training), measured at realistic duty cycles
Pilot design and procurement checkpoints
- Run head‑to‑head pilots on representative long‑context jobs; confirm compiler features (attention‑aware fusion, sparsity/quantization paths, near‑memory scheduling) on your target models 123
- Validate node/rack envelopes early: thermals, power delivery, and fabric topology under mixed traffic; ensure network paths don’t erase near‑memory gains 123
- Structure flexible contracts to allow heterogeneous capacity mixes while software/tooling matures; prioritize observability and profiling access across stacks 123
Risks, validation, and open questions
Benchmarks, software maturity, and supply timelines
Independent measurements should precede scale‑up. Prioritize apples‑to‑apples runs with clear model configs and traffic mixes to verify long‑sequence latency, sustained throughput under mixed workloads, and energy per token/step 123. Software readiness is a gating factor: each vendor outlined compiler/runtime work, but field maturity varies and gaps can force hand‑tuning of memory residency or attention kernels, delaying migrations 123. On supply and operations, packaging choices concentrate different risks (chiplet assembly and interconnect complexity vs. large‑die yields), and all designs remain sensitive to HBM availability and thermal design limits—common industry constraints that can shift sampling and volume timelines. Seek clarity on thermal envelopes, serviceability at rack scale, and how each vendor manages component variability across a product’s life 123.
18–36 month outlook
This is an analyst forecast based on Hot Chips materials and early adoption patterns. Expect near‑term wins to cluster around inference and real‑time reasoning tiers where long‑context latency and memory behavior drive business value, with broader training uptake following once compilers and runtimes consistently sustain utilization on attention‑heavy phases without manual tuning 123. Cloud providers are likely to pilot vendor‑specific stacks in targeted services—RAG endpoints, long‑document summarization, and tool‑use agents—before broader fleet conversions. Heterogeneous nodes should become more common: chiplet‑based designs for flexible scaling, SoC‑integrated nodes for predictable ops, and reasoning‑centric parts for latency‑sensitive tiers. Ultimately, what wins in production is not peak FLOPs, but stable long‑context latency, sustained throughput, and software that reliably delivers both at scale 123.




