
What began with Scale-Up and Scale-Out architectures has expanded into Scale-Across, where distributed sites operate as one. These three dimensions now define the blueprint for performance, resilience, and efficiency in large-scale AI.
In modern AI Data Centers (AIDC), “scaling” is the central theme that defines computing and networking evolution. Traditionally, two major expansion models have shaped data center design:
However, with the explosive rise of Generative AI (AIGC) and trillion-parameter models, NVIDIA introduced a third dimension:
Together, these three models form the foundation of today’s AI compute and networking architecture.

Established shortly after ChatGPT’s launch, with the support of Wistron, Foxconn, and Pegatron, Zettabyte emerged to combine the world’s leading GPU and data center supply chain with a sovereign-grade, neutral software stack.
Established shortly after ChatGPT’s launch, with the support of Wistron, Foxconn, and Pegatron, Zettabyte emerged to combine the world’s leading GPU and data center supply chain with a sovereign-grade, neutral software stack.
Scale-Up improves the performance of a single compute node by increasing GPU count, memory bandwidth, or chip-to-chip interconnect performance.
Key Characteristics:
Typical Technologies:
Scale-Out increases total compute capacity by adding more servers or racks connected via RDMA-capable networks.
Key Characteristics:
Typical Deployments:
| Workload Type | Communication Pattern | Best Scaling Method |
|---|---|---|
| Tensor Parallelism | Heavily memory-dependent, frequent small exchanges | Scale-Up |
| Expert Parallelism (MoE) | High bandwidth + low latency | Scale-Up |
| Data Parallelism | Bulk gradient exchanges | Scale-Out |
| Pipeline Parallelism | Stage-level messaging | Scale-Out |
Scale-Up networks use memory-semantic load–store communication, while Scale-Out relies on message-passing semantics. This fundamental difference makes the two architectures inherently non-interchangeable.
Static Latency
Dynamic Latency
| Item | Scale-Up | Scale-Out |
|---|---|---|
| Medium | Copper cables | Optical modules |
| Bandwidth | TB/s-level | 100–800 Gb/s, up to several Tb/s |
| Latency | Extremely low | Higher (due to DSP, routing, etc.) |
| Power | Very low | Higher (optical DSP power) |
At scale: Scale-Up uses copper, Scale-Out uses optics.
Both provide 256 GPUs, but communication characteristics differ drastically.
NVIDIA NVL72 integrates 36 Grace CPUs and 72 Blackwell GPUs into a liquid-cooled cabinet, forming the first true cabinet-scale SuperNode.
NVLink 5 + NVSwitch Characteristics:
Performance Highlights:
NVL72 behaves as a multi-server, cabinet-level unified compute node.
A SuperPOD of 8× NVL72 units (576 GPUs) is built by:
Two-Tier Network Architecture:
Scale-Across extends AI networking beyond a single physical site.
Single data centers face hard limits:
Example:
This is infeasible within a single facility.
Key Components:
HCF Benefits:
Scale-Across introduces a new inter-DC communication layer on top of traditional L3 architectures.
| Technology | Role |
|---|---|
| Hollow-Core Fiber | Ultra-low latency transport (light travels in air) |
| Coherent Transceivers | High-speed long-haul inter-DC communication |
| Optical Circuit Switches (OCS) | Massive bandwidth + port density, ideal for Spine/DCI tiers |
OCS is projected to capture >50% of traditional switch market share over time.
| Dimension | Scale-Up | Scale-Out | Scale-Across |
|---|---|---|---|
| Physical Scope | Server / chassis / cabinet | Within one data center | Across multiple data centers |
| Communication Semantics | Load–store (memory semantic) | Message passing | Cross-region unified fabric |
| Latency | ns | μs–ms | ms (optimized) |
| Bandwidth | TB/s | 100–800 Gb/s | Depends on optical modules |
| Medium | Copper | Optical | Coherent optics / HCF |
| Key Technologies | NVLink, UALink, SUE | InfiniBand, RoCE, UEC | Spectrum-XGS, OCS |
| Bottlenecks | PCB area, cooling, power | Congestion control, topology | DCI cost, optical power |
| Main Use Cases | TP/EP | DP/PP | Giant-scale training, AI factories |



The three dimensions will define the architecture of next-generation AI factories:
Future AI facilities will:
Scale-Across is the foundational step toward this future.