Explore

Break Free from Vendor Constraints with zFABRIC™

zFABRIC™ creates one high-performance fabric across all your GPUs, maximizing throughput and minimizing latency for distributed AI workloads.

Neutral backgroundOverjoyed, amazed woman
Why choose us?

Flexible, High-Performance AI Networking

Heterogeneous clusters
Resilient fabric
Scales globally
No vendor lock-in
Lower TCO
Sovereign-ready
100k+
GPUs Unified
0
Proprietary Constraints
99.99%
Uptime Target
What’s New

zFABRIC™ Capabilities

Heterogeneous Fabric
Unifies diverse GPU and accelerator types into one network, so you can scale across mixed hardware without performance loss.
Vendor Independence
Standard RoCEv2 networking works across multiple switch vendors, giving you InfiniBand-like performance on Ethernet with no single-vendor lock-in.
Resilient Networking
Built-in failover and health checks ensure high availability. If a link or node fails, zFABRIC™ reroutes traffic instantly to keep AI workloads running.

Frequently Asked Questions

Have more questions?
Here are quick answers to common inquiries about zFABRIC™

What is zFABRIC?

zFABRIC is a high performance RDMA networking solution purpose built for AI and GPU clusters. zFABRIC is like using high quality, non dealer performance parts in your race car enabling AI clusters to scale efficiently across racks and data centers without relying on closed or vendor specific networking. zFABRIC delivers the performance required for distributed AI training while giving operators flexibility in hardware sourcing avoiding vendor lock-in thus allowing faster deployments and lower long-term operating costs.

How does zFABRIC improve total cost of ownership (TCO)?

zFABRIC lowers CAPEX and OPEX for our customers by enabling mixed hardware generations, supporting multiple network vendors, and reducing downtime through automated recovery. Customers who deploy zFABRIC avoid vendor lock-in, extend hardware lifespan, and reduce operational overhead, significantly improving TCO.

How does zFABRIC improve reliability and uptime?

zFABRIC is designed to keep AI systems productive even when underlying components fail. Through automated failover, continuous link health monitoring, intelligent rerouting, and rapid recovery, zFABRIC minimizes disruption to training and inference workloads. This reduces GPU hang time, protects delivery timelines, and allows operators to meet SLA expectations minimizing manual intervention, resulting in more predictable operations and fewer costly interruptions, or mean time to recovery (MTTR). Overall, zFABRIC and Zettabyte's full product offerings allow organizations to bring systems online quickly while maintaining control and operational continuity.

Is zFABRIC limited to NVIDIA GPUs?

No, zFABRIC is vendor agnostic and supports heterogeneous GPU and accelerator environments based on open RDMA standards such as RoCEv2. This allows organizations to deploy and operate AI infrastructure using NVIDIA, AMD, or other accelerators without being locked into a single vendor ecosystem. As a result, customers can source hardware more flexibly, extend the usable life of existing assets, adapt faster to supply or pricing changes, and reduce long-term infrastructure costs while maintaining consistent performance at scale.

Which networking protocol does zFABRIC use and why?

zFABRIC primarily uses RoCEv2 (RDMA over Converged Ethernet) to deliver high-performance GPU networking on standard Ethernet infrastructure. This enables near InfiniBand performance while using widely available switches, optics, and cabling. As a result, customers can deploy AI clusters more quickly, scale across vendors and sites with less friction, and achieve high performance without the cost and constraints of proprietary networking stacks.

How many GPUs does zFABRIC support?

zFABRIC is designed to scale from thousands to hundreds of thousands of GPUs within a single AI environment. Scaling limits are determined by physical factors such as optics speed, switch capacity, and data center power and cooling, not by the zFABRIC software itself. This allows organizations to start at practical cluster sizes and expand over time without redesigning the network thus reducing deployment delays, protecting existing investments, and avoiding premature infrastructure replacement.

Can zFABRIC support cross data center AI clusters?

Yes, zFABRIC enables AI training and inference to run across geographically distributed data centers, allowing organizations to scale beyond a single site without redesigning their network. This makes it possible to bring capacity online faster, use existing facilities more effectively, and avoid costly overbuild in one location. By supporting long distance interconnection with production ready designs, zFABRIC allows teams to operate distributed AI systems reliably while improving utilization and lowering the total cost of scaling AI infrastructure.

Our products

Integrated Solutions forHigh-Performance AI Infrastructure