Gigabyte per second to Tebibyte per second

GBps

1 GBps

TiB/s

0.00090949470177292824 TiB/s

Conversion History

ConversionReuseDelete

1 GBps (Gigabyte per second) → 0.00090949470177292824 TiB/s (Tebibyte per second)

Just now

Entries per page:

1–1 of 1


Quick Reference Table (Gigabyte per second to Tebibyte per second)

Gigabyte per second (GBps)Tebibyte per second (TiB/s)
0.50.00045474735088646412
10.00090949470177292824
60.00545696821063756943
100.00909494701772928238
160.01455191522836685181
640.05820766091346740723
1280.11641532182693481445

About Gigabyte per second (GBps)

A gigabyte per second (GB/s or GBps) equals 8,000,000,000 bits per second and is used to measure the performance of high-speed storage interfaces, memory buses, and data center links. PCIe 4.0 ×4 NVMe SSDs achieve around 6–7 GB/s sequential read. DDR5 memory operates at 50–100 GB/s of bandwidth. GPU memory bandwidth reaches 1–2 TB/s on the fastest cards. At 1 GB/s, a 4K movie (50 GB) transfers in about 50 seconds.

A Samsung 990 Pro NVMe SSD reads sequentially at about 7.45 GB/s. PCIe 5.0 ×16 slots provide up to 128 GB/s of theoretical bandwidth.

About Tebibyte per second (TiB/s)

A tebibyte per second (TiB/s) equals 1,099,511,627,776 bytes per second and represents the bandwidth scale of cutting-edge AI accelerator memory and high-performance computing interconnects. The HBM3e memory on NVIDIA H200 GPUs provides approximately 4.8 TiB/s of bandwidth. At this scale, the 10% difference between tebibytes (binary) and terabytes (decimal) matters in system design — a buffer sized for 1 TiB/s must handle 1,099 GB/s in decimal bandwidth.

NVIDIA H200 SXM features 4.8 TiB/s of HBM3e memory bandwidth. Top-end AI training clusters aggregate several TiB/s of storage I/O.


Gigabyte per second – Frequently Asked Questions

CPUs constantly shuttle data between RAM and their caches. DDR5-6000 provides about 96 GB/s of bandwidth in dual-channel mode. In games, insufficient RAM bandwidth causes frame drops during complex scenes. In productivity tasks like video encoding, it directly limits how fast the CPU can process data.

Thunderbolt 4 runs at 40 Gbps, which is 5 GB/s. Thunderbolt 5, released in 2024, doubles this to 80 Gbps (10 GB/s) with a burst mode up to 120 Gbps (15 GB/s). This is fast enough to run an external NVMe SSD at near-internal speeds.

Both, depending on generation. A PCIe 3.0 ×4 interface caps at ~3.5 GB/s, bottlenecking modern NAND. PCIe 4.0 ×4 raises this to ~7 GB/s, and PCIe 5.0 ×4 to ~14 GB/s. The drive's NAND flash and controller also have limits — the fastest SSDs and the fastest interfaces are in a constant leapfrog.

GPUs use wide memory buses (256–384 bits) with very fast HBM or GDDR6X memory running at high clock speeds. An RTX 4090 has a 384-bit bus with GDDR6X at 21 Gbps per pin, totalling 1,008 GB/s. HBM3 in data center GPUs achieves 3,000+ GB/s through stacked memory with 4096-bit buses.

At multi-GB/s rates, CPU processing speed, software efficiency, and thermal throttling become bottlenecks. A 14 GB/s PCIe 5.0 SSD can deliver data faster than most applications can consume it. Decompression, parsing, and memory allocation in software often cannot keep up with raw storage bandwidth.

Tebibyte per second – Frequently Asked Questions

AMD's MI300X stacks 8 HBM3 memory modules and multiple compute chiplets on a single package using advanced 2.5D packaging with silicon interposers. The short physical distance between compute and memory dies — millimeters instead of centimeters — dramatically reduces signal latency and power per bit. This allows a 5.3 TB/s aggregate bandwidth that would be physically impossible with traditional socketed memory. The trend toward chiplet packaging is how the industry keeps scaling bandwidth despite hitting limits in single-die manufacturing.

Significantly. When provisioning an AI training cluster with hundreds of GPUs, a 10% bandwidth miscalculation cascades through the entire system design — buffer sizes, interconnect capacity, cooling, and power. Getting the units wrong could mean the difference between a training run finishing in 30 days vs 33 days.

Training large language models (100B+ parameters), molecular dynamics simulations, weather modeling, and fluid dynamics at scale. These workloads move enormous matrices through memory billions of times. The TiB/s memory bandwidth of modern GPUs is what makes training models like GPT-4 possible in months rather than decades.

Memory bandwidth dwarfs network bandwidth. Each H100 GPU has 3.35 TiB/s of internal memory bandwidth but connects to the network at only 0.05 TiB/s (400 Gbps InfiniBand). This 60:1 ratio is why AI chip designers obsess over keeping computations local to each GPU and minimising network communication.

Not in the same way. Quantum computers process information through qubits that exist in superposition, so they do not shuttle classical data around at TiB/s. However, the classical control systems that manage quantum processors and process measurement results do need high bandwidth — current quantum-classical interfaces operate at modest Gbps rates.

© 2026 TopConverters.com. All rights reserved.