Gibibyte per second to Tebibyte per second
GiB/s
TiB/s
Conversion History
| Conversion | Reuse | Delete |
|---|---|---|
1 GiB/s (Gibibyte per second) → 0.0009765625 TiB/s (Tebibyte per second) Just now |
Quick Reference Table (Gibibyte per second to Tebibyte per second)
| Gibibyte per second (GiB/s) | Tebibyte per second (TiB/s) |
|---|---|
| 0.5 | 0.00048828125 |
| 1 | 0.0009765625 |
| 7 | 0.0068359375 |
| 12 | 0.01171875 |
| 50 | 0.048828125 |
| 100 | 0.09765625 |
| 1,008 | 0.984375 |
About Gibibyte per second (GiB/s)
A gibibyte per second (GiB/s) equals 1,073,741,824 bytes per second and is used in high-performance storage and memory bandwidth measurements when binary precision is required. GPU memory bandwidth figures in technical documentation sometimes appear in GiB/s — an NVIDIA RTX 4090 features 1,008 GiB/s of GDDR6X memory bandwidth. NVMe SSD sequential read speeds are often reported as both GB/s (decimal) and GiB/s (binary) in reviews and datasheets.
The NVIDIA RTX 4090 GPU has 1,008 GiB/s of memory bandwidth (~1,082 GB/s in decimal). DDR5-6400 dual-channel memory provides about 100 GiB/s.
About Tebibyte per second (TiB/s)
A tebibyte per second (TiB/s) equals 1,099,511,627,776 bytes per second and represents the bandwidth scale of cutting-edge AI accelerator memory and high-performance computing interconnects. The HBM3e memory on NVIDIA H200 GPUs provides approximately 4.8 TiB/s of bandwidth. At this scale, the 10% difference between tebibytes (binary) and terabytes (decimal) matters in system design — a buffer sized for 1 TiB/s must handle 1,099 GB/s in decimal bandwidth.
NVIDIA H200 SXM features 4.8 TiB/s of HBM3e memory bandwidth. Top-end AI training clusters aggregate several TiB/s of storage I/O.
Gibibyte per second – Frequently Asked Questions
Why do GPU specs sometimes use GiB/s instead of GB/s?
GPU memory is addressed in binary (power-of-2 bus widths like 256-bit or 384-bit), so binary units naturally describe the actual hardware capability. Some vendors use GiB/s to be precise, while marketing materials prefer the larger-sounding GB/s number. The RTX 4090's 1,008 GiB/s is 1,082 GB/s — the latter sounds faster.
How much GiB/s bandwidth does DDR5 RAM provide?
DDR5-6000 in dual-channel mode provides about 93 GiB/s (100 GB/s). Quad-channel DDR5 on workstation platforms doubles this to ~186 GiB/s. The actual usable bandwidth depends on memory access patterns — random access achieves far less than sequential streaming.
What is the difference between memory bandwidth and storage bandwidth?
Memory bandwidth (50–100+ GiB/s for DDR5) measures how fast the CPU can read/write RAM. Storage bandwidth (3–14 GiB/s for NVMe SSDs) measures persistent data transfer. Memory is 10–30× faster because DRAM has nanosecond latency while NAND flash has microsecond latency. They serve different roles in the data hierarchy.
Can I measure GiB/s bandwidth on my own system?
Yes. For memory bandwidth, run a STREAM benchmark (available for Linux and Windows). For storage, use fio or CrystalDiskMark. GPU memory bandwidth can be tested with gpu-burn or vendor-provided tools. All will report in either GiB/s or GB/s depending on the tool — check which one.
At what GiB/s does data transfer become limited by physics?
Electrical signalling on copper traces maxes out around 112 Gbps (about 13 GiB/s) per lane with current technology. Beyond that, optics take over — silicon photonics interconnects can push individual channels to 200+ Gbps. The physical speed of light in fiber is not the limit; it is the modulation and detection electronics.
Tebibyte per second – Frequently Asked Questions
How do chiplet architectures like AMD's MI300X achieve massive bandwidth through packaging?
AMD's MI300X stacks 8 HBM3 memory modules and multiple compute chiplets on a single package using advanced 2.5D packaging with silicon interposers. The short physical distance between compute and memory dies — millimeters instead of centimeters — dramatically reduces signal latency and power per bit. This allows a 5.3 TB/s aggregate bandwidth that would be physically impossible with traditional socketed memory. The trend toward chiplet packaging is how the industry keeps scaling bandwidth despite hitting limits in single-die manufacturing.
How much does the 10% TiB vs TB difference matter for AI training?
Significantly. When provisioning an AI training cluster with hundreds of GPUs, a 10% bandwidth miscalculation cascades through the entire system design — buffer sizes, interconnect capacity, cooling, and power. Getting the units wrong could mean the difference between a training run finishing in 30 days vs 33 days.
What workloads actually need TiB/s of bandwidth?
Training large language models (100B+ parameters), molecular dynamics simulations, weather modeling, and fluid dynamics at scale. These workloads move enormous matrices through memory billions of times. The TiB/s memory bandwidth of modern GPUs is what makes training models like GPT-4 possible in months rather than decades.
How does TiB/s memory bandwidth compare to network bandwidth in AI clusters?
Memory bandwidth dwarfs network bandwidth. Each H100 GPU has 3.35 TiB/s of internal memory bandwidth but connects to the network at only 0.05 TiB/s (400 Gbps InfiniBand). This 60:1 ratio is why AI chip designers obsess over keeping computations local to each GPU and minimising network communication.
Could quantum computers need TiB/s bandwidth?
Not in the same way. Quantum computers process information through qubits that exist in superposition, so they do not shuttle classical data around at TiB/s. However, the classical control systems that manage quantum processors and process measurement results do need high bandwidth — current quantum-classical interfaces operate at modest Gbps rates.