Terabyte per second to Mebibit per second
TBps
Mibps
Conversion History
| Conversion | Reuse | Delete |
|---|---|---|
| No conversion history to show. | ||
Quick Reference Table (Terabyte per second to Mebibit per second)
| Terabyte per second (TBps) | Mebibit per second (Mibps) |
|---|---|
| 0.001 | 7,629.39453125 |
| 0.01 | 76,293.9453125 |
| 0.1 | 762,939.453125 |
| 1 | 7,629,394.53125 |
| 3.35 | 25,558,471.6796875 |
| 10 | 76,293,945.3125 |
About Terabyte per second (TBps)
A terabyte per second (TB/s or TBps) equals 8 terabits per second and represents the bandwidth scale of GPU memory systems, high-performance computing interconnects, and the fastest data center storage fabrics. The HBM3 memory stacks on high-end AI accelerators provide 3–4 TB/s of internal bandwidth. InfiniBand NDR connections used in supercomputers reach 400 Gbps per link, with multiple links aggregated to TB/s totals. At 1 TB/s, the entire contents of a 1 PB data store could transfer in about 17 minutes.
The NVIDIA H100 GPU features 3.35 TB/s of HBM3 memory bandwidth. Top-tier supercomputers like Frontier aggregate over 75 TB/s of storage I/O bandwidth.
About Mebibit per second (Mibps)
A mebibit per second (Mibps) equals 1,048,576 bits per second — the binary IEC equivalent of megabit per second. It is approximately 4.9% larger than 1 Mbps. Mibps appears in network performance specifications written to IEC standards, and in operating system tools on Linux and some Unix variants that apply binary prefixes strictly. When a Linux system reports "ethtool: speed 100MiB/s", this distinction from 100 MB/s (decimal) matters in precise bandwidth budgeting.
A 100 Mibps figure represents 104.86 Mbps in decimal — about 5% more data. Network engineers use Mibps when exact binary calculations are required for buffer sizing.
Terabyte per second – Frequently Asked Questions
Why do AI chips need TB/s of memory bandwidth?
Large language models have billions of parameters that must be read from memory for every inference pass. An LLM with 70 billion parameters at 16-bit precision needs 140 GB of data read per forward pass. At 3 TB/s, the H100 can perform roughly 20 inference passes per second — bandwidth directly determines tokens-per-second output.
Why is memory bandwidth the main bottleneck for large language model inference?
During LLM inference each token requires reading all model weights from memory. A 70-billion-parameter model at 16-bit precision means 140 GB read per forward pass. At 30 tokens per second, that is 4.2 TB/s of memory reads — right at the limit of an H100's HBM3. This is why AI inference is "memory-bound": the GPU's compute cores sit idle waiting for data. Quantising weights to 8-bit or 4-bit halves or quarters the bandwidth demand, directly increasing tokens per second.
What is the fastest memory bandwidth ever achieved in a commercial chip?
The NVIDIA B200 GPU with HBM3e achieves approximately 8 TB/s of memory bandwidth as of 2025. Each generation roughly doubles bandwidth — from 2 TB/s (A100) to 3.35 TB/s (H100) to 4.8 TB/s (H200) to 8 TB/s (B200). The trajectory suggests 16+ TB/s within a few years.
How long would it take to transfer a petabyte at 1 TB/s?
About 16.7 minutes. A petabyte is 1,000 terabytes, so at 1 TB/s, the math is simple division. For context, the Library of Congress contains roughly 10–20 petabytes of data. Transferring it all at 1 TB/s would take about 3–6 hours.
Is there anything beyond TB/s?
Yes — petabytes per second (PB/s). Experimental optical interconnects and photonic computing architectures are pushing toward PB/s-class bandwidth. Some supercomputer storage systems already aggregate into the PB/s range when all nodes operate simultaneously. It is the next frontier for AI training clusters.
Mebibit per second – Frequently Asked Questions
When would I encounter mebibits instead of megabits?
Mainly in Linux system tools, IEC-compliant technical specifications, and some enterprise storage documentation. The iperf3 network testing tool can report in Mibps if configured to use binary units. Most consumer-facing software and ISPs use megabits exclusively.
How do I convert Mibps to Mbps?
Multiply by 1.048576. So 100 Mibps = 104.86 Mbps. To go from Mbps to Mibps, divide by 1.048576. At small values the difference is negligible, but at gigabit scales it can mean a meaningful amount of data.
Why does Linux sometimes use binary units for networking?
Linux kernel developers historically followed IEC recommendations to use binary prefixes where applicable. Some tools like dd and rsync default to binary (MiB/s) for disk operations. However, network-facing tools like ethtool and ip still use decimal Mbps because that is what the hardware reports.
Does the 5% difference between Mibps and Mbps matter in practice?
For casual use, no. For capacity planning and SLA compliance, yes. If a contract guarantees 100 Mibps and the provider measures in Mbps, the customer might get 100 Mbps (only 95.4 Mibps) and technically be short-changed. Data center SLAs should specify which unit system applies.
Is my ISP cheating me by using megabits instead of mebibits?
No — ISPs legitimately use decimal megabits because Ethernet and fiber standards are decimal. A "100 Mbps" plan genuinely delivers 100,000,000 bits per second. The confusion arises only when comparing with binary-unit tools. ISPs are not hiding anything; the two systems just coexist awkwardly.