Mebibyte per second to Terabyte per second
MiBps
TBps
Conversion History
| Conversion | Reuse | Delete |
|---|---|---|
1 MiBps (Mebibyte per second) → 0.000001048576 TBps (Terabyte per second) Just now |
Quick Reference Table (Mebibyte per second to Terabyte per second)
| Mebibyte per second (MiBps) | Terabyte per second (TBps) |
|---|---|
| 1 | 0.000001048576 |
| 10 | 0.00001048576 |
| 60 | 0.00006291456 |
| 125 | 0.000131072 |
| 550 | 0.0005767168 |
| 1,000 | 0.001048576 |
| 7,000 | 0.007340032 |
About Mebibyte per second (MiBps)
A mebibyte per second (MiB/s) equals 1,048,576 bytes per second and is the binary unit most commonly seen in operating system disk and memory bandwidth reports. Linux tools like dd, rsync, and hdparm report I/O speeds in MiB/s. Windows Task Manager and Resource Monitor use MB/s, which is decimal. A USB 2.0 high-speed connection peaks at about 60 MiB/s; a SATA SSD reads at 500–600 MiB/s; an NVMe SSD reaches 3,500–7,000 MiB/s.
Running dd on Linux to test disk speed shows results in MiB/s. A SATA III SSD typically reads at around 550 MiB/s.
About Terabyte per second (TBps)
A terabyte per second (TB/s or TBps) equals 8 terabits per second and represents the bandwidth scale of GPU memory systems, high-performance computing interconnects, and the fastest data center storage fabrics. The HBM3 memory stacks on high-end AI accelerators provide 3–4 TB/s of internal bandwidth. InfiniBand NDR connections used in supercomputers reach 400 Gbps per link, with multiple links aggregated to TB/s totals. At 1 TB/s, the entire contents of a 1 PB data store could transfer in about 17 minutes.
The NVIDIA H100 GPU features 3.35 TB/s of HBM3 memory bandwidth. Top-tier supercomputers like Frontier aggregate over 75 TB/s of storage I/O bandwidth.
Mebibyte per second – Frequently Asked Questions
Why does dd report in MiB/s while manufacturers advertise in MB/s?
dd uses binary units because Linux filesystems work in binary block sizes (4 KiB, etc.). Drive manufacturers use decimal MB/s because it makes speeds look about 5% higher and aligns with their decimal capacity marketing. A "550 MB/s" SSD shows roughly 524 MiB/s in dd.
How do I benchmark my disk speed in MiB/s on Linux?
Run "dd if=/dev/zero of=testfile bs=1M count=1024 oflag=direct" and it will report write speed in MiB/s. For read speed, use "dd if=testfile of=/dev/null bs=1M". The oflag=direct flag bypasses filesystem cache to measure actual disk performance.
Is 550 MiB/s the same as 550 MB/s?
No — 550 MiB/s is about 577 MB/s, and 550 MB/s is about 524 MiB/s. The ~5% difference means an SSD advertised at 550 MB/s will show around 524 MiB/s in Linux tools. It is not a defect or false advertising, just different unit systems measuring the same physical speed.
What MiB/s should I expect from a RAID array?
A RAID 0 stripe of two SATA SSDs gives roughly 1,000–1,100 MiB/s sequential reads. Four NVMe SSDs in RAID 0 can hit 12,000–14,000 MiB/s. RAID 5/6 arrays sacrifice some write speed for redundancy — expect 70–90% of raw stripe performance on writes.
Why is random I/O speed so much lower than sequential MiB/s?
Sequential reads let the drive stream data from contiguous locations, maximising throughput. Random I/O forces the controller to seek different addresses, adding latency per operation. An NVMe SSD might do 7,000 MiB/s sequential but only 50–80 MiB/s random (at 4 KiB block size), because the bottleneck shifts from bandwidth to IOPS.
Terabyte per second – Frequently Asked Questions
Why do AI chips need TB/s of memory bandwidth?
Large language models have billions of parameters that must be read from memory for every inference pass. An LLM with 70 billion parameters at 16-bit precision needs 140 GB of data read per forward pass. At 3 TB/s, the H100 can perform roughly 20 inference passes per second — bandwidth directly determines tokens-per-second output.
Why is memory bandwidth the main bottleneck for large language model inference?
During LLM inference each token requires reading all model weights from memory. A 70-billion-parameter model at 16-bit precision means 140 GB read per forward pass. At 30 tokens per second, that is 4.2 TB/s of memory reads — right at the limit of an H100's HBM3. This is why AI inference is "memory-bound": the GPU's compute cores sit idle waiting for data. Quantising weights to 8-bit or 4-bit halves or quarters the bandwidth demand, directly increasing tokens per second.
What is the fastest memory bandwidth ever achieved in a commercial chip?
The NVIDIA B200 GPU with HBM3e achieves approximately 8 TB/s of memory bandwidth as of 2025. Each generation roughly doubles bandwidth — from 2 TB/s (A100) to 3.35 TB/s (H100) to 4.8 TB/s (H200) to 8 TB/s (B200). The trajectory suggests 16+ TB/s within a few years.
How long would it take to transfer a petabyte at 1 TB/s?
About 16.7 minutes. A petabyte is 1,000 terabytes, so at 1 TB/s, the math is simple division. For context, the Library of Congress contains roughly 10–20 petabytes of data. Transferring it all at 1 TB/s would take about 3–6 hours.
Is there anything beyond TB/s?
Yes — petabytes per second (PB/s). Experimental optical interconnects and photonic computing architectures are pushing toward PB/s-class bandwidth. Some supercomputer storage systems already aggregate into the PB/s range when all nodes operate simultaneously. It is the next frontier for AI training clusters.