Mebibyte per second to Gigabyte per second
MiBps
GBps
Conversion History
| Conversion | Reuse | Delete |
|---|---|---|
1 MiBps (Mebibyte per second) → 0.001048576 GBps (Gigabyte per second) Just now |
Quick Reference Table (Mebibyte per second to Gigabyte per second)
| Mebibyte per second (MiBps) | Gigabyte per second (GBps) |
|---|---|
| 1 | 0.001048576 |
| 10 | 0.01048576 |
| 60 | 0.06291456 |
| 125 | 0.131072 |
| 550 | 0.5767168 |
| 1,000 | 1.048576 |
| 7,000 | 7.340032 |
About Mebibyte per second (MiBps)
A mebibyte per second (MiB/s) equals 1,048,576 bytes per second and is the binary unit most commonly seen in operating system disk and memory bandwidth reports. Linux tools like dd, rsync, and hdparm report I/O speeds in MiB/s. Windows Task Manager and Resource Monitor use MB/s, which is decimal. A USB 2.0 high-speed connection peaks at about 60 MiB/s; a SATA SSD reads at 500–600 MiB/s; an NVMe SSD reaches 3,500–7,000 MiB/s.
Running dd on Linux to test disk speed shows results in MiB/s. A SATA III SSD typically reads at around 550 MiB/s.
About Gigabyte per second (GBps)
A gigabyte per second (GB/s or GBps) equals 8,000,000,000 bits per second and is used to measure the performance of high-speed storage interfaces, memory buses, and data center links. PCIe 4.0 ×4 NVMe SSDs achieve around 6–7 GB/s sequential read. DDR5 memory operates at 50–100 GB/s of bandwidth. GPU memory bandwidth reaches 1–2 TB/s on the fastest cards. At 1 GB/s, a 4K movie (50 GB) transfers in about 50 seconds.
A Samsung 990 Pro NVMe SSD reads sequentially at about 7.45 GB/s. PCIe 5.0 ×16 slots provide up to 128 GB/s of theoretical bandwidth.
Mebibyte per second – Frequently Asked Questions
Why does dd report in MiB/s while manufacturers advertise in MB/s?
dd uses binary units because Linux filesystems work in binary block sizes (4 KiB, etc.). Drive manufacturers use decimal MB/s because it makes speeds look about 5% higher and aligns with their decimal capacity marketing. A "550 MB/s" SSD shows roughly 524 MiB/s in dd.
How do I benchmark my disk speed in MiB/s on Linux?
Run "dd if=/dev/zero of=testfile bs=1M count=1024 oflag=direct" and it will report write speed in MiB/s. For read speed, use "dd if=testfile of=/dev/null bs=1M". The oflag=direct flag bypasses filesystem cache to measure actual disk performance.
Is 550 MiB/s the same as 550 MB/s?
No — 550 MiB/s is about 577 MB/s, and 550 MB/s is about 524 MiB/s. The ~5% difference means an SSD advertised at 550 MB/s will show around 524 MiB/s in Linux tools. It is not a defect or false advertising, just different unit systems measuring the same physical speed.
What MiB/s should I expect from a RAID array?
A RAID 0 stripe of two SATA SSDs gives roughly 1,000–1,100 MiB/s sequential reads. Four NVMe SSDs in RAID 0 can hit 12,000–14,000 MiB/s. RAID 5/6 arrays sacrifice some write speed for redundancy — expect 70–90% of raw stripe performance on writes.
Why is random I/O speed so much lower than sequential MiB/s?
Sequential reads let the drive stream data from contiguous locations, maximising throughput. Random I/O forces the controller to seek different addresses, adding latency per operation. An NVMe SSD might do 7,000 MiB/s sequential but only 50–80 MiB/s random (at 4 KiB block size), because the bottleneck shifts from bandwidth to IOPS.
Gigabyte per second – Frequently Asked Questions
Why does RAM bandwidth matter for gaming and productivity?
CPUs constantly shuttle data between RAM and their caches. DDR5-6000 provides about 96 GB/s of bandwidth in dual-channel mode. In games, insufficient RAM bandwidth causes frame drops during complex scenes. In productivity tasks like video encoding, it directly limits how fast the CPU can process data.
How fast is Thunderbolt 4 in GB/s?
Thunderbolt 4 runs at 40 Gbps, which is 5 GB/s. Thunderbolt 5, released in 2024, doubles this to 80 Gbps (10 GB/s) with a burst mode up to 120 Gbps (15 GB/s). This is fast enough to run an external NVMe SSD at near-internal speeds.
What limits SSD speed — the drive or the interface?
Both, depending on generation. A PCIe 3.0 ×4 interface caps at ~3.5 GB/s, bottlenecking modern NAND. PCIe 4.0 ×4 raises this to ~7 GB/s, and PCIe 5.0 ×4 to ~14 GB/s. The drive's NAND flash and controller also have limits — the fastest SSDs and the fastest interfaces are in a constant leapfrog.
How does GPU memory bandwidth reach over 1,000 GB/s?
GPUs use wide memory buses (256–384 bits) with very fast HBM or GDDR6X memory running at high clock speeds. An RTX 4090 has a 384-bit bus with GDDR6X at 21 Gbps per pin, totalling 1,008 GB/s. HBM3 in data center GPUs achieves 3,000+ GB/s through stacked memory with 4096-bit buses.
At GB/s speeds, what becomes the bottleneck?
At multi-GB/s rates, CPU processing speed, software efficiency, and thermal throttling become bottlenecks. A 14 GB/s PCIe 5.0 SSD can deliver data faster than most applications can consume it. Decompression, parsing, and memory allocation in software often cannot keep up with raw storage bandwidth.