Gigabyte per second to Byte per second
GBps
Bps
Conversion History
| Conversion | Reuse | Delete |
|---|---|---|
1 GBps (Gigabyte per second) → 1000000000 Bps (Byte per second) Just now |
Quick Reference Table (Gigabyte per second to Byte per second)
| Gigabyte per second (GBps) | Byte per second (Bps) |
|---|---|
| 0.5 | 500,000,000 |
| 1 | 1,000,000,000 |
| 6 | 6,000,000,000 |
| 10 | 10,000,000,000 |
| 16 | 16,000,000,000 |
| 64 | 64,000,000,000 |
| 128 | 128,000,000,000 |
About Gigabyte per second (GBps)
A gigabyte per second (GB/s or GBps) equals 8,000,000,000 bits per second and is used to measure the performance of high-speed storage interfaces, memory buses, and data center links. PCIe 4.0 ×4 NVMe SSDs achieve around 6–7 GB/s sequential read. DDR5 memory operates at 50–100 GB/s of bandwidth. GPU memory bandwidth reaches 1–2 TB/s on the fastest cards. At 1 GB/s, a 4K movie (50 GB) transfers in about 50 seconds.
A Samsung 990 Pro NVMe SSD reads sequentially at about 7.45 GB/s. PCIe 5.0 ×16 slots provide up to 128 GB/s of theoretical bandwidth.
About Byte per second (Bps)
A byte per second (B/s or Bps) is the base byte-based unit of data transfer rate, equal to 8 bits per second. While ISPs advertise in bits per second, download managers, operating systems, and file transfer tools display speeds in bytes per second — a direct measure of how quickly usable file data arrives. The conversion between bits and bytes is constant: divide Mbps by 8 to get MB/s. At 1 B/s, transferring a 1 MB file would take about 11.5 days.
An old dial-up connection at 56 kbps delivered roughly 7,000 B/s (7 kB/s) of actual file data. USB 2.0 maxes out at about 60,000,000 B/s (60 MB/s).
Gigabyte per second – Frequently Asked Questions
Why does RAM bandwidth matter for gaming and productivity?
CPUs constantly shuttle data between RAM and their caches. DDR5-6000 provides about 96 GB/s of bandwidth in dual-channel mode. In games, insufficient RAM bandwidth causes frame drops during complex scenes. In productivity tasks like video encoding, it directly limits how fast the CPU can process data.
How fast is Thunderbolt 4 in GB/s?
Thunderbolt 4 runs at 40 Gbps, which is 5 GB/s. Thunderbolt 5, released in 2024, doubles this to 80 Gbps (10 GB/s) with a burst mode up to 120 Gbps (15 GB/s). This is fast enough to run an external NVMe SSD at near-internal speeds.
What limits SSD speed — the drive or the interface?
Both, depending on generation. A PCIe 3.0 ×4 interface caps at ~3.5 GB/s, bottlenecking modern NAND. PCIe 4.0 ×4 raises this to ~7 GB/s, and PCIe 5.0 ×4 to ~14 GB/s. The drive's NAND flash and controller also have limits — the fastest SSDs and the fastest interfaces are in a constant leapfrog.
How does GPU memory bandwidth reach over 1,000 GB/s?
GPUs use wide memory buses (256–384 bits) with very fast HBM or GDDR6X memory running at high clock speeds. An RTX 4090 has a 384-bit bus with GDDR6X at 21 Gbps per pin, totalling 1,008 GB/s. HBM3 in data center GPUs achieves 3,000+ GB/s through stacked memory with 4096-bit buses.
At GB/s speeds, what becomes the bottleneck?
At multi-GB/s rates, CPU processing speed, software efficiency, and thermal throttling become bottlenecks. A 14 GB/s PCIe 5.0 SSD can deliver data faster than most applications can consume it. Decompression, parsing, and memory allocation in software often cannot keep up with raw storage bandwidth.
Byte per second – Frequently Asked Questions
Why is a byte the fundamental unit of file storage but not of network speed?
Files are stored in bytes because CPUs address memory in byte-sized (8-bit) chunks — the smallest unit a program can read or write. Networks measure in bits because physical signals on a wire or fiber are serial: one bit at a time, clocked at a specific frequency. A 1 GHz signal produces 1 Gbps, not 1 GBps. The two worlds evolved independently and neither adopted the other's convention, leaving users to divide by 8 forever.
Is a byte always 8 bits?
In modern computing, yes — a byte is universally 8 bits. Historically, some architectures used 6, 7, or 9-bit bytes, which is why the unambiguous term "octet" exists in networking standards. But for all practical bandwidth conversions today, 1 byte = 8 bits.
Why is actual file download speed always less than the connection speed in bytes?
Network protocols add overhead — TCP headers, encryption (TLS), error correction, and packet framing all consume bandwidth without contributing to file data. A 100 Mbps connection might deliver 11 MB/s instead of the theoretical 12.5 MB/s because 10–15% goes to protocol overhead.
How many bytes per second does USB 3.0 actually transfer?
USB 3.0 has a theoretical maximum of 625 MB/s (5 Gbps ÷ 8), but real-world sustained transfers hit 300–400 MB/s due to protocol overhead and controller limitations. USB 3.2 Gen 2 doubles this to about 700–900 MB/s in practice.
What came first — the bit or the byte?
The bit came first, coined by Claude Shannon in 1948. The byte was introduced at IBM in the mid-1950s by Werner Buchholz to describe the smallest addressable group of bits in the IBM Stretch computer. Originally it could be any size; the 8-bit byte became standard with the IBM System/360 in 1964.