Terabyte per second to Bit per second
TBps
bps
Conversion History
| Conversion | Reuse | Delete |
|---|---|---|
1 TBps (Terabyte per second) → 8000000000000 bps (Bit per second) Just now |
Quick Reference Table (Terabyte per second to Bit per second)
| Terabyte per second (TBps) | Bit per second (bps) |
|---|---|
| 0.001 | 8,000,000,000 |
| 0.01 | 80,000,000,000 |
| 0.1 | 800,000,000,000 |
| 1 | 8,000,000,000,000 |
| 3.35 | 26,800,000,000,000 |
| 10 | 80,000,000,000,000 |
About Terabyte per second (TBps)
A terabyte per second (TB/s or TBps) equals 8 terabits per second and represents the bandwidth scale of GPU memory systems, high-performance computing interconnects, and the fastest data center storage fabrics. The HBM3 memory stacks on high-end AI accelerators provide 3–4 TB/s of internal bandwidth. InfiniBand NDR connections used in supercomputers reach 400 Gbps per link, with multiple links aggregated to TB/s totals. At 1 TB/s, the entire contents of a 1 PB data store could transfer in about 17 minutes.
The NVIDIA H100 GPU features 3.35 TB/s of HBM3 memory bandwidth. Top-tier supercomputers like Frontier aggregate over 75 TB/s of storage I/O bandwidth.
About Bit per second (bps)
A bit per second (bps) is the base unit of data transfer rate, representing one binary digit transmitted every second. It is the foundation from which all larger bandwidth units are built. In practice, raw bps figures are useful only for extremely low-speed links — early telegraph systems, narrowband IoT sensors, and some serial control lines operate at tens to thousands of bps. Modern connections are described in kbps, Mbps, or Gbps, making raw bps a reference unit rather than a practical measurement for everyday networking.
Early Morse code telegraph lines transmitted at roughly 10–50 bps. Modern IoT sensors on LoRaWAN networks communicate at 250–50,000 bps.
Terabyte per second – Frequently Asked Questions
Why do AI chips need TB/s of memory bandwidth?
Large language models have billions of parameters that must be read from memory for every inference pass. An LLM with 70 billion parameters at 16-bit precision needs 140 GB of data read per forward pass. At 3 TB/s, the H100 can perform roughly 20 inference passes per second — bandwidth directly determines tokens-per-second output.
Why is memory bandwidth the main bottleneck for large language model inference?
During LLM inference each token requires reading all model weights from memory. A 70-billion-parameter model at 16-bit precision means 140 GB read per forward pass. At 30 tokens per second, that is 4.2 TB/s of memory reads — right at the limit of an H100's HBM3. This is why AI inference is "memory-bound": the GPU's compute cores sit idle waiting for data. Quantising weights to 8-bit or 4-bit halves or quarters the bandwidth demand, directly increasing tokens per second.
What is the fastest memory bandwidth ever achieved in a commercial chip?
The NVIDIA B200 GPU with HBM3e achieves approximately 8 TB/s of memory bandwidth as of 2025. Each generation roughly doubles bandwidth — from 2 TB/s (A100) to 3.35 TB/s (H100) to 4.8 TB/s (H200) to 8 TB/s (B200). The trajectory suggests 16+ TB/s within a few years.
How long would it take to transfer a petabyte at 1 TB/s?
About 16.7 minutes. A petabyte is 1,000 terabytes, so at 1 TB/s, the math is simple division. For context, the Library of Congress contains roughly 10–20 petabytes of data. Transferring it all at 1 TB/s would take about 3–6 hours.
Is there anything beyond TB/s?
Yes — petabytes per second (PB/s). Experimental optical interconnects and photonic computing architectures are pushing toward PB/s-class bandwidth. Some supercomputer storage systems already aggregate into the PB/s range when all nodes operate simultaneously. It is the next frontier for AI training clusters.
Bit per second – Frequently Asked Questions
Why is a bit the smallest unit of data transfer?
A bit represents a single binary choice — 0 or 1 — which is the fundamental quantum of digital information. Every larger unit (byte, kilobit, megabit) is just a multiple of bits. You cannot meaningfully subdivide a binary digit, so bps is the floor of data rate measurement.
What modern devices still operate at bits per second?
LoRaWAN IoT sensors, some RFID readers, and legacy serial ports (RS-232 at 300–9600 baud) still deal in raw bps ranges. Satellites communicating with deep-space probes also use very low bps — NASA's Voyager 1 transmits at about 160 bps from interstellar space.
Is baud the same as bits per second?
Not exactly. Baud measures symbol changes per second, while bps measures bits per second. If each symbol encodes one bit, they are equal. But modern modems encode multiple bits per symbol — a 2400-baud modem using 16-QAM transmits 9600 bps because each symbol carries 4 bits.
How many bits per second does a human speaking convey?
Research suggests human speech carries about 39 bits per second of actual information content, regardless of language. Italian speakers talk faster but convey less information per syllable than Japanese speakers, balancing out to roughly the same bps across all studied languages.
Why did early modems top out at 56,000 bps?
The 56 kbps limit came from the Shannon-Hartley theorem applied to analogue phone lines. The 3.1 kHz bandwidth of a voice telephone channel, combined with its signal-to-noise ratio, creates a theoretical ceiling near 56 kbps. FCC power regulations further capped actual downstream to 53.3 kbps.