"590 GFLOPS" appears at the top of your 9bench result. So does "this RTX 4080 hits 48 TFLOPS" in tech reviews. What does it actually mean, and how should you compare it across hardware?
The simple definition
GFLOPS = Giga-FLoating-point Operations Per Second.
Breaking it down:
- Giga = billion (10⁹). Like gigabyte but for operations.
- FLoating-point = math on real numbers (3.14, -0.5, 1.0e-7) — distinct from integer math (1, 2, 3).
- Operations Per Second = math operations completed in a second. Operations are usually multiply-add (a × b + c) which is fundamental in graphics + ML.
So a GPU rated at 500 GFLOPS can do 500 billion floating-point math operations every second. A laptop CPU at 50 GFLOPS does 50 billion. The difference is 10×, which is why GPUs are so much faster than CPUs at math-heavy work like graphics, machine learning, and physics simulation.
The unit ladder
| Unit | Operations/sec | Typical hardware |
|---|---|---|
| FLOPS | 1 | (too small to be useful) |
| KFLOPS | 1,000 | 1980s pocket calculator |
| MFLOPS | 1,000,000 | 1990s desktop CPU |
| GFLOPS | 10⁹ (1 billion) | 2010s GPUs, modern CPUs |
| TFLOPS | 10¹² (1 trillion) | Modern gaming GPUs (RTX 4080 = 48 TFLOPS) |
| PFLOPS | 10¹⁵ (1 quadrillion) | Top-500 supercomputers |
| EFLOPS | 10¹⁸ (1 quintillion) | Frontier supercomputer (1.7 EFLOPS, 2023) |
1 TFLOPS = 1000 GFLOPS. Manufacturers like to use TFLOPS for marketing because bigger numbers sound better ("our card is 48 TFLOPS!"). Browser benchmarks like 9bench typically report GFLOPS because the numbers fit better in result cards (a 5000 GFLOPS GPU = 5 TFLOPS). Same measurement, different prefix.
What "good" GFLOPS looks like in 2026
These are real-world WebGPU compute benchmark numbers (lower than manufacturer theoretical peaks by 5-15% browser-API overhead):
Integrated GPUs (laptops + entry-level)
- Apple M3 base GPU: 800-1200 GFLOPS
- Apple M4 base GPU: 1100-1500 GFLOPS
- Intel Xe Iris (laptop): 300-700 GFLOPS
- AMD Radeon 780M (laptop): 600-900 GFLOPS
- Older Intel UHD: 100-300 GFLOPS
Mid-range gaming GPUs
- RTX 4060 (laptop): 1800-2400 GFLOPS
- RTX 4060 Ti (desktop): 2500-3500 GFLOPS
- RTX 4070: 4000-5500 GFLOPS
- RX 7800 XT: 3500-4500 GFLOPS
- Apple M3 Pro GPU: 2500-3500 GFLOPS
High-end gaming GPUs
- RTX 4080: 5500-7500 GFLOPS
- RTX 4090: 7500-10000 GFLOPS
- RTX 5090: 9000-12000 GFLOPS
- RX 7900 XTX: 6500-8500 GFLOPS
- Apple M5 Max GPU: 4500-6000 GFLOPS
These are browser WebGPU compute measurements via 9bench's matrix multiplication test. Manufacturer-quoted "TFLOPS" specs are higher because they measure theoretical peak — real workloads (browser or native) achieve 60-85% of theoretical.
Why GFLOPS isn't the only thing that matters
GFLOPS measures compute throughput — how fast the GPU can do math. But real GPU performance also depends on:
1. Memory bandwidth (GB/s)
GPUs need to feed data to compute units. If memory can't deliver fast enough, compute units sit idle. Modern GPUs have 200-1000 GB/s memory bandwidth (Apple M5 Max unified memory: 800 GB/s, RTX 4090: 1008 GB/s, mid-range RTX 4070: 504 GB/s).
A high-GFLOPS GPU with low memory bandwidth gets bottlenecked. Real-world performance is often bandwidth-limited rather than compute-limited, especially for large texture work or big ML models.
2. Special-purpose units
Modern GPUs have dedicated hardware for specific tasks:
- Ray-tracing cores (NVIDIA RT cores, AMD Ray Accelerators): hardware-accelerated ray-triangle intersection for realistic lighting
- Tensor cores / Matrix engines: dedicated mixed-precision math for AI inference
- Video encode/decode (NVENC, AMF, QuickSync): hardware H.264/H.265/AV1
These don't show up in GFLOPS measurements but matter enormously for specific workloads. A GPU with weaker raw GFLOPS but stronger ray-tracing units beats a higher-GFLOPS rival in ray-traced games.
3. Driver + software optimization
Real game frame rates depend on driver maturity, game-engine optimization, and platform-specific quirks. Two GPUs with identical GFLOPS can have 30%+ different frame rates in a specific game based on driver work.
4. Thermal headroom
Sustained workloads (long games, ML inference, video encoding) hit thermal limits. A laptop GPU with 4000 GFLOPS peak might sustain only 2500 GFLOPS after 10 minutes due to thermal throttling. Desktop GPUs with better cooling sustain closer to peak.
How GFLOPS translates to real workloads
Gaming
Roughly: every 1000 GFLOPS adds 10-30 fps in mid-settings 1080p gaming, depending on game. Modern AAA games at 4K Ultra need 5000+ GFLOPS for 60fps. Esports titles at 1080p run on 500-1000 GFLOPS easily.
Machine learning inference
Local LLM inference (Llama 3 8B, Mistral 7B): 1500+ GFLOPS gets you ~10 tokens/sec. Stable Diffusion 1.5 at 512×512: 2000+ GFLOPS for ~30 sec/image. Larger models scale linearly with GFLOPS up to memory limits.
Video editing + encoding
1080p editing: 1000+ GFLOPS handles real-time preview. 4K editing: 3000+ GFLOPS for smooth scrub. H.265 hardware encode: GFLOPS less relevant — dedicated NVENC/QuickSync units do this.
3D rendering (Blender, Cinema 4D)
GPU rendering scales linearly with GFLOPS. A 5000-GFLOPS GPU renders ~3× faster than 1500-GFLOPS. For big scenes, RAM matters too — large scenes need 12GB+ VRAM.
Browser tasks
Most everyday browser use needs <100 GFLOPS. WebGL games (browser-based 3D games), WebGPU apps, and AI in browser (Transformers.js) benefit from more — 1000-2000 GFLOPS is a sweet spot for browser-AI workloads.
Why your GFLOPS might be lower than expected
1. Browser overhead
WebGPU adds 5-15% overhead vs native Metal/Vulkan/DirectX. Expect 9bench to show 85-95% of Geekbench Compute or native benchmarks.
2. Power / battery throttling
Laptop on battery: GPU clocks drop 30-50%. Plug in for accurate measurement.
3. Other GPU work in progress
YouTube playing in another tab, Discord running, screen recording active — all consume GPU. Close them for clean measurement.
4. Browser GPU acceleration disabled
Chrome, Firefox, and Edge have flags that can disable hardware acceleration. Check
chrome://gpu (Chrome) or about:support (Firefox) — should show
"Hardware accelerated" for graphics features.
5. Driver issues
Outdated GPU drivers can drop performance 20-40%. Update via NVIDIA GeForce Experience, AMD Adrenalin, Intel Driver Updater, or your laptop manufacturer's update tool.
Test your GFLOPS in 15 seconds
Run 9bench — it measures your GPU compute throughput via WebGPU matrix multiplication and reports GFLOPS in the result card. Compare to the brackets above to see where your hardware sits.
Want absolute-accurate GFLOPS? Use Geekbench 6's GPU Compute test (Metal/OpenCL/Vulkan native) — it'll be 5-15% higher than 9bench's WebGPU number, which is the true peak your hardware can achieve.
The honest closing
GFLOPS is a useful first-order GPU performance metric. It tells you compute throughput. It does not tell you the full story — memory bandwidth, special-purpose units, drivers, and thermals all matter for real-world workloads.
Best use: compare GFLOPS within the same generation/architecture. RTX 4070 vs RTX 4080: GFLOPS comparison is meaningful. RTX 4070 vs RX 7800 XT: GFLOPS gives a rough idea but architecture differences make comparisons less precise. Always cross-reference with real review benchmarks for the workload you care about.