TL;DR
GFLOPS = Giga-FLoating-point Operations Per Second. 1 GFLOP = 1 billion math ops/sec on real numbers. Modern 2026 hardware: integrated laptop GPUs ~200-800 GFLOPS, mid-range desktop ~3000-5000, high-end ~6000-10000+. Browser benchmarks measure 85-95% of native. More GFLOPS usually = better gaming, ML, video editing — but memory bandwidth + driver optimization matter too. Test your GPU's GFLOPS in 15 seconds.

"590 GFLOPS" appears at the top of your 9bench result. So does "this RTX 4080 hits 48 TFLOPS" in tech reviews. What does it actually mean, and how should you compare it across hardware?

The simple definition

GFLOPS = Giga-FLoating-point Operations Per Second.

Breaking it down:

So a GPU rated at 500 GFLOPS can do 500 billion floating-point math operations every second. A laptop CPU at 50 GFLOPS does 50 billion. The difference is 10×, which is why GPUs are so much faster than CPUs at math-heavy work like graphics, machine learning, and physics simulation.

The unit ladder

UnitOperations/secTypical hardware
FLOPS1(too small to be useful)
KFLOPS1,0001980s pocket calculator
MFLOPS1,000,0001990s desktop CPU
GFLOPS10⁹ (1 billion)2010s GPUs, modern CPUs
TFLOPS10¹² (1 trillion)Modern gaming GPUs (RTX 4080 = 48 TFLOPS)
PFLOPS10¹⁵ (1 quadrillion)Top-500 supercomputers
EFLOPS10¹⁸ (1 quintillion)Frontier supercomputer (1.7 EFLOPS, 2023)

1 TFLOPS = 1000 GFLOPS. Manufacturers like to use TFLOPS for marketing because bigger numbers sound better ("our card is 48 TFLOPS!"). Browser benchmarks like 9bench typically report GFLOPS because the numbers fit better in result cards (a 5000 GFLOPS GPU = 5 TFLOPS). Same measurement, different prefix.

What "good" GFLOPS looks like in 2026

These are real-world WebGPU compute benchmark numbers (lower than manufacturer theoretical peaks by 5-15% browser-API overhead):

Integrated GPUs (laptops + entry-level)

Mid-range gaming GPUs

High-end gaming GPUs

These are browser WebGPU compute measurements via 9bench's matrix multiplication test. Manufacturer-quoted "TFLOPS" specs are higher because they measure theoretical peak — real workloads (browser or native) achieve 60-85% of theoretical.

Why GFLOPS isn't the only thing that matters

GFLOPS measures compute throughput — how fast the GPU can do math. But real GPU performance also depends on:

1. Memory bandwidth (GB/s)

GPUs need to feed data to compute units. If memory can't deliver fast enough, compute units sit idle. Modern GPUs have 200-1000 GB/s memory bandwidth (Apple M5 Max unified memory: 800 GB/s, RTX 4090: 1008 GB/s, mid-range RTX 4070: 504 GB/s).

A high-GFLOPS GPU with low memory bandwidth gets bottlenecked. Real-world performance is often bandwidth-limited rather than compute-limited, especially for large texture work or big ML models.

2. Special-purpose units

Modern GPUs have dedicated hardware for specific tasks:

These don't show up in GFLOPS measurements but matter enormously for specific workloads. A GPU with weaker raw GFLOPS but stronger ray-tracing units beats a higher-GFLOPS rival in ray-traced games.

3. Driver + software optimization

Real game frame rates depend on driver maturity, game-engine optimization, and platform-specific quirks. Two GPUs with identical GFLOPS can have 30%+ different frame rates in a specific game based on driver work.

4. Thermal headroom

Sustained workloads (long games, ML inference, video encoding) hit thermal limits. A laptop GPU with 4000 GFLOPS peak might sustain only 2500 GFLOPS after 10 minutes due to thermal throttling. Desktop GPUs with better cooling sustain closer to peak.

How GFLOPS translates to real workloads

Gaming

Roughly: every 1000 GFLOPS adds 10-30 fps in mid-settings 1080p gaming, depending on game. Modern AAA games at 4K Ultra need 5000+ GFLOPS for 60fps. Esports titles at 1080p run on 500-1000 GFLOPS easily.

Machine learning inference

Local LLM inference (Llama 3 8B, Mistral 7B): 1500+ GFLOPS gets you ~10 tokens/sec. Stable Diffusion 1.5 at 512×512: 2000+ GFLOPS for ~30 sec/image. Larger models scale linearly with GFLOPS up to memory limits.

Video editing + encoding

1080p editing: 1000+ GFLOPS handles real-time preview. 4K editing: 3000+ GFLOPS for smooth scrub. H.265 hardware encode: GFLOPS less relevant — dedicated NVENC/QuickSync units do this.

3D rendering (Blender, Cinema 4D)

GPU rendering scales linearly with GFLOPS. A 5000-GFLOPS GPU renders ~3× faster than 1500-GFLOPS. For big scenes, RAM matters too — large scenes need 12GB+ VRAM.

Browser tasks

Most everyday browser use needs <100 GFLOPS. WebGL games (browser-based 3D games), WebGPU apps, and AI in browser (Transformers.js) benefit from more — 1000-2000 GFLOPS is a sweet spot for browser-AI workloads.

Why your GFLOPS might be lower than expected

1. Browser overhead

WebGPU adds 5-15% overhead vs native Metal/Vulkan/DirectX. Expect 9bench to show 85-95% of Geekbench Compute or native benchmarks.

2. Power / battery throttling

Laptop on battery: GPU clocks drop 30-50%. Plug in for accurate measurement.

3. Other GPU work in progress

YouTube playing in another tab, Discord running, screen recording active — all consume GPU. Close them for clean measurement.

4. Browser GPU acceleration disabled

Chrome, Firefox, and Edge have flags that can disable hardware acceleration. Check chrome://gpu (Chrome) or about:support (Firefox) — should show "Hardware accelerated" for graphics features.

5. Driver issues

Outdated GPU drivers can drop performance 20-40%. Update via NVIDIA GeForce Experience, AMD Adrenalin, Intel Driver Updater, or your laptop manufacturer's update tool.

Test your GFLOPS in 15 seconds

Run 9bench — it measures your GPU compute throughput via WebGPU matrix multiplication and reports GFLOPS in the result card. Compare to the brackets above to see where your hardware sits.

Want absolute-accurate GFLOPS? Use Geekbench 6's GPU Compute test (Metal/OpenCL/Vulkan native) — it'll be 5-15% higher than 9bench's WebGPU number, which is the true peak your hardware can achieve.

The honest closing

GFLOPS is a useful first-order GPU performance metric. It tells you compute throughput. It does not tell you the full story — memory bandwidth, special-purpose units, drivers, and thermals all matter for real-world workloads.

Best use: compare GFLOPS within the same generation/architecture. RTX 4070 vs RTX 4080: GFLOPS comparison is meaningful. RTX 4070 vs RX 7800 XT: GFLOPS gives a rough idea but architecture differences make comparisons less precise. Always cross-reference with real review benchmarks for the workload you care about.

Sources + further reading