(Image credit: Tom’s Hardware)
For anyone navigating the world of PC gaming or demanding graphical applications, the graphics processing unit (GPU) stands as the cornerstone of performance. Whether your passion lies in immersive gaming experiences, powering through artificial intelligence tasks like Stable Diffusion, or executing professional video editing, the GPU’s capabilities are paramount. Even the best CPUs for Gaming, while crucial, often play a supporting role to the graphical prowess of your GPU. Tom’s Hardware provides exhaustive benchmarks for a vast array of GPUs, both current and previous generations, solidifying its place as the go-to resource for GPU comparisons. This article delves into the comprehensive GPU benchmarks and performance hierarchy offered by Tom’s Hardware, ensuring you’re equipped to make informed decisions when choosing your next graphics card.
The GPU landscape is ever-evolving, with recent refreshes from both Nvidia and AMD. Nvidia rolled out the Super series, including the RTX 4070 Super, RTX 4070 Ti Super, and RTX 4080 Super. AMD countered with the RX 7600 XT and the US release of the RX 7900 GRE. As we look ahead, significant architectural shifts are anticipated with the arrival of next-generation GPUs like the Nvidia Blackwell RTX 50-series, Intel Battlemage, and AMD RDNA 4 GPUs, expected to begin surfacing in late 2024 and early 2025.
Tom’s Hardware is also poised to update its GPU testing methodology, incorporating new game titles and a revised testing platform. Following issues encountered with the Core i9-13900K, the AMD Ryzen 7 9800X3D is under consideration as the new testbed CPU. This shift will necessitate a comprehensive re-evaluation of GPUs across the performance spectrum. In the interim, recent GPU reviews on Tom’s Hardware utilize the 13900K platform with an expanded suite of games, with these results reflected in the benchmark charts presented below.
Tom’s Hardware’s GPU hierarchy is segmented into two primary benchmarks: traditional rasterization and ray tracing. The rasterization hierarchy provides a performance ranking based on standard rendering techniques, while the ray tracing hierarchy focuses on GPUs capable of handling demanding ray tracing workloads. Notably, ray tracing benchmarks naturally include only GPUs from AMD’s RX 7000/6000 series, Intel’s Arc series, and Nvidia’s RTX lineup, as these are equipped with the necessary hardware. All benchmark results are derived from native resolutions, without the artificial performance boosts of upscaling technologies like DLSS, FSR, or XeSS, ensuring a direct comparison of raw GPU power.
Nvidia’s current RTX 40-series GPUs are built upon the Ada Lovelace architecture, introducing features such as DLSS 3 Frame Generation and Nvidia DLSS 3.5 Ray Reconstruction. AMD’s RDNA 3 architecture powers the RX 7000 series, while Intel’s Arc Alchemist architecture marks Intel’s entry into the dedicated GPU market.
For historical context, Tom’s Hardware also maintains a 2020–2021 benchmark suite using an older test suite on a Core i9-9900K system. Additionally, a legacy GPU hierarchy, sorted by theoretical performance, is available for reference.
The performance rankings presented in the following sections are based purely on Tom’s Hardware’s gaming benchmarks at 1080p “ultra” settings for the main suite and 1080p “medium” for the DXR (ray tracing) suite. Factors such as price, graphics card power consumption, efficiency, and specific features are not considered in these rankings, which are solely performance-driven. The 2024 benchmarks are conducted on an Alder Lake Core i9-12900K testbed, ensuring consistency and relevance for contemporary hardware assessments.
GPU Benchmarks Ranking 2025: Rasterization Performance
Image 1 of 4
(Image credit: Tom’s Hardware)
Image 2 of 4
(Image credit: Tom’s Hardware)
Image 3 of 4
(Image credit: Tom’s Hardware)
Image 4 of 4
(Image credit: Tom’s Hardware)
Tom’s Hardware’s comprehensive GPU benchmarks encompass nearly every graphics card released in the last seven years, tested at 1080p medium, 1080p ultra, and, where applicable, 1440p ultra and 4K ultra settings. The tables are sorted by 1080p ultra performance, with scores scaled relative to the GeForce RTX 4090, the top performer in their test suite, especially at higher 1440p and 4K resolutions.
The summary chart above provides a visual representation of relative performance across GPU generations at 1080p ultra. You can navigate through the image gallery to view charts for 1080p medium, 1440p, and 4K ultra settings. While some niche cards like the GT 1030, RX 550, and certain Titan models are excluded for clarity, the hierarchy is remarkably comprehensive. Detailed data for older GPUs is also available within the tables below.
The eight games constituting Tom’s Hardware’s standard GPU benchmark suite are Borderlands 3 (DX12), Far Cry 6 (DX12), Flight Simulator (DX11 Nvidia, DX12 AMD/Intel), Forza Horizon 5 (DX12), Horizon Zero Dawn (DX12), Red Dead Redemption 2 (Vulkan), Total War Warhammer 3 (DX11), and Watch Dogs Legion (DX12). The overall fps score is calculated as the geometric mean of frame rates across these eight titles, ensuring a balanced representation of performance. For detailed specifications and individual reviews, refer to the provided links in the ‘Specifications’ column within the tables.
GPU Rasterization Hierarchy: Key Performance Insights
Swipe to scroll horizontally
Graphics Card | Lowest Price | 1080p Ultra | 1080p Medium | 1440p Ultra | 4K Ultra | Specifications (Links to Review) |
---|---|---|---|---|---|---|
GeForce RTX 4090 | $2,529 | 100.0% (154.1fps) | 100.0% (195.7fps) | 100.0% (146.1fps) | 100.0% (114.5fps) | AD102, 16384 shaders, 2520MHz, 24GB GDDR6X@21Gbps, 1008GB/s, 450W |
Radeon RX 7900 XTX | $869 | 96.7% (149.0fps) | 97.2% (190.3fps) | 92.6% (135.3fps) | 83.1% (95.1fps) | Navi 31, 6144 shaders, 2500MHz, 24GB GDDR6@20Gbps, 960GB/s, 355W |
GeForce RTX 4080 Super | No Stock | 96.2% (148.3fps) | 98.5% (192.7fps) | 91.0% (133.0fps) | 80.3% (91.9fps) | AD103, 10240 shaders, 2550MHz, 16GB GDDR6X@23Gbps, 736GB/s, 320W |
GeForce RTX 4080 | $1,699 | 95.4% (147.0fps) | 98.1% (192.0fps) | 89.3% (130.4fps) | 78.0% (89.3fps) | AD103, 9728 shaders, 2505MHz, 16GB [email protected], 717GB/s, 320W |
Radeon RX 7900 XT | $649 | 93.4% (143.9fps) | 95.8% (187.6fps) | 86.1% (125.9fps) | 71.0% (81.2fps) | Navi 31, 5376 shaders, 2400MHz, 20GB GDDR6@20Gbps, 800GB/s, 315W |
GeForce RTX 4070 Ti Super | $899 | 92.3% (142.3fps) | 96.8% (189.4fps) | 83.5% (122.0fps) | 68.7% (78.6fps) | AD103, 8448 shaders, 2610MHz, 16GB GDDR6X@21Gbps, 672GB/s, 285W |
GeForce RTX 4070 Ti | $759 | 89.8% (138.3fps) | 95.7% (187.2fps) | 79.8% (116.5fps) | 63.8% (73.0fps) | AD104, 7680 shaders, 2610MHz, 12GB GDDR6X@21Gbps, 504GB/s, 285W |
Radeon RX 7900 GRE | No Stock | 88.1% (135.8fps) | 94.1% (184.3fps) | 78.0% (113.9fps) | 60.5% (69.3fps) | Navi 31, 5120 shaders, 2245MHz, 16GB GDDR6@18Gbps, 576GB/s, 260W |
GeForce RTX 4070 Super | $609 | 87.1% (134.2fps) | 94.6% (185.1fps) | 75.2% (109.8fps) | 57.8% (66.1fps) | AD104, 7168 shaders, 2475MHz, 12GB GDDR6X@21Gbps, 504GB/s, 220W |
Radeon RX 6950 XT | $859 | 84.7% (130.5fps) | 91.7% (179.4fps) | 75.3% (110.1fps) | 58.6% (67.1fps) | Navi 21, 5120 shaders, 2310MHz, 16GB GDDR6@18Gbps, 576GB/s, 335W |
GeForce RTX 3090 Ti | $1,899 | 84.7% (130.5fps) | 90.5% (177.1fps) | 77.1% (112.7fps) | 66.3% (75.9fps) | GA102, 10752 shaders, 1860MHz, 24GB GDDR6X@21Gbps, 1008GB/s, 450W |
Radeon RX 7800 XT | $489 | 83.9% (129.3fps) | 91.5% (179.1fps) | 72.4% (105.8fps) | 54.4% (62.3fps) | Navi 32, 3840 shaders, 2430MHz, 16GB [email protected], 624GB/s, 263W |
GeForce RTX 3090 | $1,530 | 81.4% (125.5fps) | 88.9% (174.0fps) | 72.5% (106.0fps) | 61.8% (70.7fps) | GA102, 10496 shaders, 1695MHz, 24GB [email protected], 936GB/s, 350W |
Radeon RX 6900 XT | $810 | 80.9% (124.6fps) | 89.6% (175.3fps) | 69.9% (102.1fps) | 53.5% (61.2fps) | Navi 21, 5120 shaders, 2250MHz, 16GB GDDR6@16Gbps, 512GB/s, 300W |
GeForce RTX 3080 Ti | $979 | 80.4% (123.9fps) | 87.8% (171.8fps) | 71.1% (103.9fps) | 60.1% (68.8fps) | GA102, 10240 shaders, 1665MHz, 12GB GDDR6X@19Gbps, 912GB/s, 350W |
Radeon RX 6800 XT | $1,150 | 79.6% (122.7fps) | 88.5% (173.2fps) | 67.8% (99.0fps) | 50.6% (57.9fps) | Navi 21, 4608 shaders, 2250MHz, 16GB GDDR6@16Gbps, 512GB/s, 300W |
GeForce RTX 3080 12GB | $829 | 79.2% (122.1fps) | 86.5% (169.4fps) | 70.0% (102.3fps) | 58.3% (66.7fps) | GA102, 8960 shaders, 1845MHz, 12GB GDDR6X@19Gbps, 912GB/s, 400W |
GeForce RTX 4070 | $549 | 79.2% (122.0fps) | 90.7% (177.5fps) | 66.9% (97.8fps) | 50.0% (57.2fps) | AD104, 5888 shaders, 2475MHz, 12GB GDDR6X@21Gbps, 504GB/s, 200W |
GeForce RTX 3080 | $788 | 76.0% (117.0fps) | 85.6% (167.6fps) | 66.0% (96.4fps) | 54.1% (62.0fps) | GA102, 8704 shaders, 1710MHz, 10GB GDDR6X@19Gbps, 760GB/s, 320W |
Radeon RX 7700 XT | $409 | 75.3% (116.1fps) | 87.7% (171.6fps) | 63.4% (92.7fps) | 45.0% (51.5fps) | Navi 32, 3456 shaders, 2544MHz, 12GB GDDR6@18Gbps, 432GB/s, 245W |
Radeon RX 6800 | $849 | 74.4% (114.6fps) | 86.2% (168.7fps) | 61.0% (89.2fps) | 44.3% (50.7fps) | Navi 21, 3840 shaders, 2105MHz, 16GB GDDR6@16Gbps, 512GB/s, 250W |
GeForce RTX 3070 Ti | $699 | 67.5% (104.0fps) | 81.6% (159.8fps) | 56.7% (82.8fps) | 41.7% (47.7fps) | GA104, 6144 shaders, 1770MHz, 8GB GDDR6X@19Gbps, 608GB/s, 290W |
Radeon RX 6750 XT | $354 | 66.8% (102.9fps) | 82.6% (161.6fps) | 52.9% (77.2fps) | 37.4% (42.8fps) | Navi 22, 2560 shaders, 2600MHz, 12GB GDDR6@18Gbps, 432GB/s, 250W |
GeForce RTX 4060 Ti 16GB | $634 | 65.3% (100.6fps) | 82.6% (161.7fps) | 51.8% (75.7fps) | 36.4% (41.6fps) | AD106, 4352 shaders, 2535MHz, 16GB GDDR6@18Gbps, 288GB/s, 160W |
GeForce RTX 4060 Ti | $399 | 65.1% (100.4fps) | 81.8% (160.1fps) | 51.7% (75.6fps) | 34.6% (39.6fps) | AD106, 4352 shaders, 2535MHz, 8GB GDDR6@18Gbps, 288GB/s, 160W |
Titan RTX | Row 25 – Cell 1 | 64.5% (99.3fps) | 80.0% (156.6fps) | 54.4% (79.5fps) | 41.8% (47.8fps) | TU102, 4608 shaders, 1770MHz, 24GB GDDR6@14Gbps, 672GB/s, 280W |
Radeon RX 6700 XT | $499 | 64.3% (99.1fps) | 80.8% (158.1fps) | 50.3% (73.4fps) | 35.3% (40.4fps) | Navi 22, 2560 shaders, 2581MHz, 12GB GDDR6@16Gbps, 384GB/s, 230W |
GeForce RTX 3070 | $495 | 64.1% (98.8fps) | 79.1% (154.8fps) | 53.2% (77.7fps) | 38.8% (44.4fps) | GA104, 5888 shaders, 1725MHz, 8GB GDDR6@14Gbps, 448GB/s, 220W |
GeForce RTX 2080 Ti | Row 28 – Cell 1 | 62.5% (96.3fps) | 77.2% (151.0fps) | 51.8% (75.6fps) | 38.0% (43.5fps) | TU102, 4352 shaders, 1545MHz, 11GB GDDR6@14Gbps, 616GB/s, 250W |
Radeon RX 7600 XT | $314 | 59.7% (91.9fps) | 77.3% (151.2fps) | 45.1% (65.9fps) | 32.4% (37.1fps) | Navi 33, 2048 shaders, 2755MHz, 16GB GDDR6@18Gbps, 288GB/s, 190W |
GeForce RTX 3060 Ti | $498 | 58.9% (90.7fps) | 75.0% (146.9fps) | 47.9% (70.0fps) | Row 30 – Cell 5 | GA104, 4864 shaders, 1665MHz, 8GB GDDR6@14Gbps, 448GB/s, 200W |
Radeon RX 6700 10GB | No Stock | 55.9% (86.1fps) | 74.4% (145.7fps) | 43.0% (62.8fps) | 28.7% (32.9fps) | Navi 22, 2304 shaders, 2450MHz, 10GB GDDR6@16Gbps, 320GB/s, 175W |
GeForce RTX 2080 Super | Row 32 – Cell 1 | 55.8% (86.0fps) | 72.2% (141.3fps) | 45.2% (66.1fps) | 32.1% (36.7fps) | TU104, 3072 shaders, 1815MHz, 8GB [email protected], 496GB/s, 250W |
GeForce RTX 4060 | $294 | 55.1% (84.9fps) | 72.7% (142.3fps) | 41.9% (61.2fps) | 27.8% (31.9fps) | AD107, 3072 shaders, 2460MHz, 8GB GDDR6@17Gbps, 272GB/s, 115W |
GeForce RTX 2080 | Row 34 – Cell 1 | 53.5% (82.5fps) | 69.8% (136.7fps) | 43.2% (63.2fps) | Row 34 – Cell 5 | TU104, 2944 shaders, 1710MHz, 8GB GDDR6@14Gbps, 448GB/s, 215W |
Radeon RX 7600 | $259 | 53.2% (82.0fps) | 72.3% (141.4fps) | 39.2% (57.3fps) | 25.4% (29.1fps) | Navi 33, 2048 shaders, 2655MHz, 8GB GDDR6@18Gbps, 288GB/s, 165W |
Radeon RX 6650 XT | $254 | 50.4% (77.7fps) | 70.0% (137.1fps) | 37.3% (54.5fps) | Row 36 – Cell 5 | Navi 23, 2048 shaders, 2635MHz, 8GB GDDR6@18Gbps, 280GB/s, 180W |
GeForce RTX 2070 Super | Row 37 – Cell 1 | 50.3% (77.4fps) | 66.2% (129.6fps) | 40.0% (58.4fps) | Row 37 – Cell 5 | TU104, 2560 shaders, 1770MHz, 8GB GDDR6@14Gbps, 448GB/s, 215W |
Intel Arc A770 16GB | $299 | 49.9% (76.9fps) | 59.4% (116.4fps) | 41.0% (59.8fps) | 30.8% (35.3fps) | ACM-G10, 4096 shaders, 2400MHz, 16GB [email protected], 560GB/s, 225W |
Intel Arc A770 8GB | No Stock | 48.9% (75.3fps) | 59.0% (115.5fps) | 39.3% (57.5fps) | 29.0% (33.2fps) | ACM-G10, 4096 shaders, 2400MHz, 8GB GDDR6@16Gbps, 512GB/s, 225W |
Radeon RX 6600 XT | $259 | 48.5% (74.7fps) | 68.2% (133.5fps) | 35.7% (52.2fps) | Row 40 – Cell 5 | Navi 23, 2048 shaders, 2589MHz, 8GB GDDR6@16Gbps, 256GB/s, 160W |
Radeon RX 5700 XT | Row 41 – Cell 1 | 47.6% (73.3fps) | 63.8% (124.9fps) | 36.3% (53.1fps) | 25.6% (29.3fps) | Navi 10, 2560 shaders, 1905MHz, 8GB GDDR6@14Gbps, 448GB/s, 225W |
GeForce RTX 3060 | Row 42 – Cell 1 | 46.9% (72.3fps) | 61.8% (121.0fps) | 36.9% (54.0fps) | Row 42 – Cell 5 | GA106, 3584 shaders, 1777MHz, 12GB GDDR6@15Gbps, 360GB/s, 170W |
Intel Arc A750 | $239 | 45.9% (70.8fps) | 56.4% (110.4fps) | 36.7% (53.7fps) | 27.2% (31.1fps) | ACM-G10, 3584 shaders, 2350MHz, 8GB GDDR6@16Gbps, 512GB/s, 225W |
GeForce RTX 2070 | Row 44 – Cell 1 | 45.3% (69.8fps) | 60.8% (119.1fps) | 35.5% (51.8fps) | Row 44 – Cell 5 | TU106, 2304 shaders, 1620MHz, 8GB GDDR6@14Gbps, 448GB/s, 175W |
Radeon VII | Row 45 – Cell 1 | 45.1% (69.5fps) | 58.2% (113.9fps) | 36.3% (53.0fps) | 27.5% (31.5fps) | Vega 20, 3840 shaders, 1750MHz, 16GB [email protected], 1024GB/s, 300W |
GeForce GTX 1080 Ti | Row 46 – Cell 1 | 43.1% (66.4fps) | 56.3% (110.2fps) | 34.4% (50.2fps) | 25.8% (29.5fps) | GP102, 3584 shaders, 1582MHz, 11GB GDDR5X@11Gbps, 484GB/s, 250W |
GeForce RTX 2060 Super | Row 47 – Cell 1 | 42.5% (65.5fps) | 57.2% (112.0fps) | 33.1% (48.3fps) | Row 47 – Cell 5 | TU106, 2176 shaders, 1650MHz, 8GB GDDR6@14Gbps, 448GB/s, 175W |
Radeon RX 6600 | $189 | 42.3% (65.2fps) | 59.3% (116.2fps) | 30.6% (44.8fps) | Row 48 – Cell 5 | Navi 23, 1792 shaders, 2491MHz, 8GB GDDR6@14Gbps, 224GB/s, 132W |
Intel Arc A580 | $169 | 42.3% (65.1fps) | 51.6% (101.1fps) | 33.4% (48.8fps) | 24.4% (27.9fps) | ACM-G10, 3072 shaders, 2300MHz, 8GB GDDR6@16Gbps, 512GB/s, 185W |
Radeon RX 5700 | Row 50 – Cell 1 | 41.9% (64.5fps) | 56.6% (110.8fps) | 31.9% (46.7fps) | Row 50 – Cell 5 | Navi 10, 2304 shaders, 1725MHz, 8GB GDDR6@14Gbps, 448GB/s, 180W |
Radeon RX 5600 XT | Row 51 – Cell 1 | 37.5% (57.8fps) | 51.1% (100.0fps) | 28.8% (42.0fps) | Row 51 – Cell 5 | Navi 10, 2304 shaders, 1750MHz, 8GB GDDR6@14Gbps, 336GB/s, 160W |
Radeon RX Vega 64 | Row 52 – Cell 1 | 36.8% (56.7fps) | 48.2% (94.3fps) | 28.5% (41.6fps) | 20.5% (23.5fps) | Vega 10, 4096 shaders, 1546MHz, 8GB [email protected], 484GB/s, 295W |
GeForce RTX 2060 | Row 53 – Cell 1 | 36.0% (55.5fps) | 51.4% (100.5fps) | 27.5% (40.1fps) | Row 53 – Cell 5 | TU106, 1920 shaders, 1680MHz, 6GB GDDR6@14Gbps, 336GB/s, 160W |
GeForce GTX 1080 | Row 54 – Cell 1 | 34.4% (53.0fps) | 45.9% (89.9fps) | 27.0% (39.4fps) | Row 54 – Cell 5 | GP104, 2560 shaders, 1733MHz, 8GB GDDR5X@10Gbps, 320GB/s, 180W |
GeForce RTX 3050 | $169 | 33.7% (51.9fps) | 45.4% (88.8fps) | 26.4% (38.5fps) | Row 55 – Cell 5 | GA106, 2560 shaders, 1777MHz, 8GB GDDR6@14Gbps, 224GB/s, 130W |
GeForce GTX 1070 Ti | Row 56 – Cell 1 | 33.1% (51.1fps) | 43.8% (85.7fps) | 26.0% (37.9fps) | Row 56 – Cell 5 | GP104, 2432 shaders, 1683MHz, 8GB GDDR5@8Gbps, 256GB/s, 180W |
Radeon RX Vega 56 | Row 57 – Cell 1 | 32.8% (50.6fps) | 43.0% (84.2fps) | 25.3% (37.0fps) | Row 57 – Cell 5 | Vega 10, 3584 shaders, 1471MHz, 8GB [email protected], 410GB/s, 210W |
GeForce GTX 1660 Super | Row 58 – Cell 1 | 30.3% (46.8fps) | 43.7% (85.5fps) | 22.8% (33.3fps) | Row 58 – Cell 5 | TU116, 1408 shaders, 1785MHz, 6GB GDDR6@14Gbps, 336GB/s, 125W |
GeForce GTX 1660 Ti | Row 59 – Cell 1 | 30.3% (46.6fps) | 43.3% (84.8fps) | 22.8% (33.3fps) | Row 59 – Cell 5 | TU116, 1536 shaders, 1770MHz, 6GB GDDR6@12Gbps, 288GB/s, 120W |
GeForce GTX 1070 | Row 60 – Cell 1 | 29.0% (44.7fps) | 38.3% (75.0fps) | 22.7% (33.1fps) | Row 60 – Cell 5 | GP104, 1920 shaders, 1683MHz, 8GB GDDR5@8Gbps, 256GB/s, 150W |
GeForce GTX 1660 | Row 61 – Cell 1 | 27.7% (42.6fps) | 39.7% (77.8fps) | 20.8% (30.3fps) | Row 61 – Cell 5 | TU116, 1408 shaders, 1785MHz, 6GB GDDR5@8Gbps, 192GB/s, 120W |
Radeon RX 5500 XT 8GB | Row 62 – Cell 1 | 25.7% (39.7fps) | 36.8% (72.1fps) | 19.3% (28.2fps) | Row 62 – Cell 5 | Navi 14, 1408 shaders, 1845MHz, 8GB GDDR6@14Gbps, 224GB/s, 130W |
Radeon RX 590 | Row 63 – Cell 1 | 25.5% (39.3fps) | 35.0% (68.5fps) | 19.9% (29.0fps) | Row 63 – Cell 5 | Polaris 30, 2304 shaders, 1545MHz, 8GB GDDR5@8Gbps, 256GB/s, 225W |
GeForce GTX 980 Ti | Row 64 – Cell 1 | 23.3% (35.9fps) | 32.0% (62.6fps) | 18.2% (26.6fps) | Row 64 – Cell 5 | GM200, 2816 shaders, 1075MHz, 6GB GDDR5@7Gbps, 336GB/s, 250W |
Radeon RX 580 8GB | Row 65 – Cell 1 | 22.9% (35.3fps) | 31.5% (61.7fps) | 17.8% (26.0fps) | Row 65 – Cell 5 | Polaris 20, 2304 shaders, 1340MHz, 8GB GDDR5@8Gbps, 256GB/s, 185W |
Radeon R9 Fury X | Row 66 – Cell 1 | 22.9% (35.2fps) | 32.6% (63.8fps) | Row 66 – Cell 4 | Row 66 – Cell 5 | Fiji, 4096 shaders, 1050MHz, 4GB HBM2@2Gbps, 512GB/s, 275W |
GeForce GTX 1650 Super | Row 67 – Cell 1 | 22.0% (33.9fps) | 34.6% (67.7fps) | 14.5% (21.2fps) | Row 67 – Cell 5 | TU116, 1280 shaders, 1725MHz, 4GB GDDR6@12Gbps, 192GB/s, 100W |
Radeon RX 5500 XT 4GB | Row 68 – Cell 1 | 21.6% (33.3fps) | 34.1% (66.8fps) | Row 68 – Cell 4 | Row 68 – Cell 5 | Navi 14, 1408 shaders, 1845MHz, 4GB GDDR6@14Gbps, 224GB/s, 130W |
GeForce GTX 1060 6GB | Row 69 – Cell 1 | 20.8% (32.1fps) | 29.5% (57.7fps) | 15.8% (23.0fps) | Row 69 – Cell 5 | GP106, 1280 shaders, 1708MHz, 6GB GDDR5@8Gbps, 192GB/s, 120W |
Radeon RX 6500 XT | $232 | 19.9% (30.6fps) | 33.6% (65.8fps) | 12.3% (18.0fps) | Row 70 – Cell 5 | Navi 24, 1024 shaders, 2815MHz, 4GB GDDR6@18Gbps, 144GB/s, 107W |
Radeon R9 390 | Row 71 – Cell 1 | 19.3% (29.8fps) | 26.1% (51.1fps) | Row 71 – Cell 4 | Row 71 – Cell 5 | Grenada, 2560 shaders, 1000MHz, 8GB GDDR5@6Gbps, 384GB/s, 275W |
GeForce GTX 980 | Row 72 – Cell 1 | 18.7% (28.9fps) | 27.4% (53.6fps) | Row 72 – Cell 4 | Row 72 – Cell 5 | GM204, 2048 shaders, 1216MHz, 4GB GDDR5@7Gbps, 256GB/s, 165W |
GeForce GTX 1650 GDDR6 | Row 73 – Cell 1 | 18.7% (28.8fps) | 28.9% (56.6fps) | Row 73 – Cell 4 | Row 73 – Cell 5 | TU117, 896 shaders, 1590MHz, 4GB GDDR6@12Gbps, 192GB/s, 75W |
Intel Arc A380 | $119 | 18.4% (28.4fps) | 27.7% (54.3fps) | 13.3% (19.5fps) | Row 74 – Cell 5 | ACM-G11, 1024 shaders, 2450MHz, 6GB [email protected], 186GB/s, 75W |
Radeon RX 570 4GB | Row 75 – Cell 1 | 18.2% (28.1fps) | 27.4% (53.6fps) | 13.6% (19.9fps) | Row 75 – Cell 5 | Polaris 20, 2048 shaders, 1244MHz, 4GB GDDR5@7Gbps, 224GB/s, 150W |
GeForce GTX 1650 | Row 76 – Cell 1 | 17.5% (27.0fps) | 26.2% (51.3fps) | Row 76 – Cell 4 | Row 76 – Cell 5 | TU117, 896 shaders, 1665MHz, 4GB GDDR5@8Gbps, 128GB/s, 75W |
GeForce GTX 970 | Row 77 – Cell 1 | 17.2% (26.5fps) | 25.0% (49.0fps) | Row 77 – Cell 4 | Row 77 – Cell 5 | GM204, 1664 shaders, 1178MHz, 4GB GDDR5@7Gbps, 256GB/s, 145W |
Radeon RX 6400 | $209 | 15.7% (24.1fps) | 26.1% (51.1fps) | Row 78 – Cell 4 | Row 78 – Cell 5 | Navi 24, 768 shaders, 2321MHz, 4GB GDDR6@16Gbps, 128GB/s, 53W |
GeForce GTX 1050 Ti | Row 79 – Cell 1 | 12.9% (19.8fps) | 19.4% (38.0fps) | Row 79 – Cell 4 | Row 79 – Cell 5 | GP107, 768 shaders, 1392MHz, 4GB GDDR5@7Gbps, 112GB/s, 75W |
GeForce GTX 1060 3GB | Row 80 – Cell 1 | Row 80 – Cell 2 | 26.8% (52.5fps) | Row 80 – Cell 4 | Row 80 – Cell 5 | GP106, 1152 shaders, 1708MHz, 3GB GDDR5@8Gbps, 192GB/s, 120W |
GeForce GTX 1630 | Row 81 – Cell 1 | 10.9% (16.9fps) | 17.3% (33.8fps) | Row 81 – Cell 4 | Row 81 – Cell 5 | TU117, 512 shaders, 1785MHz, 4GB GDDR6@12Gbps, 96GB/s, 75W |
Radeon RX 560 4GB | Row 82 – Cell 1 | 9.6% (14.7fps) | 16.2% (31.7fps) | Row 82 – Cell 4 | Row 82 – Cell 5 | Baffin, 1024 shaders, 1275MHz, 4GB GDDR5@7Gbps, 112GB/s, 60-80W |
GeForce GTX 1050 | Row 83 – Cell 1 | Row 83 – Cell 2 | 15.2% (29.7fps) | Row 83 – Cell 4 | Row 83 – Cell 5 | GP107, 640 shaders, 1455MHz, 2GB GDDR5@7Gbps, 112GB/s, 75W |
Radeon RX 550 4GB | Row 84 – Cell 1 | Row 84 – Cell 2 | 10.0% (19.5fps) | Row 84 – Cell 4 | Row 84 – Cell 5 | Lexa, 640 shaders, 1183MHz, 4GB GDDR5@7Gbps, 112GB/s, 50W |
GeForce GT 1030 | Row 85 – Cell 1 | Row 85 – Cell 2 | 7.5% (14.6fps) | Row 85 – Cell 4 | Row 85 – Cell 5 | GP108, 384 shaders, 1468MHz, 2GB GDDR5@6Gbps, 48GB/s, 30W |





*: GPU couldn’t run all tests, so the overall score is slightly skewed at 1080p ultra.
While the RTX 4090, as Tom’s Hardware benchmarks reveal, leads the pack at 1080p ultra, its dominance truly emerges at 1440p and 4K resolutions. At 1080p ultra, it registers a marginal <2% performance increase over the RTX 4080 Super. However, this gap widens significantly to 9% at 1440p and an impressive 25% at 4K. It’s important to note that Tom’s Hardware’s fps metrics are a composite score, weighting average fps more heavily than 1% low fps, providing a balanced performance indicator.
The rasterization benchmark hierarchy from Tom’s Hardware intentionally excludes ray tracing and DLSS results to maintain a level playing field across both current and older generation GPUs. Since DLSS is an Nvidia-specific technology (and DLSS 3 is exclusive to RTX 40-series), its inclusion would skew direct comparisons. For those interested in DLSS performance, Tom’s Hardware’s RTX 4070 review includes DLSS 2/3 and FSR 2 upscaling benchmarks, demonstrating the performance uplift these technologies can offer.
Despite its price point, the RTX 4090, as highlighted in Tom’s Hardware’s analysis, presents a compelling value proposition, especially when compared to its predecessor, the RTX 3090. While the 3090 offered only incremental performance gains over the 3080 at launch, the 4090 represents a more substantial leap, driven by increased core counts, clock speeds, and power limits. However, Tom’s Hardware also points out challenges associated with the RTX 4090, including limited availability at MSRP due to AI sector demand and concerns regarding 450W power draw through the 16-pin connector.
Stepping down from the RTX 4090, Tom’s Hardware’s benchmarks show a performance trade-off between the RTX 4080 Super and RX 7900 XTX at higher resolutions, with CPU bottlenecks becoming more pronounced at 1080p. As mentioned earlier, Tom’s Hardware is transitioning to a new testbed in the near future, with current results from the 13900K testing platform available in the charts at the end of this article.
(Image credit: Intel)
Beyond the latest offerings from AMD and Nvidia, Tom’s Hardware notes that the RX 6000- and RTX 30-series GPUs remain viable options, especially for users not immediately seeking an upgrade. Intel’s Arc GPUs, as tested by Tom’s Hardware, are also noteworthy, representing a third player in the GPU market.
Tom’s Hardware’s continuous testing and driver updates have addressed initial anomalies with Arc GPUs, such as issues with Minecraft. While Arc GPUs may not lead in efficiency, the A750, in particular, offers a compelling balance of performance and price, according to Tom’s Hardware’s reviews.
Looking at previous generations, Tom’s Hardware’s hierarchy positions the RTX 20-series and GTX 16-series cards, alongside the RX 5000-series, across the performance spectrum. A general trend observed by Tom’s Hardware is that newer architectures offer a performance uplift equivalent to one or two “model upgrades.” For instance, the RTX 2080 Super closely trails the RTX 3060 Ti, while the RX 5700 XT performs similarly to the newer, more budget-friendly RX 6600 XT.
Tom’s Hardware’s benchmarks underscore the increasing VRAM demands of modern games, particularly at ultra settings. Cards with 4GB VRAM or less struggle significantly, highlighting that 8GB VRAM is now a minimum recommendation, with 12GB+ preferred for mainstream GPUs and 16GB+ for high-end and enthusiast cards. Older GPUs like the GTX 1060 3GB and GTX 1050 encountered test failures, skewing their overall scores, despite potentially performing better at 1080p medium settings.
Ray Tracing GPU Benchmarks Ranking 2025
(Image credit: Techland)
Enabling ray tracing, especially in demanding games within Tom’s Hardware’s DXR test suite, can drastically reduce frame rates. Tom’s Hardware tests ray tracing performance at “medium” and “ultra” settings. “Medium” typically entails using medium graphics presets with ray tracing effects enabled, while “ultra” maximizes all ray tracing options for highest visual fidelity.
Due to the performance-intensive nature of ray tracing, Tom’s Hardware sorts these results by 1080p medium scores. Entry-level ray tracing capable GPUs like the RX 6500 XT, RX 6400, and Arc A380 struggle even at these settings, making higher resolutions impractical for comparative benchmarking at this level.
Tom’s Hardware’s ray tracing benchmark suite comprises five DX12 / DX12 Ultimate games: Bright Memory Infinite, Control Ultimate Edition, Cyberpunk 2077, Metro Exodus Enhanced, and Minecraft. The fps score is the geometric mean across these five titles, scaled relative to the GeForce RTX 4090, the top performer in ray tracing as well, according to Tom’s Hardware’s tests.
For a glimpse into the future of ray tracing, Tom’s Hardware recommends exploring their Alan Wake 2 benchmarks, where path tracing pushes even high-end GPUs to their limits, even with upscaling. However, Tom’s Hardware emphasizes that games where ray tracing significantly enhances visuals remain limited, suggesting rasterization remains a more practical rendering approach for most titles.
Image 1 of 4
(Image credit: Tom’s Hardware)
Image 2 of 4
(Image credit: Tom’s Hardware)
Image 3 of 4
(Image credit: Tom’s Hardware)
Image 4 of 4
(Image credit: Tom’s Hardware)
GPU Ray Tracing Hierarchy: Key Performance Takeaways
Swipe to scroll horizontally
Graphics Card | Lowest Price | 1080p Medium | 1080p Ultra | 1440p Ultra | 4K Ultra | Specifications (Links to Review) |
---|---|---|---|---|---|---|
GeForce RTX 4090 | $2,643 | 100.0% (165.9fps) | 100.0% (136.3fps) | 100.0% (103.9fps) | 100.0% (55.9fps) | AD102, 16384 shaders, 2520MHz, 24GB GDDR6X@21Gbps, 1008GB/s, 450W |
GeForce RTX 4080 Super | No Stock | 86.8% (144.0fps) | 85.3% (116.3fps) | 75.6% (78.6fps) | 70.5% (39.4fps) | AD103, 10240 shaders, 2550MHz, 16GB GDDR6X@23Gbps, 736GB/s, 320W |
GeForce RTX 4080 | $1,725 | 85.4% (141.6fps) | 83.4% (113.6fps) | 73.1% (76.0fps) | 67.7% (37.8fps) | AD103, 9728 shaders, 2505MHz, 16GB [email protected], 717GB/s, 320W |
GeForce RTX 4070 Ti Super | $819 | 77.3% (128.2fps) | 73.5% (100.3fps) | 63.5% (66.0fps) | 58.4% (32.6fps) | AD103, 8448 shaders, 2610MHz, 16GB GDDR6X@21Gbps, 672GB/s, 285W |
GeForce RTX 3090 Ti | $1,899 | 71.9% (119.3fps) | 68.4% (93.2fps) | 59.6% (62.0fps) | 56.9% (31.8fps) | GA102, 10752 shaders, 1860MHz, 24GB GDDR6X@21Gbps, 1008GB/s, 450W |
GeForce RTX 4070 Ti | $739 | 71.5% (118.6fps) | 67.1% (91.6fps) | 56.9% (59.1fps) | 52.3% (29.2fps) | AD104, 7680 shaders, 2610MHz, 12GB GDDR6X@21Gbps, 504GB/s, 285W |
GeForce RTX 4070 Super | $609 | 68.1% (113.0fps) | 62.7% (85.6fps) | 52.4% (54.5fps) | 47.8% (26.7fps) | AD104, 7168 shaders, 2475MHz, 12GB GDDR6X@21Gbps, 504GB/s, 220W |
GeForce RTX 3090 | $1,389 | 67.7% (112.4fps) | 63.5% (86.6fps) | 55.1% (57.2fps) | 51.8% (28.9fps) | GA102, 10496 shaders, 1695MHz, 24GB [email protected], 936GB/s, 350W |
GeForce RTX 3080 Ti | $979 | 66.5% (110.4fps) | 62.2% (84.8fps) | 53.2% (55.3fps) | 48.6% (27.1fps) | GA102, 10240 shaders, 1665MHz, 12GB GDDR6X@19Gbps, 912GB/s, 350W |
Radeon RX 7900 XTX | $869 | 66.1% (109.6fps) | 61.7% (84.1fps) | 53.2% (55.3fps) | 48.6% (27.2fps) | Navi 31, 6144 shaders, 2500MHz, 24GB GDDR6@20Gbps, 960GB/s, 355W |
GeForce RTX 3080 12GB | $829 | 64.9% (107.6fps) | 59.9% (81.7fps) | 50.8% (52.8fps) | 46.3% (25.8fps) | GA102, 8960 shaders, 1845MHz, 12GB GDDR6X@19Gbps, 912GB/s, 400W |
GeForce RTX 4070 | $519 | 61.2% (101.4fps) | 54.2% (73.9fps) | 45.1% (46.9fps) | 40.7% (22.7fps) | AD104, 5888 shaders, 2475MHz, 12GB GDDR6X@21Gbps, 504GB/s, 200W |
Radeon RX 7900 XT | $689 | 60.4% (100.3fps) | 55.3% (75.3fps) | 46.7% (48.5fps) | 41.6% (23.3fps) | Navi 31, 5376 shaders, 2400MHz, 20GB GDDR6@20Gbps, 800GB/s, 315W |
GeForce RTX 3080 | $829 | 60.2% (99.8fps) | 54.5% (74.3fps) | 46.1% (47.9fps) | 41.8% (23.3fps) | GA102, 8704 shaders, 1710MHz, 10GB GDDR6X@19Gbps, 760GB/s, 320W |
Radeon RX 7900 GRE | No Stock | 52.9% (87.7fps) | 46.8% (63.7fps) | 39.6% (41.2fps) | 35.7% (19.9fps) | Navi 31, 5120 shaders, 2245MHz, 16GB GDDR6@18Gbps, 576GB/s, 260W |
GeForce RTX 3070 Ti | $499 | 50.6% (84.0fps) | 43.0% (58.6fps) | 35.7% (37.1fps) | Row 15 – Cell 5 | GA104, 6144 shaders, 1770MHz, 8GB GDDR6X@19Gbps, 608GB/s, 290W |
Radeon RX 6950 XT | $1,199 | 48.3% (80.1fps) | 41.4% (56.4fps) | 34.3% (35.7fps) | 31.0% (17.3fps) | Navi 21, 5120 shaders, 2310MHz, 16GB GDDR6@18Gbps, 576GB/s, 335W |
GeForce RTX 3070 | $399 | 47.2% (78.2fps) | 39.9% (54.4fps) | 32.8% (34.1fps) | Row 17 – Cell 5 | GA104, 5888 shaders, 1725MHz, 8GB GDDR6@14Gbps, 448GB/s, 220W |
Radeon RX 7800 XT | $489 | 46.7% (77.5fps) | 41.9% (57.1fps) | 34.9% (36.3fps) | 31.0% (17.3fps) | Navi 32, 3840 shaders, 2430MHz, 16GB [email protected], 624GB/s, 263W |
Radeon RX 6900 XT | $811 | 45.4% (75.4fps) | 38.3% (52.3fps) | 32.1% (33.3fps) | 28.8% (16.1fps) | Navi 21, 5120 shaders, 2250MHz, 16GB GDDR6@16Gbps, 512GB/s, 300W |
GeForce RTX 4060 Ti | $399 | 45.2% (75.1fps) | 38.7% (52.8fps) | 32.3% (33.5fps) | 24.8% (13.9fps) | AD106, 4352 shaders, 2535MHz, 8GB GDDR6@18Gbps, 288GB/s, 160W |
GeForce RTX 4060 Ti 16GB | $449 | 45.2% (75.0fps) | 38.8% (53.0fps) | 32.7% (34.0fps) | 29.5% (16.5fps) | AD106, 4352 shaders, 2535MHz, 16GB GDDR6@18Gbps, 288GB/s, 160W |
Titan RTX | Row 22 – Cell 1 | 44.8% (74.4fps) | 39.1% (53.3fps) | 33.7% (35.0fps) | 31.2% (17.4fps) | TU102, 4608 shaders, 1770MHz, 24GB GDDR6@14Gbps, 672GB/s, 280W |
GeForce RTX 2080 Ti | Row 23 – Cell 1 | 42.7% (70.9fps) | 37.2% (50.7fps) | 31.6% (32.9fps) | Row 23 – Cell 5 | TU102, 4352 shaders, 1545MHz, 11GB GDDR6@14Gbps, 616GB/s, 250W |
Radeon RX 6800 XT | $1,099 | 42.2% (70.0fps) | 35.6% (48.5fps) | 29.9% (31.1fps) | 26.8% (15.0fps) | Navi 21, 4608 shaders, 2250MHz, 16GB GDDR6@16Gbps, 512GB/s, 300W |
GeForce RTX 3060 Ti | $453 | 41.9% (69.5fps) | 35.0% (47.7fps) | 28.8% (30.0fps) | Row 25 – Cell 5 | GA104, 4864 shaders, 1665MHz, 8GB GDDR6@14Gbps, 448GB/s, 200W |
Radeon RX 7700 XT | $404 | 41.3% (68.4fps) | 36.5% (49.7fps) | 30.6% (31.8fps) | 27.2% (15.2fps) | Navi 32, 3456 shaders, 2544MHz, 12GB GDDR6@18Gbps, 432GB/s, 245W |
Radeon RX 6800 | $849 | 36.3% (60.1fps) | 30.2% (41.2fps) | 25.4% (26.3fps) | Row 27 – Cell 5 | Navi 21, 3840 shaders, 2105MHz, 16GB GDDR6@16Gbps, 512GB/s, 250W |
GeForce RTX 2080 Super | Row 28 – Cell 1 | 35.8% (59.4fps) | 30.8% (42.0fps) | 26.1% (27.1fps) | Row 28 – Cell 5 | TU104, 3072 shaders, 1815MHz, 8GB [email protected], 496GB/s, 250W |
GeForce RTX 4060 | $294 | 35.4% (58.8fps) | 30.6% (41.7fps) | 24.9% (25.8fps) | Row 29 – Cell 5 | AD107, 3072 shaders, 2460MHz, 8GB GDDR6@17Gbps, 272GB/s, 115W |
GeForce RTX 2080 | Row 30 – Cell 1 | 34.4% (57.1fps) | 29.1% (39.7fps) | 24.6% (25.5fps) | Row 30 – Cell 5 | TU104, 2944 shaders, 1710MHz, 8GB GDDR6@14Gbps, 448GB/s, 215W |
Intel Arc A770 8GB | No Stock | 32.7% (54.2fps) | 28.4% (38.7fps) | 24.0% (24.9fps) | Row 31 – Cell 5 | ACM-G10, 4096 shaders, 2400MHz, 8GB GDDR6@16Gbps, 512GB/s, 225W |
Intel Arc A770 16GB | $299 | 32.6% (54.1fps) | 28.3% (38.6fps) | 25.3% (26.2fps) | Row 32 – Cell 5 | ACM-G10, 4096 shaders, 2400MHz, 16GB [email protected], 560GB/s, 225W |
GeForce RTX 3060 | Row 33 – Cell 1 | 31.7% (52.5fps) | 25.7% (35.1fps) | 21.1% (22.0fps) | Row 33 – Cell 5 | GA106, 3584 shaders, 1777MHz, 12GB GDDR6@15Gbps, 360GB/s, 170W |
GeForce RTX 2070 Super | Row 34 – Cell 1 | 31.6% (52.4fps) | 26.8% (36.6fps) | 22.3% (23.1fps) | Row 34 – Cell 5 | TU104, 2560 shaders, 1770MHz, 8GB GDDR6@14Gbps, 448GB/s, 215W |
Intel Arc A750 | $189 | 30.7% (51.0fps) | 26.8% (36.6fps) | 22.6% (23.5fps) | Row 35 – Cell 5 | ACM-G10, 3584 shaders, 2350MHz, 8GB GDDR6@16Gbps, 512GB/s, 225W |
Radeon RX 6750 XT | $359 | 30.0% (49.8fps) | 25.3% (34.5fps) | 20.7% (21.5fps) | Row 36 – Cell 5 | Navi 22, 2560 shaders, 2600MHz, 12GB GDDR6@18Gbps, 432GB/s, 250W |
Radeon RX 6700 XT | $519 | 28.1% (46.6fps) | 23.7% (32.3fps) | 19.1% (19.9fps) | Row 37 – Cell 5 | Navi 22, 2560 shaders, 2581MHz, 12GB GDDR6@16Gbps, 384GB/s, 230W |
GeForce RTX 2070 | Row 38 – Cell 1 | 27.9% (46.3fps) | 23.5% (32.1fps) | 19.7% (20.4fps) | Row 38 – Cell 5 | TU106, 2304 shaders, 1620MHz, 8GB GDDR6@14Gbps, 448GB/s, 175W |
Intel Arc A580 | $169 | 27.5% (45.6fps) | 24.0% (32.7fps) | 20.3% (21.1fps) | Row 39 – Cell 5 | ACM-G10, 3072 shaders, 2300MHz, 8GB GDDR6@16Gbps, 512GB/s, 185W |
GeForce RTX 2060 Super | Row 40 – Cell 1 | 26.8% (44.5fps) | 22.4% (30.5fps) | 18.5% (19.3fps) | Row 40 – Cell 5 | TU106, 2176 shaders, 1650MHz, 8GB GDDR6@14Gbps, 448GB/s, 175W |
Radeon RX 7600 XT | $314 | 26.6% (44.2fps) | 22.6% (30.8fps) | 18.3% (19.0fps) | 16.0% (8.9fps) | Navi 33, 2048 shaders, 2755MHz, 16GB GDDR6@18Gbps, 288GB/s, 190W |
Radeon RX 6700 10GB | No Stock | 25.9% (42.9fps) | 21.4% (29.2fps) | 16.8% (17.5fps) | Row 42 – Cell 5 | Navi 22, 2304 shaders, 2450MHz, 10GB GDDR6@16Gbps, 320GB/s, 175W |
GeForce RTX 2060 | Row 43 – Cell 1 | 23.2% (38.4fps) | 18.6% (25.4fps) | Row 43 – Cell 4 | Row 43 – Cell 5 | TU106, 1920 shaders, 1680MHz, 6GB GDDR6@14Gbps, 336GB/s, 160W |
Radeon RX 7600 | $249 | 23.1% (38.3fps) | 18.9% (25.7fps) | 14.7% (15.2fps) | Row 44 – Cell 5 | Navi 33, 2048 shaders, 2655MHz, 8GB GDDR6@18Gbps, 288GB/s, 165W |
Radeon RX 6650 XT | $254 | 22.7% (37.6fps) | 18.8% (25.6fps) | Row 45 – Cell 4 | Row 45 – Cell 5 | Navi 23, 2048 shaders, 2635MHz, 8GB GDDR6@18Gbps, 280GB/s, 180W |
GeForce RTX 3050 | $169 | 22.3% (36.9fps) | 18.0% (24.6fps) | Row 46 – Cell 4 | Row 46 – Cell 5 | GA106, 2560 shaders, 1777MHz, 8GB GDDR6@14Gbps, 224GB/s, 130W |
Radeon RX 6600 XT | $239 | 22.1% (36.7fps) | 18.2% (24.8fps) | Row 47 – Cell 4 | Row 47 – Cell 5 | Navi 23, 2048 shaders, 2589MHz, 8GB GDDR6@16Gbps, 256GB/s, 160W |
Radeon RX 6600 | $189 | 18.6% (30.8fps) | 15.2% (20.7fps) | Row 48 – Cell 4 | Row 48 – Cell 5 | Navi 23, 1792 shaders, 2491MHz, 8GB GDDR6@14Gbps, 224GB/s, 132W |
Intel Arc A380 | $119 | 11.0% (18.3fps) | Row 49 – Cell 3 | Row 49 – Cell 4 | Row 49 – Cell 5 | ACM-G11, 1024 shaders, 2450MHz, 6GB [email protected], 186GB/s, 75W |
Radeon RX 6500 XT | $139 | 5.9% (9.9fps) | Row 50 – Cell 3 | Row 50 – Cell 4 | Row 50 – Cell 5 | Navi 24, 1024 shaders, 2815MHz, 4GB GDDR6@18Gbps, 144GB/s, 107W |
Radeon RX 6400 | $139 | 5.0% (8.3fps) | Row 51 – Cell 3 | Row 51 – Cell 4 | Row 51 – Cell 5 | Navi 24, 768 shaders, 2321MHz, 4GB GDDR6@16Gbps, 128GB/s, 53W |
Tom’s Hardware’s ray tracing benchmarks highlight Nvidia’s architectural advantage in ray tracing performance with their Ada Lovelace architecture. The RTX 4090 demonstrates a significant lead, outpacing the previous-generation RTX 3090 Ti by 41% at 1080p medium, a lead that expands to 53% at 1080p ultra and nearly 64% at 1440p. Without DLSS 3, the RTX 4090 still achieves a remarkable 72% performance increase over the 3090 Ti at 4K in ray tracing workloads, according to Tom’s Hardware’s testing.
AMD, as observed by Tom’s Hardware, prioritizes rasterization performance and cost-effectiveness through chiplet designs in their RDNA 3 GPUs, resulting in less competitive ray tracing performance. The RX 7900 XTX, AMD’s flagship, performs roughly on par with Nvidia’s previous-generation RTX 3080 12GB in ray tracing, placing it just ahead of the RTX 4070. While RDNA 3 shows minor ray tracing improvements, as seen with the RX 7800 XT outperforming the RX 6800 XT in DXR benchmarks despite similar rasterization performance, AMD still lags behind Nvidia in ray tracing capabilities as per Tom’s Hardware’s findings.
Intel’s Arc A7-series GPUs, as tested by Tom’s Hardware, present a mixed performance profile in ray tracing. The A750 surpasses the RTX 3060 overall in ray tracing workloads. With driver optimizations, Arc GPUs now exhibit more consistent performance in DXR titles, including Minecraft.
(Image credit: Tom’s Hardware)
Tom’s Hardware also assessed the impact of DLSS Quality mode on ray tracing performance in their RTX 4090 review, observing a substantial 78% performance boost at 4K ultra. DLSS 3 frame generation further increased frame rates by 30% to 100% in their tests, though they caution against relying solely on FPS metrics with frame generation enabled due to potential latency and perceived smoothness issues.
Overall, with DLSS 2 enabled, Tom’s Hardware’s ray tracing benchmarks indicate the RTX 4090 is nearly four times faster than AMD’s RX 7900 XTX. While AMD’s FSR 2 and FSR 3 offer upscaling alternatives, Tom’s Hardware notes that DLSS maintains an advantage in both game support and overall image quality. Native 4K ray tracing remains challenging for most GPUs, with only the RTX 3090 Ti and above consistently exceeding 30 fps in Tom’s Hardware’s composite DXR benchmark.
AMD’s FSR 3 frame generation, while promising, introduces latency, and its Anti-Lag+ feature, designed to mitigate this, is AMD GPU-exclusive, potentially disadvantaging non-AMD users. Tom’s Hardware’s testing in games like Avatar: Frontiers of Pandora shows variable FSR 3 performance, with quality and latency remaining inconsistent across different titles.
Mid-range GPUs like the RTX 3070 and RX 6700 XT generally handle 1080p ultra ray tracing, while entry-level DXR-capable GPUs struggle to maintain playable frame rates even at 1080p medium. The RX 6500 XT, in Tom’s Hardware’s testing, falls short even at 1080p medium, with single-digit frame rates in most DXR benchmarks.
Intel’s Arc A380 surprisingly outperforms the RX 6500 XT in ray tracing, despite having fewer Ray Tracing Units (RTUs), highlighting Intel’s ray tracing hardware efficiency, as noted by Tom’s Hardware. However, the limited number of RTUs in Arc GPUs, even in the top-tier A770, caps overall ray tracing performance, preventing it from significantly exceeding the RTX 3060 in DXR benchmarks. Arc A750 and above cards, however, surpass AMD’s RX 6750 XT in DXR performance, further illustrating AMD’s relative weakness in ray tracing compared to both Nvidia and Intel, according to Tom’s Hardware’s comprehensive testing.
Tom’s Hardware’s generational performance analysis of Nvidia RTX cards reveals that the older RTX 2060 still marginally outperforms the newer RTX 3050. However, the RTX 2080 Ti, despite being from the 20-series, comes close to the RTX 3070 in ray tracing performance. Overall, the performance scaling within the RTX 30-series is more pronounced than in the 20-series generation, as per Tom’s Hardware’s data.
(Image credit: Tom’s Hardware)
Test System and How Tom’s Hardware Tests GPUs
Tom’s Hardware employs multiple test platforms for GPU benchmarking, with the latest 2022–2024 configuration utilizing an Alder Lake CPU and platform, while previous tests used a Coffee Lake and Z390 setup. The most recent benchmark charts, as presented in this article, are based on a Core i9-13900K system with an updated game selection. The detailed specifications of Tom’s Hardware’s test PCs are as follows:
Tom’s Hardware 2022–2024 GPU Testbed:
- CPU: Intel Core i9-12900K
- Motherboard: MSI Pro Z690-A WiFi DDR4
- Memory: Corsair 2x16GB DDR4-3600 CL16
- SSD: Crucial P5 Plus 2TB
- Power Supply: Cooler Master MWE 1250 V2 Gold
- CPU Cooler: Cooler Master PL360 Flux
- Case: Cooler Master HAF500
- Operating System: Windows 11 Pro 64-bit
Tom’s Hardware 2020–2021 GPU Testbed:
- CPU: Intel Core i9-9900K
- CPU Cooler: Corsair H150i Pro RGB
- Motherboard: MSI MEG Z390 Ace
- Memory: Corsair 2x16GB DDR4-3200
- SSD: XPG SX8200 Pro 2TB
- Operating System: Windows 10 Pro (21H1)
Tom’s Hardware follows a consistent testing methodology for each GPU. This involves an initial “warm-up” benchmark run, followed by at least two benchmark passes at each setting and resolution. Results are validated for consistency, with retesting conducted to address any anomalies or outliers, ensuring data accuracy and reliability in their GPU comparisons. Tom’s Hardware also actively monitors driver updates and game patches, periodically retesting sample cards to maintain benchmark validity and relevance. New games may be added to their test suite based on popularity and suitability for benchmarking, adhering to their criteria for good game benchmarks.
GPU Benchmarks: Individual Game Charts (Updated November 11, 2024)
For a more granular performance breakdown, Tom’s Hardware provides individual game charts for both standard rasterization and ray tracing benchmarks. These charts, updated as of November 11, 2024, present detailed performance data for recent GPUs across each game in the test suites, using their latest test PC configuration.
(Note: Due to the extensive number of charts, only image captions are provided below. Please refer to the original Tom’s Hardware article for the full interactive charts.)
GPU Benchmarks — 1080p Medium Individual Game Charts:
Image 1 of 22
(Image credit: Tom’s Hardware)
… (Images 2-22 follow a similar pattern for Far Cry 6, Flight Simulator, Forza Horizon 5, Horizon Zero Dawn, Red Dead Redemption 2, Total War Warhammer 3, and Watch Dogs Legion)
GPU Benchmarks — 1080p Ultra Individual Game Charts:
Image 1 of 22
(Image credit: Tom’s Hardware)
… (Images 2-22 follow a similar pattern for Far Cry 6, Flight Simulator, Forza Horizon 5, Horizon Zero Dawn, Red Dead Redemption 2, Total War Warhammer 3, and Watch Dogs Legion)
GPU Benchmarks — 1440p Ultra Individual Game Charts:
Image 1 of 22
(Image credit: Tom’s Hardware)
… (Images 2-22 follow a similar pattern for Far Cry 6, Flight Simulator, Forza Horizon 5, Horizon Zero Dawn, Red Dead Redemption 2, Total War Warhammer 3, and Watch Dogs Legion)
GPU Benchmarks — 4K Ultra Individual Game Charts:
Image 1 of 22
(Image credit: Tom’s Hardware)
… (Images 2-22 follow a similar pattern for Far Cry 6, Flight Simulator, Forza Horizon 5, Horizon Zero Dawn, Red Dead Redemption 2, Total War Warhammer 3, and Watch Dogs Legion)
GPU Benchmarks — Power, Clocks, and Temperatures
Beyond raw gaming performance, Tom’s Hardware also provides detailed analysis of GPU power consumption, clock speeds, and temperatures. These metrics are crucial for understanding GPU efficiency, thermal management, and overall system requirements.
Image 1 of 4
(Image credit: Tom’s Hardware)
Image 2 of 4
(Image credit: Tom’s Hardware)
Image 3 of 4
(Image credit: Tom’s Hardware)
Image 4 of 4
(Image credit: Tom’s Hardware)
(Note: Due to the extensive number of charts for each metric, only image captions are provided above. Please refer to the original Tom’s Hardware article for the full interactive charts for power, clocks and temperatures across all tested GPUs.)
For access to Tom’s Hardware’s legacy GPU hierarchy and older benchmark data, please refer to page two of their article. For further discussion and community engagement regarding GPU benchmarks, visit Tom’s Hardware forums.
Choosing the Right Graphics Card: Tom’s Hardware Guide
To guide your graphics card selection, Tom’s Hardware has created this comprehensive GPU benchmarks hierarchy, encompassing a wide range of GPUs from the last four generations. Their testing consistently shows that the latest cards from Nvidia Ada Lovelace and AMD RDNA 3 architectures offer the highest performance. AMD GPUs excel in rasterization, while Nvidia cards typically lead in ray tracing, especially when leveraging DLSS. While GPU prices have become more reasonable, making upgrades more accessible, careful consideration of your specific needs and budget remains essential.
Beyond gaming, Tom’s Hardware emphasizes that GPUs are crucial for various professional applications. Their full GPU reviews include professional GPU benchmarks, demonstrating GPU performance in content creation and compute workloads. A capable gaming GPU generally translates to strong performance in these areas as well. Selecting a top-tier card ensures high-resolution, high-framerate gaming with maximum visual settings and provides ample power for demanding content creation tasks. Lower-tier GPUs necessitate compromises in settings to achieve acceptable performance in both gaming and professional applications.
Tom’s Hardware reminds readers that CPU selection is equally vital for gaming. An underpowered CPU can bottleneck even the most powerful GPU. They recommend consulting their Best CPUs for gaming guide and CPU Benchmarks Hierarchy to ensure a balanced system configuration for optimal gaming experiences.
- 1
- 2
Current page: GPU Benchmarks Hierarchy 2025
Next Page: 2020-2021 and Legacy GPU Benchmarks Hierarchy
Stay Informed: Tom’s Hardware Newsletter
For the latest tech news and in-depth reviews, subscribe to the Tom’s Hardware Newsletter.
TOPICS: GPU