amd radeon rx 9070 xt

amd radeon rx 9070 xt

Thumbnail

AMD’s New Radeon RX 9070 XT Could Cost Just $699

Image

ASUS TUF Gaming Radeon RX 9070 XT OC Edition graphics card

If you think PC graphics card prices are high then you’re not alone, but there’s hope on the horizon thanks to AMD, whose highly anticipated Radeon RX 9700 XT and 9700 have been listed at Micro Center for just $699 and $649 respectively.

Spotted by Twitter user @gawroskiT who saw several new models on Micro Center’s website, the pricing undercuts Nvidia’s RTX 5070 Ti by $50, and that’s before you factor in actual pricing, which has been $100 or more above its $750 launch price, with many PC gamers complaining about the inflated cost compared to Nvidia’s stated pricing. Currently all RTX 5070 Ti models are also out of stock on Newegg’s website.

The RX 9070 XT is meant to compete with Nvidia x070 models such as the RTX 5070 Ti, 5070, 4070 Ti Super and RTX 4070 Super, while some leaked performance numbers show favorable results compared to the RTX 4080 too. The pricing is still more than recent rumors, that put it south of $600, but all eyes will be on retail pricing once it hits shelves given how inflated Nvidia’s pricing has been of all three of its RTX 50-series cards released so far.

AMD has named its new graphics cards the RX 9070 XT and RX 9070 to allow for clearer comparison with … [+]

Forbes Daily: Join over 1 million Forbes Daily subscribers and get our best stories, exclusive reporting and essential analysis of the day’s news in your inbox every weekday.

The most significant news is that AMD will fully reveal the Radeon RX 9070 XT and 9070 tomorrow – Feb. 28 and the cards themselves will be available early March. You can view AMD’s announcement at 8am EST on its YouTube channel.

The RX 9070 XT is expected to have 16GB of video memory, although there’s no confirmation of the amount or the type of memory being used. Rumors were flying this week about a possible 32GB version of the card but these were quashed this week by AMD itself.

Sadly, this appears to be the highest-performing model in the range, meaning AMD will not be aiming to take the performance crown away from Nvidia’s RTX 5090, and we may not see performance top AMD’s own Radeon RX 7900 XTX either. However, recent leaked information suggest it could outperform the Radeon RX 7900GRE by up to 42% at 4K resolution in games.

I’ll be covering AMD’s announcement tomorrow to be sure to follow me here on Forbes using the blue button below.

One Community. Many Voices. Create a free account to share your thoughts.

Why Is The Radeon RX 9070 XT Pricing Significant?

What Else Do We Know About The Radeon RX 9070 XT

How Fast Will The Radeon RX 9070 XT Be Compared To Previous Models?

Join The Conversation

Trump’s DOJ Publicly Releases ‘Phase 1’ Of Jeffrey Epstein Files—What To Know

Sudden $1 Trillion Crypto Market Shock Sees Bitcoin Plunge Under $80,000 Price

Here’s Where Trump’s Government Layoffs Are Targeted—As Judge Blocks Mass Firings Of Probationary Employees

AMD RDNA 4 and Radeon RX 9000-series GPUs start at $549: Specifications, release date, pricing, and more revealed

Image

Everything you need to know about AMD’s RDNA 4 GPU architecture and graphics cards.

When you purchase through links on our site, we may earn an affiliate commission. Here’s how it works.

AMD has announced the Radeon RX 9070 at $549 and the RX 9070 XT at $599, both of which ship on March 6. The AMD RDNA 4 architecture and Radeon RX 9000-series GPUs were partially revealed at CES 2025, except they weren’t part of AMD’s keynote. Very little was known (officially) other than the names of the first two graphics cards for the family. That changes today, with AMD detailing many of the architectural upgrades, specifications, and more, during a video presentation. These will go up against the Nvidia Blackwell RTX 50-series GPUs and the Intel Battlemage Arc B-series GPUs and will likely join the ranks of the best graphics cards in the coming days.

Like Nvidia’s RTX 50-series graphics cards, AMD’s RDNA 4 launch seems to have been delayed, though perhaps for different reasons. There were rumors that the cards would be revealed at CES 2025 and launched in January, then February, and finally March. That last is no longer a rumor, with the RX 9070 XT and RX 9070 set to go on sale on March 6 — and in typical fashion, the “MSRP” or base model cards will have reviews go up the day before, followed by the overclocked non-MSRP models on the launch date. Nvidia’s RTX 5070 will likely land right around the same time, just to make things even more exciting.

But if you look at graphics card availability right now, what becomes immediately clear is that virtually everything is sold out or, at the very least, seriously overpriced. AMD has had difficulties with GPU transitions in the past, with the prior generation hanging around for too long and competing with the new parts. This time, it seems to have gone the opposite way, with RX 7000-series GPUs mostly having disappeared from retail shelves in December and January. Only the lower tier RX 7600 and RX 7600 XT are still in stock at MSRP (or close to it).

The result has been dramatically increased demand for everything from mainstream to high-end graphics cards, and Nvidia’s RTX 5090, RTX 5080, and RTX 5070 Ti all sold out almost instantly at launch. Will AMD’s 9070 XT and 9070 fare better? We can hope so, but we suspect there’s so much pent-up demand that even with another two months’ worth of production and supply, it will still be insufficient. Hopefully, things will settle down later this year, but in the near term, we expect inadequate supplies and increased retail prices — and, yes, scalping.

No doubt Nvidia’s record profits driven by AI are a big contributor, and while AMD isn’t selling quite as many data center GPUs, a lot of its wafer allocation from TSMC is likely going to data center CPUs and GPUs as well. Gamers are no longer the top priority for either company, in other words; for the time being, they just get the scraps that fall from the AI table.

But enough sad talk. Let’s check out the specifications for AMD’s RDNA 4 GPUs, talk about architectural updates, and dig into all the other details. We even have pricing information, though as you can guess, that’s worth about as much as the paper this is printed on. We’ll continue updating this article as additional details become available, but for now, here’s everything you need to know about the AMD RDNA 4 and Radeon RX 9000-series GPUs.

Here are the known specifications for the RX 9070 series GPUs, along with placeholder information on the RX 9060 series. AMD did share the 9060 name at CES 2025, but no other details have been shared. There are rumors, however, which we’ve used to flesh out the table — these are indicated by question marks in the various cells.

Graphics Card

RX 9070 XT

RX 9070

RX 9060 XT?

RX 9060?

Architecture

Navi 48

Navi 48

Navi 48?

Navi 44?

Process Technology

TSMC N4P

TSMC N4P

TSMC N4P

TSMC N4P?

Transistors (Billion)

53.9

53.9

53.9

22?

Die size (mm^2)

356.5

356.5

356.5

153?

SMs / CUs / Xe-Cores

64

56

32?

20?

GPU Shaders (ALUs)

4096

3584

2048?

1280?

Tensor / AI Cores

128

112

64?

40?

Ray Tracing Cores

64

56

32?

20?

Boost Clock (MHz)

2970

2520

2790?

2700?

VRAM Speed (Gbps)

20

20

20?

20?

VRAM (GB)

16

16

12?

8?

VRAM Bus Width

256

256

192?

128?

L2 / Infinity Cache

64

64

48?

32?

Render Output Units

128

128

96?

64?

Texture Mapping Units

256

224

128?

80?

TFLOPS FP32 (Boost)

48.7

36.1

22.9?

13.8?

TFLOPS FP16 (FP4/FP8 TFLOPS)

389 (1557)

289 (1156)

183 (731)?

111 (442)?

Bandwidth (GB/s)

640

640

480?

320?

TBP (watts)

304

220

150?

120?

Launch Date

Mar 2025

Mar 2025

Apr–Jun 2025?

Apr–Jun 2025?

Launch Price

$599

$549

$399?

$299?

The RX 9070 XT and 9070 columns should be fully accurate. We’re reasonably sure there will be a trimmed-down RX 9060 XT using the same Navi 48 die as the 9070 cards, just with fewer CUs (Compute Units) and memory controllers enabled. Below that, things get murky.

The RX 9060 could use a further binned Navi 48, or it could use Navi 44. Most of the details on Navi 44 are questionable at best, but we’ll certainly find out more in the coming months. There might even be RX 9050-class GPUs at some point, but we’ve avoided listing those for the time being.

Looking at the RX 9070 XT, it uses a fully enabled Navi 48 die that includes 64 RDNA 4 CUs. Combined with a 2.97 GHz boost clock and a 256-bit memory interface with 20 Gbps GDDR6 VRAM, the other specifications mostly come from straight mathematical calculations. The RX 9070 is mostly the same configuration, just with 56 CUs and a 2.52 GHz boost clock — substantially lower than its bigger sibling, though we’ll have to wait and see what real-world clocks actually look like.

Power targets also play a role in the final clock speeds, and where the 9070 XT has a 304W TBP (Total Board Power), the 9070 cuts that all the way down to 220W. That’s probably a big factor in the 450 MHz difference in boost clocks.

Raw compute works out to 48.7 TFLOPS FP32 on the 9070 XT and 36.1 TFLOPS on the 9070. On paper, that makes the XT up to 35% faster. In practice, we suspect the two chips will be quite a bit closer and that the actual clocks in most games may only be a couple hundred MHz apart, despite what the specs suggest.

AMD has also given the Ray Accelerators and AI Accelerators in the CUs a massive overhaul compared to RDNA 3. For AI, each can do twice as many FP16 operations per cycle and they now support sparse operations. Sparsity can skip up to half of the zero multiply operations to potentially double performance, and it’s a feature Nvidia has supported since its second-generation RTX 30-series GPUs. (AMD has also supported sparse operations on its CDNA family of GPUs for several years.)

Moreover, the AI units also support FP8, INT8, BF8, and INT4 operations, with the 8-bit calculations being twice as fast as 16-bit, and 4-bit integers double that again. Put it all together, and you get 389 TFLOPS of sparse FP16 compute and up to 1557 TOPS of sparse INT4 compute.

Keep in mind that the previous generation RDNA 3 architecture featured GPUs with up to 96 CUs and a 384-bit memory interface on the RX 7900 XTX, so while RDNA 4 GPUs are faster on a per-CU basis, AMD doesn’t expect the RX 9070 XT to beat the RX 7900 XTX in all workloads.

There’s a lot more going on than the raw specs will tell you. First, let’s cover the pricing and launch date, then move on to the architectural deep dive.

How much will the RX 9000-series GPUs cost? AMD has announced the Radeon RX 9070 will start at $549 and the RX 9070 XT will start at $599, placing them firmly in the “mainstream” segment. Given the current market conditions, however, it probably doesn’t matter what AMD has given as the MSRP. Short-term, certainly, we expect the cards will all sell out and end up costing much more than the MSRP.

As we said in the Nvidia Blackwell overview, for dedicated desktop graphics cards, we’re now living in a world where “budget” means around $250–$300, “mainstream” means $400–$700, “high-end” is for GPUs costing $800 to $1,000, and the “enthusiast” segment targets $1,500 or more. AMD is going after the mainstream segment with the 9070 series, and possibly the lower mainstream and upper budget segments with future 9060 series parts.

Depending on supply, as well as performance, the RDNA 4 series should be worth the price AMD is asking. More likely is that there simply won’t be enough cards to satisfy the current demand, not for many months. Given the reasonably low-ish prices, don’t be surprised if scalpers and retail markups step in and push the prices up.

It’s basically a repeat of the cryptocurrency GPU mining shortages, only this time it’s caused by AI and demand from that sector may not go away for many years. Let’s hope we’re wrong, but the RTX 50-series launches so far have not been promising.

We listed the March 6, 2025 release date for the RX 9070 cards already, but AMD has also at least partially teased an RX 9060 family of GPUs. Will there be multiple cards or only one? Will there be lower-tier RX 9050 cards as well? The short answer: We don’t know. The nebulous answer: Sometime between April and the end of the year, hopefully sooner than later.

We’ve seen rumored die sizes for Navi 44 that suggest it’s a much smaller chip, like more of a replacement for the current Navi 33 (RX 7600 series). If that’s correct, it may not come out any time soon. There still appear to be plenty of RX 7600 and RX 7600 XT GPUs floating around, and that’s because when those launched there were still a lot of similar performing Navi 23 (RX 6650 XT / RX 6600 XT / RX 6600) cards still available, at lower prices.

The naming scheme from AMD suggests that the RX 9060 will compete with the RTX 5060 family. That would perhaps require a larger chip than what’s indicated. But RX 7600 does technically compete with the RTX 4060, and there’s no RTX 4050 and probably won’t be an RTX 5050.

Will AMD be making a “true budget” RDNA 4 chip? Again, rumors suggest that’s at least possible, perhaps even likely. At less than half the size of Navi 48, AMD may try to create a $200~$250 graphics card to go after budget-minded gamers — and OEMs. Certainly it could get a lot more chips per wafer with the rumored 150~160 mm^2 die size.

But if the cards then only sell for $250 or less? That hardly seems worth the effort, not when companies can charge tens of thousands for data center GPUs.

The above slide gallery covers the architectural briefing AMD provided in advance of today’s reveal, including some of the specifications discussed above as well as finer details. AMD worked to refine the underlying architecture to improve per-CU performance in all major workloads: rasterization, ray tracing, and AI.

Rasterization performance sees the smallest generational gains, but it’s still about 40% faster than RDNA 3 according to AMD. Ray tracing performance is basically doubled, and AI performance is doubled for dense FP16 compute, with lower precision formats delivering even higher performance.

The specific details of the rasterization improvements are a bit nebulous. RDNA 4 supports out of order memory requests, which AMD specifically notes as being helpful for ray tracing, but it can help rasterization tasks as well — we just don’t have any details on how much. The other major change involves dynamic register allocation. RDNA 3 (and earlier) allocated registers for the worst case for shaders. By dynamically allocating extra registers only when needed, AMD provides an example use case where it could have an extra wave in flight. The slides show three waves versus four waves, which would be a 33% increase, but we don’t know if that’s representative of real workloads or just for illustrative purposes.

Moving on to ray tracing, this is where AMD spent a significant amount of effort. It doubled the ray/triangle and ray/box intersection rates per RT unit as a start. THen it offers some enhancements including hardware instance transforms (rather than doing a lot of the work via GPU shaders), oriented bounding boxes, an improved BVH (Bounding Volume Hierarchy) structure and traversal, the above-mentioned out of order memory returns, and better ray hardware stack management.

Most of the improvements come from the doubling of intersection rates and BVH compression, but the other aspects combine to deliver a solid improvement as well. How does RDNA 4 compare with Nvidia’s latest hardware? That’s not fully clear, but certainly it’s going to do better per CU than what we saw with RDNA 3 and RDNA 2. It likely won’t match Blackwell, but it might be better than Ampere and at least closer to Ada levels of performance.

AI, as already noted, sees the biggest changes. Nvidia has been iterating on its AI tensor cores since the RTX 20-series, and even before that the Volta data center GPU had tensor cores. So Nvidia is on its fifth generation of AI matrix cores while AMD is mostly on its second generation — mostly because it looks like AMD took a lot of the work that’s been happening in its CDNA GPUs and brought it over to RDNA 4.

RDNA 3 CUs could do 512 FP16 operations per cycle, with no sparsity support, or 1024 INT4 operations per cycle. With RDNA 4, AMD doubles the baseline FP16 throughput for dense operations, doubles that again for sparse operations, and doubles that again for FP8 workloads — which are proving useful in the AI space. That’s up to 8X higher AI throughput for FP8 on RDNA 4 compared to FP16 on RDNA 3, and the INT4 throughput sees a similar up to 8X improvement.

AMD gave a real-world example of how this affects AI performance using Stable Diffusion XL. The RX 9070 XT with 64 CUs took on the RX 7900 XT with 84 CUs. That gives the older GPU a 31% advantage in compute units, but the 9070 XT ended up delivering very close to 2X the performance. That will prove very helpful for other AI and machine learning workloads, including ML-based upscaling and frame generation (see FSR 4 below).

Alongside these changes, AMD has reworked some of the cache and memory hierarchy with RDNA 4. It didn’t provide any clear details on what has changed, but it notes that this is the third generation of Infinity Cache. The capacity remains 64MB, the same as what was present on the 256-bit 7900 GRE and 7800 XT, but now the cache is again part of the monolithic chip, so it likely has better latencies and throughput.

Get Tom’s Hardware’s best news and in-depth reviews, straight to your inbox.

RDNA 4 isn’t just about core architecture upgrades. Along with the above rasterization, ray tracing, and AI enhancements, AMD has also upgraded a few other areas. One of the big changes is with the media encoding hardware. Last time we checked video encoding performance and quality, AMD came in last place, clearly behind Nvidia and Intel. It looks like RDNA 4 will close the gap.

AMD says it has improved H.264 (AVC) quality by up to 25%, H.265 (HEVC) by 11%, and improved the AV1 encoding efficiency. It also has better support for AV1 and VP9 decoding and reduced memory accesses.

Besides the quality improvements, RDNA 4 adds a dual media engine. Nvidia did this with its Ada architecture, and AMD seems to be taking a similar approach. It likely doesn’t help all workloads equally, but AMD says it doubles the AV1 encoding throughput.

Realistically, there’s only so far you can go with improving video encoding quality, particularly with hardware encoders. Intel and Nvidia are pretty comparable, but AMD was behind on quality while being ahead on performance with RDNA 3. It sounds like RDNA 4 will continue to be faster while offering similar quality to the competition, which is a good thing.

Another change with RDNA 4 is that AMD has added hardware flip queue support, which offloads video frame scheduling to the GPU. While Nvidia discussed something similar for MFG (Multi Frame Generation), it sounds like AMD’s solution is focused on improving video playback by reducing CPU load, as opposed to being something to improve the scheduling of generated frames.

Radeon Image Sharpening (RIS) has also been updated, to RIS2. This is a driver level sharpening solution that’s based on AMD’s CAS algorithm (Contrast Aware Sharpening), only now the quality is supposed to be better. It’s a single click toggle to apply RIS2 across all APIs.

Finally, RDNA 4 GPUs will support PCIe 5.0 interfaces. That doubles the throughput over the x16 link, though in practice most workloads likely won’t see much benefit. Gaming in particular doesn’t tend to need more than PCIe 3.0, or perhaps 4.0, when using a full x16 connection. However, AI and certain content creation tasks can benefit from the added bandwidth. Don’t be surprised if the future Navi 48 chips and possibly even the RX 9060 XT cut the interface down to x8 or even x4 widths.

One thing that isn’t changing from RDNA 3 is the memory support. While Nvidia has moved all of the announced Blackwell RTX 50-series solutions to GDDR7 memory, AMD will continue to use GDDR6 memory, clocked at up to 20 Gbps. Coupled with a 256-bit interface on the 9070 XT and 9070 GPUs, that results in 640 GB/s of memory bandwidth. That’s the same VRAM capacity as the RX 7900 GRE and RX 7800 XT, and also the same as Nvidia’s RTX 5070 Ti and RTX 4070 Ti Super.

The 64MB Infinity Cache will improve the effective bandwidth, though AMD didn’t elect to provide any estimates of cache hit rates so far. The RX 7900 GRE and RX 7800 XT both had 64MB Infinity Caches, and AMD provided effective bandwidth rates that were about 4X the base memory bandwidth with those GPUs, so we’d anticipate the Navi 48 GPUs will see similar results.

It’s also possible that further improvements to the Infinity Cache have made it less critical for AMD to move to GDDR7 at present. Considering that Nvidia gets a 40% improvement in raw bandwidth from 28 Gbps GDDR7 compared to AMD’s 20 Gbps GDDR6, that might seem like a sizeable advantage. However, effective bandwidth after factoring in the large caches may not be all that different.

Plus, there’s only so much bandwidth needed to drive a 64 CU GPU. Nvidia’s RTX 5070 Ti for example has 70 SMs (Streaming Multiprocessors), which are roughly analogous to AMD’s CUs, and the 5070 Ti has a 48MB L2 cache. Putting a larger 64MB L3 cache with fewer GPU processing clusters could reduce the need for higher memory speeds.

AMD continues to use 16Gb (2GB) GDDR6 modules, and we’re unaware of any companies currently pursuing 24Gb (3GB) capacities. That’s one area where GDDR7 support could prove beneficial for Nvidia in the future, though so far only the RTX 5090 Laptop GPU is using the higher capacity chips.

One of the interesting changes with RDNA 4 is that AMD is, at least for now, ditching the GPU chiplets approach. It may come back to that in the future, but the Navi 48 and presumably the rumored Navi 44 will be monolithic chips. Along with that design choice, AMD is also upgrading from TSMC’s N5 process node used on RDNA 3 to the N4P node for RDNA 4.

N4P provides for modest improvements in performance and efficiency compared to the N4 node, which in turn refines the base N5 node. Our understanding is that N4P may introduce some additional metal layers, and N4 used more EUV than N5. What’s not entirely clear is how N4P compares to 4N and 4NP — the “for Nvidia” variants that are used with Hopper, Ada, and Blackwell. It’s probably pretty similar in most respects, which means that AMD will be on node parity with Nvidia this round.

But AMD isn’t really trying to take down Nvidia’s top GPUs. The lack of GDDR7 memory and the lack of a larger design prove this. The Navi 48 chip will house 53.9 billion transistors in a 356.5 mm^2 die. Nvidia’s GB203 used in the RTX 5080 and 5070 Ti contains 45.6 billion transistors in a 378 mm^2 die… which might suggest AMD actually has a superior process node and/or design. But we can’t really conclude that.

While die sizes are pretty straightforward, transistor counts are not. They’re more of a mathematical estimate, and there are different ways of counting what constitutes a “transistor.” Perhaps AMD does have a denser design with more transistors, perhaps not. Ultimately, we’ll have to see how the various GPUs perform.

One interesting side note here is that Navi 31, the top solution from the RDNA 3 family, had a 300 mm^2 GCD (Graphics Compute Die) with six 37.5 mm^2 MCDs (Memory Cache Dies). I wondered when AMD revealed the specs just how much it was actually saving by going the chiplet route. The GCD had 45.6 billion transistors, which means the overall transistor density — looking at the RDNA 3 GCD compared to the RDNA 4 monolithic design — is basically identical (152 MTrans/mm^2 on Navi 31 GCD compared to 151.2 Mtrans/mm^2 on Navi 48).

But let’s not get too carried away. It’s known that scaling on external interfaces — like the GDDR6 memory controllers — is quite poor with newer process nodes. Navi 31 used twelve 32-bit controllers while Navi 48 has eight 32-bit controllers. If AMD had attempted to make a 384-bit interface on a monolithic design, it would have certainly required a larger chip. Putting that on an older process node for the prior generation did make financial sense at the time, and may yet prove a smart approach for a future AMD product.

If it wasn’t clear yet, AMD will not be making or selling its own reference model RX 9070 series graphics cards. Despite providing some slides of what appear to be MBA (Made By AMD) cards, these are merely graphical renders rather than photos of actual hardware. There were certainly prototype cards created during the design, testing, and validation process, but what those looked like isn’t really important.

All of the RX 9070 series graphics cards will be made by AMD’s add-in board (AIB) partners. That means two things. First, we’ll see a lot of variation in final clock speeds and power draw, not to mention things like the number of fans and RGB lighting. But more importantly, it means AMD has a lot less say in the actual retail graphics card prices.

Very likely AMD has a requirement that all of the AIBs have at least one model for each GPU that will be nominally priced at the stated MSRP. Beyond that, however, all bets are off. “Here’s our RX 9070 XT Red Herring for $599… and we sold all of those. Sorry! But you can pick up our Redder Herring OC model for $799!” We saw something like this with the RTX 5070 Ti cards, where there also isn’t a reference model from Nvidia.

Long-term, if there’s insufficient supply to meet the demand, most AIBs are going to produce higher tier models with a few minor extras and drastically inflated prices. If on the other hand the supply catches up to demand, then it’s easy enough to drop prices as needed.

Besides the hardware, AMD has been working on a variety of feature improvements. The biggest one is undoubtedly FSR 4. The fourth iteration of AMD’s FidelityFX Super Resolution (FSR) algorithm, it will break with tradition in a couple of key ways.

First, FSR 4 will leverage the more potent AI accelerators in the RDNA 4 GPUs. At launch, it will require an RDNA 4 GPU. Down the road a few months, AMD may try back-porting the algorithm so that it can run on RDNA 3 and maybe even RDNA 2 GPUs… but it seems unlikely.

Instead, FSR 4 will basically co-exist with FSR 3.1, or rather, the non-AI upscaling will continue to be offered. It’s not entirely clear exactly how this will play out, but keeping everything unified under one name makes more sense. What we do know is that AMD plans to allow gamers to use the more potent FSR 4 algorithm on games that have FSR 3.1 support. Will that happen automatically or require a driver settings toggle? It seems like the latter but we’ll have to wait and see.

FSR 4 isn’t just for upscaling, either; it also has frame generation. From our understanding, both upscaling and framegen will use the AI accelerators of the RX 9000-series GPUs. AMD also says RDNA 4 is “neural rendering ready” without really going into further detail. Presumably that’s related to Microsoft’s new Cooperative Vectors feature, which is something Nvidia also talked about with Blackwell.

We’ve asked for additional details on how FSR 4 works, in terms of the computations. AMD hasn’t responded, but one slide does note that the RX 9070 XT offers “up to 779 TOPS AI Acceleration” while talking about FSR 4. Now, that’s either dense INT4 operations or sparse INT8 operations, as the 9070 XT hits double that figure for sparse INT4, but we don’t have a direct answer on whether the algorithms are using INT4 or INT8 yet. Either way, that’s a lot more theoretical compute than what you can get from prior generation AMD GPUs, which is why we don’t anticipate the AI upscaling and framegen models getting backported.

We also asked if FSR 4 was using a transformers-based network or a convolutional neural network. DLSS 4 offers better image fidelity than DLSS 3 by using transformers, and AMD may have skipped the CNN approach since it’s late to the AI-powered upscaling and framegen party. However, we don’t have a direct answer yet. We do have some image quality comparisons from AMD, in the slides above, and FSR 4 definitely looks better than FSR 3.1.

As with Nvidia’s use of performance mode upscaling with framegen, we don’t generally focus on the promised performance after all these extras. Framegen in particular is very heavy on marketing in our experience. It’s less problematic when you already have a high base framerate, but then it’s also less necessary when you’re already getting 100+ FPS.

AMD says it will have over 30 games with FSR 4 enabled for the RX 9070 series launch, with 75+ games coming in 2025.

AMD also talked about HYPR-RX, which combines a variety of driver-level performance boosting features and can be enabled with a single click. We’ve poked at it a bit in the past, and it can be useful in some cases, but we prefer sticking with apples to apples comparisons. If you’re just playing games, however, enabling HYPR-RX to apply all of the features including FSR/RSR, Anti-Lag, Radeon Boost, and AFMF 2 could be useful.

AMD also has a new AFMF 2.1 release that imprves the quality of the algorithm, reducing ghosting, improving fine features, and detecting and handling overlays better.

The last item AMD discussed is its new Adrenalin 25.3.1 drivers along with some new software. While most of the driver interface will be familiar to AMD GPU users, there are some new additions along with some behind the scenes changes. AI plays a role in both areas. We’ve already discussed FSR 4 upscaling and framegen, so let’s talk about the other AI uses.

First, AMD is using AI to help find rendering errors and to detect instability and other issues. AMD claims its new 25.3.1 drivers will be perhaps the best and most stable drivers it has ever released, with fewer rendering errors. We’ll have to wait and see how that goes…

Moving on, similar to Nvidia’s Chat RTX and other tools, AMD is providing some easy to access AI-powered features. These are all managed by a new AMD Install Manager that sits alongside the usual AMD Software in your system tray. Besides your GPU drivers, it can also detect if you have an AMD platform and keep your chipset drivers updated. And then there are some new extras: AMD Chat, AMD Image Inspector, and the AI Apps Manager (among others).

AMD Chat is a chatbot designed to answer questions specifically about your PC hardware and GPU. You can ask it about GPU temperatures, performance, and more. It’s a hefty 25GB download, though, so you might not want to install if it you’re low on space — or if chatting with your PC isn’t something you plan on doing.

The AI Apps Manager provides a list of software and utilities that can use AI that are installed or available to install. So if you have Adobe CC, some of those apps might show up. Or you can use it to install Amuse, AMD’s tuned AI image and video generation tool.

Finally, the Image Inspector is a feature to help with finding and reporting rendering errors and bugs. AMD is already using AI to help it find issues internally, and the Image Inspector is an opt-in feature that allows you to participate. Using spare GPU resources (so it won’t go crazy and use all your GPU power if you’re in a demanding game), it can automatically capture rendering errors and submit them to AMD, should you enable the feature. It sounds interesting, but we suspect there might be a performance hit still, even if it’s small.

Frankly, RDNA 4 feels like what AMD should have been doing with RDNA 3 rather than pursuing the abandoned-for-now GPU chiplets approach. AMD has finally decided to put serious effort into ray tracing hardware and AI in its consumer product line. We can understand why RDNA 2 was lacking in these areas — Nvidia’s GeForce RTX feature set probably caught the company off guard — but when RDNA 3 arrived in 2022, it really needed to do more and be more.

What’s interesting is that all of these new hardware features haven’t caused a massive bloat in the die size. Navi 48 is 357 mm^2 on a 5nm-class node (N4P). Navi 31 was 300 mm^2 on a 5nm-class node (N5), with Infinity Fabric links to the external memory and cache chiplets. Rip out those links, rework the cores, and this was all possible several years ago. Which is obvious, since Nvidia already did that, but it felt like RDNA 3 doubled down on the “ray tracing and AI aren’t really that important” marketing and got left behind. RDNA 4 finally rights that misstep, or at least attempts to. Now we just need to see how the actual hardware performs in a variety of tasks.

AMD’s RDNA 4 GPUs will have to compete with Nvidia Blackwell RTX and Intel Arc Battlemage solutions. As we discussed already, supply and availability of all graphics cards has become quite poor of late. (Yes, that’s a sarcastic understatement.) Every recently launched GPU has sold out quickly, with many cards then selling at prices far above the original MSRP. It started with the Arc B580 and became especially painful with Nvidia’s Blackwell launches.

Things aren’t going to get better in the near term. The big issue is that there are a lot of companies competing for a limited supply of silicon manufacturing. TSMC only has the ability to process so many wafers in a month. Right now, AMD, Apple, Intel, and Nvidia are all using TSMC for various chips, and there are plenty of other companies as well — Broadcom, Facebook, Google, Amazon… the list can get quite large.

Even if a company pays for a certain number of wafers in a given month, what to do with those wafers is still up for debate. Just looking at the main PC companies, AMD could make RDNA 4 GPUs on TSMC’s N4P node, sure. Or it could make Zen 5 CPU chiplets for both Ryzen and EPYC CPUs, CDNA 3 data center GPUs (MI300X), other Ryzen APU designs for laptops and handhelds, or the future CDNA 4 GPUs that are likely coming this year. Nvidia has Grace CPUs, Hopper and Blackwell data center GPUs, NVLink processors, Ada Lovelace previous generation GPUs, and the new Blackwell RTX GPUs that are all using variants of TSMC’s 5nm-class nodes. And Intel has leveraged TSMC for all or part of its Arrow Lake, Lunar Lake, and Battlemage product lines.

Nvidia made record profits last year of $130 billion, primarily driven by AI. Its consumer gaming division only accounted for $11.35 billion, 8.7% of the total. And going forward, Nvidia will likely invest even more heavily in data center GPUs. That will eat up a lot of wafers, needless to say, and gaming will have to compete for its share.

The good news is that more manufacturing capacity is coming online. A lot of that will likely go to create more AI processors, but the more capacity that exists, the more likely it is for other, less profitable chips — like consumer GPUs — to get made. And maybe AMD and Intel will try to grow their gaming GPU divisions while Nvidia is otherwise occupied. Or maybe Nvidia will treat gaming as a passion project that started the company and so it will try to ensure at least a reasonable number of chips get made. Maybe, maybe…

Whatever happens, what’s clear right now is that, as long as AI continues to grow as an industry, gaming GPUs are now a lower priority item for most of the biggest players in the graphics space. Let’s hope that, like cyptocurrency mining, this turns out to be just a passing phase. But we wouldn’t bet the farm on that.

Jarred Walton is a senior editor at Tom’s Hardware focusing on everything GPU. He has been working as a tech journalist since 2004, writing for AnandTech, Maximum PC, and PC Gamer. From the first S3 Virge ‘3D decelerators’ to today’s GPUs, Jarred keeps up with all the latest graphics trends and is the one to ask about game performance.

Sapphire does the unthinkable and puts 16-pin power connector on an RX 9070 XT

AMD RDNA4 Navi 48 is 25% denser than Nvidia Blackwell GPUs — 53.9 billion transistors in a die smaller than GB203

Intel delays $100 billion Ohio chipmaking site to next decade: First fab now coming online in 2030

Tom’s Hardware is part of Future US Inc, an international media group and leading digital publisher. Visit our corporate site.

© Future US, Inc. Full 7th Floor, 130 West 42nd Street, New York, NY 10036.

Stay On the Cutting Edge: Get the Tom’s Hardware Newsletter

RDNA 4 GPU SPECIFICATIONS

RX 9000-SERIES PRICING

RDNA 4 RELEASE DATES

RDNA 4 CORE GPU ARCHITECTURE

RDNA 4 OTHER ARCHITECTURAL IMPROVEMENTS

STICKING WITH GDDR6 VRAM

MONOLITHIC GPUS, BUILT ON TSMC N4P

NO “MADE BY AMD” REFERENCE CARDS

FSR 4 AND HYPR-RX

DRIVERS AND SOFTWARE

THE GPU LANDSCAPE

MOST POPULAR

Graphics Card

RX 9070 XT

RX 9070

RX 9060 XT?

RX 9060?

Architecture

Navi 48

Navi 48

Navi 48?

Navi 44?

Process Technology

TSMC N4P

TSMC N4P

TSMC N4P

TSMC N4P?

Transistors (Billion)

53.9

53.9

53.9

22?

Die size (mm^2)

356.5

356.5

356.5

153?

SMs / CUs / Xe-Cores

64

56

32?

20?

GPU Shaders (ALUs)

4096

3584

2048?

1280?

Tensor / AI Cores

128

112

64?

40?

Ray Tracing Cores

64

56

32?

20?

Boost Clock (MHz)

2970

2520

2790?

2700?

VRAM Speed (Gbps)

20

20

20?

20?

VRAM (GB)

16

16

12?

8?

VRAM Bus Width

256

256

192?

128?

L2 / Infinity Cache

64

64

48?

32?

Render Output Units

128

128

96?

64?

Texture Mapping Units

256

224

128?

80?

TFLOPS FP32 (Boost)

48.7

36.1

22.9?

13.8?

TFLOPS FP16 (FP4/FP8 TFLOPS)

389 (1557)

289 (1156)

183 (731)?

111 (442)?

Bandwidth (GB/s)

640

640

480?

320?

TBP (watts)

304

220

150?

120?

Launch Date

Mar 2025

Mar 2025

Apr–Jun 2025?

Apr–Jun 2025?

Launch Price

$599

$549

$399?

$299?

AMD’s Radeon RX 9070 and 9070 XT start at $549, ship March 6th

Image

AMD is finally ready to compete with Nvidia’s new RTX 50-series GPUs.

by Sean Hollister and Tom Warren

AMD’s answer to Nvidia’s RTX 50-series GPUs is arriving next week — and its new graphics cards are aggressively priced against Nvidia’s $749 GeForce RTX 5070 Ti and $549 RTX 5070. The AMD Radeon RX 9070 and 9070 XT ship on March 6th for $549 and $599, respectively, one day after the RTX 5070 arrives.

AMD says they both offer “4K gaming at a 1440p price,” though it’s making some unusual comparisons to do so — according to AMD, the 9070 XT is 51 percent faster on average than a four-year-old RX 6900 XT at 4K and max settings and 26 percent faster than a four-year-old RTX 3090 at the same settings, cards that filled a pricey niche two generations ago.

While leaks had suggested the pricing of AMD’s 9070 could start at $649, AMD surprised everyone today with suggested retail prices at $549 for this card — putting it head-to-head with Nvidia’s $549 RTX 5070. The $599 9700 XT could even challenge Nvidia’s $749 pricing for its RTX 5070 Ti, as long as AMD has managed to deliver performance that can challenge both of these RTX 50-series cards.

We won’t know the full answer to that until reviews for AMD’s cards appear next week as well as reviews for Nvidia’s RTX 5070. AMD is dropping some performance hints about where its 9070 series cards might fit in, though. AMD says the 9070 XT is 42 percent more potent at 4K Ultra (and 38 percent at 1440p Ultra) than AMD’s RX 7900 GRE, a card that challenged the RTX 4070 at $549 but vanished after less than a year. And it’s “just barely slightly slower” than a 7900 XTX, AMD’s 2022 flagship, we heard on a call.

The vanilla RX 9070, meanwhile, is 21 percent faster than the same 7900 GRE at 4K, suggesting it’s not far off the XT in performance: 38 percent faster than the AMD 6800 XT and 26 percent faster than the Nvidia RTX 3080, according to AMD. At 220 watts of board power, AMD’s director of graphics product management, Scott Olschewsky, says the 4nm monolithic chip is “the most efficient GPU we’ve ever built.”

Each card also offers 16GB of GDDR6 memory, DisplayPort 2.1a and HDMI 2.1b ports, and are PCIe 5.0 cards — not that your motherboard needs a 5.0 new slot or would necessarily benefit from one. They will use standard 8-pin PCIe power connectors.

While both the 9070 and 9070 XT have far fewer graphics compute units than the 7900 GRE (55, 64, and 80, respectively), they don’t need as many because they’re the first cards to ship with RDNA 4, which AMD says offers twice the rasterization (read: non-ray traced) graphics performance per compute unit as RDNA 2. (RDNA 3 was closer to 1.4x that of RDNA 2.)

AMD says it’s taken larger steps than that in ray tracing, too, and almost fully doubled its FP16 machine learning performance since RDNA 3, and up to 779 TOPS, though performance will differ from app to app: AMD’s only seeing 12 percent more performance in Adobe Lightroom super resolution over the 7900 GRE, as one example.

But even if you’re not playing with AI very intentionally, AMD will put that hardware to use in gaming with FSR 4, which comes with a new AI upscaling algorithm for its super resolution tech that’ll exclusively run on these RDNA 4 and newer cards.

Olschewsky says FSR 4 can increase frame rates “while looking just as good as native rendering” at 4K resolution. AMD used some images from Warhammer 40,000: Space Marine 2 — a title that averages just 53fps at 4K Ultra settings — to show more fine detail in the distant background of an image while averaging at 182fps with FSR 4 and frame generation turned on.

And while AMD didn’t share expected performance for that game without the fake frames added, it did break out the differences for a handful of other games like Call of Duty: Black Ops 6 and Ratchet & Clank:

AMD says it’ll have 30 FSR4 games at launch and over 75 by the end of the year — one obvious theme is a lot of PS5 games, making us wonder how closely the PS5 Pro’s “PlayStation Spectral Super Resolution,” an AI upscaler powered by AMD hardware, might be related to it. But AMD also says FSR 3.1 games can be updated to FSR 4 “with a simple toggle.”

The new cards also have an enhanced media engine with higher image quality for gameplay recording, as the last one “did not meet quality expectations from the gamers trying to record and stream.” AMD says its Fluid Motion Frames (AFMF), its generic fake frame generation you can apply to any game yourself, is being upgraded to AFMF 2.1, with less ghosting and smearing of details.

As with any graphics card these days, the actual amount you pay will be determined by whether AMD can actually make enough of the things, and if AMD’s board partners don’t jack up the price. Olschewsky had some fighting words for Nvidia there, though: “We believe that our 9070 XT will be going toe-to-toe with what you’ve seen from the 5070 Ti that users may or may not be able to buy,” adding that the 9070 “is going to look very strong against their upcoming 5070.”

Unlike Nvidia, which warned of shortages with its new cards, he says AMD expects “strong availability at launch.”

“We are working with all of our partners to go and deliver the most competitive prices on GPUs around the world; our focus is ensuring cards hit target price points around the world. It’s difficult to predict what’s going to happen, but we’ll be working with partners as we see pricing show up.”

At the end of AMD’s event today the company also teased that its RX 9060 series cards will arrive in the second quarter to “broaden the family even further.” AMD plans to disclose more information about these cards “later,” but it certainly looks like it’s getting ready to challenge Nvidia’s unannounced RTX 5060 cards.

A weekly newsletter by David Pierce designed to tell you everything you need to download, watch, read, listen to, and explore that fits in The Verge’s universe.

© 2025 VOX MEDIA, LLC. ALL RIGHTS RESERVED

MOST POPULAR

Installer

More in News

Top Stories