• Hey, guest user. Hope you're enjoying NeoGAF! Have you considered registering for an account? Come join us and add your take to the daily discourse.

NVIDIA RTX 4090 is 300% Faster than AMD’s RX 7900 XTX in Cyberpunk 2077: Phantom Liberty Overdrive Mode, 500% Faster with Frame Gen

SolidQ

Member
but it's RT using voxels. So it's probably more similar to Crytec's Svogi.
Each frame of the scene was pre-computed into a baked voxel blob that we could re-light and move around in real time. (c)

Those RT in C2077, reminds me this times
aec93706c58737ca73c92fd93aec7d61.png
d0b8fe152c529608607154593c0fe370.png
a45454761222caa60642d24b235b6afb.png
 
Last edited:

RoadHazard

Gold Member
Looking at the bar graphs, they're lying in the article, or they don't understand how percentages work. What they mean is "300% as fast" (or "3x as fast") etc, which is equivalent to 200% faster, not 300% faster.

In case anyone wants to argue this with me: If one card does 100fps and the other does 50fps, that means the first card is 200% (2x) as fast as the second one, but only 100% faster.

Actually they even mix these two ways of counting, making it even more misleading, as in:

"The RTX 4070 is nearly 60% faster than the RX 7900 XTX, while the 4080 is 240% faster with Ray Reconstruction enabled."

The accurate statement here (as the "nearly 60%" is correctly calculated) would be:

"The RTX 4070 is nearly 60% faster than the RX 7900 XTX, while the 4080 is nearly 140% faster with Ray Reconstruction enabled."

Might sound nitpicky, but it makes quite a big difference in understanding the differences.

But yeah, either way the Nvidia advantage is massive obviously.
 
Last edited:

shamoomoo

Banned
There is the matter of "if". As there was no GPU at the time that could do RT.
But even if there was, it would be very bad. Even Turing that released in 2018 had limited RT capabilities. Now imagine what a GPU in 2012 would be cable off, regarding RT.
Maybe Imagination Technologies would've been the only option if AMD didn't have anything.
 

JohnnyFootball

GerAlt-Right. Ciriously.
Hmm, yes, I hadn't thought of it from that angle. Nvidia would pretty much have total domination of gaming if that were the case and that wouldn't be good.

I do hope AMD are able to significantly improve their ray-tracing for the next consoles. Hopefully Sony & Microsoft can help them with that.
nvidia certainly has nice tech, but putting it in all consoles would mean $1000 consoles and $2000 GPUs. Cheap consoles have been a huge help in preventing GPU prices from getting much worse than they already are. Trust me, it could get worse when you consider how little nvidia actually cares about the GPU market.
 

XesqueVara

Member
They are a generation behind with Alchemist, which you could say was a proof of concept.
The A770 keeps up with the 3060Ti and 6700XT......all of them are approx 400mm².
The A770 spanks the 6700XT in RT and upscaling tech.
Intel is approx 2 years behind Nvidia......AMD is 5 or more in term of tech and die usage.
The A770 barely keeps up with the 3060 and the 6600XT in Average utilizing ~2x more area and more power than both, It just reaches the performance of the 3060ti in some games.
Their GPU IP is still way behind of Nvidia and AMD, Gen 12 is... zzz
 
nvidia certainly has nice tech, but putting it in all consoles would mean $1000 consoles and $2000 GPUs. Cheap consoles have been a huge help in preventing GPU prices from getting much worse than they already are. Trust me, it could get worse when you consider how little nvidia actually cares about the GPU market.
The cheap consoles perform like shit in RT and these days are actively holding PC gaming back because games are designed for the shit performance of the consoles
 

Three

Gold Member
The harsh reality is that there is absolutely nothing they can do because they are relying on AMD for that.

Lots of wishful thinking around on how Sony will use some ultra secret mega sauce and fix the situation with the “pro” version of the PS5, not to mention the 6.. but is not going to happen. Some of the devs might be able to remedy the situation, but what you see here with the 4090 is not coming to consoles for at least 10-15 years.
Weird that you replied to somebody mentioning MSFT and Sony (ie consoles) and concentrated so heavily on the discussion of RT on PS pro. Who has said path tracing will be remedied with new consoles or the pro?

The answer is that they won't use the same technology on console but still have RT and that itself is why there are so few games with path tracing to begin with, you can count them with half a hand. You should be hoping that path tracing does come to console sooner rather than later otherwise you will just get games using alternatives to path tracing like lumen which perform just fine on consoles and AMD. I get that to some the urge to be better than others outweighs everyone being better together but this isn't something you should be celebrating.
 
Last edited:

winjer

Gold Member
Weird that you replied to somebody mentioning MSFT and Sony (ie consoles) and concentrated so heavily on the discussion of RT on PS pro. Who has said path tracing will be remedied with new consoles or the pro?

The answer is that they won't use the same technology on console but still have RT and that itself is why there are so few games with path tracing to begin with, you can count them with half a hand. You should be hoping that path tracing does come to console sooner rather than later otherwise you will just get games using alternatives to path tracing like lumen which perform just fine on consoles and AMD. I get that to some the urge to be better than others outweighs everyone being better together but this isn't something you should be celebrating.

Consoles already have hardware that can do ray-tracing and path-tracing. The problem is performance, as path-tracing cast more rays and more bounces, even using the Monte Carlo simulation to randomize and optimize the amount of rays cast.

Lumen on consoles does not use path-tracing. Nor ray-tracing. It uses SDF's that are a very simplified and much less performance intensive way.
 

ShaiKhulud1989

Gold Member
Cool to know that $1000+ GPU is doing fine.

Now, could Nvidia fix at least basic RT FPS on my simpleton 3080, because after 2.0 everything runs worse with any settings, RT or not.
 

//DEVIL//

Member
It runs 8K 120 fps ray traced on Nvidia!*

*with DLSS lowering internal resolution to 360p and applying fake frames
Who gives a fuck what technology they used to give you these frames ? Call them fake frames call them cheating. It enhance the experience motion and make it feel and play like a 120 or whatever high fps it can reach.

People calling it weird names is just bitterness and salt to a high level.

As far as lowering internal resolution, when a Game running on DLSS has better smoothness to the edges and less antialaising than native in many games, then I'll take it.

This is just a sad post to be honest
 

Three

Gold Member
Consoles already have hardware that can do ray-tracing and path-tracing. The problem is performance, as path-tracing cast more rays and more bounces, even using the Monte Carlo simulation to randomize and optimize the amount of rays cast.
Exactly, a lot of console games have raytracing already and next gen and the pro consoles will remedy the performance to make it easier to have RT by default. Nobody has said the pro will remedy path tracing though.
Lumen on consoles does not use path-tracing. Nor ray-tracing. It uses SDF's that are a very simplified and much less performance intensive way.
Yeah, that's my point, they will use more performant alternatives to path tracing to achieve realistic lighting meaning fewer games get actual path tracing if consoles don't support it at acceptable framerates. He shouldn't be celebrating that. If consoles right now were capable of path tracing at good framerates we would have a lot more games with path tracing for PC too.
 
Last edited:

k_trout

Member
Each frame of the scene was pre-computed into a baked voxel blob that we could re-light and move around in real time. (c)

Those RT in C2077, reminds me this times
aec93706c58737ca73c92fd93aec7d61.png
d0b8fe152c529608607154593c0fe370.png
a45454761222caa60642d24b235b6afb.png
jeeze i bought a 5800ultra, worst card I have owned, dx9 perf was just terrible, next card I had was the 6800ultra which was a whole different story
 

twilo99

Member
Weird that you replied to somebody mentioning MSFT and Sony (ie consoles) and concentrated so heavily on the discussion of RT on PS pro. Who has said path tracing will be remedied with new consoles or the pro?

The answer is that they won't use the same technology on console but still have RT and that itself is why there are so few games with path tracing to begin with, you can count them with half a hand. You should be hoping that path tracing does come to console sooner rather than later otherwise you will just get games using alternatives to path tracing like lumen which perform just fine on consoles and AMD. I get that to some the urge to be better than others outweighs everyone being better together but this isn't something you should be celebrating.

It’s regarding both consoles since they both use AMD hardware. The PS5 pro is more relevant only because it will be released sooner with newer hardware.. it will definitely have better RT performance, but nowhere near what you are seeing here with the 4090.
 

Klosshufvud

Member
Who gives a fuck what technology they used to give you these frames ? Call them fake frames call them cheating. It enhance the experience motion and make it feel and play like a 120 or whatever high fps it can reach.

People calling it weird names is just bitterness and salt to a high level.

As far as lowering internal resolution, when a Game running on DLSS has better smoothness to the edges and less antialaising than native in many games, then I'll take it.

This is just a sad post to be honest
I just don't enjoy DLSS in motion. For the games I've tested, it has introduced weird motion artifacts hard to put in words. I'll argue this sort of development is all wrong. Instead of optimizing their codes/engines for higher resolutions, devs now can just rely on these crutches to carry their asses at 1080p resolution. It's a sad development.
 

//DEVIL//

Member
I just don't enjoy DLSS in motion. For the games I've tested, it has introduced weird motion artifacts hard to put in words. I'll argue this sort of development is all wrong. Instead of optimizing their codes/engines for higher resolutions, devs now can just rely on these crutches to carry their asses at 1080p resolution. It's a sad development.
Asking game devs to better optimize their shit game is a totally different subject and I agree with. You having DLSS feedback you don't like it's you and your opinion only. Even die hard AMD fans ( not saying you are ) agree that DLSS is amazing. And many games showed it run better than native when native is just TAA.
 
Top Bottom