• Hey Guest. Check out your NeoGAF Wrapped 2025 results here!

007 First Light recommended specs revealed and... oof.

god i remember being able to play diablo 3 near launch on an dell ultrasharp 10bit screen in true 1440p and it fucking blasting my eyes in highres. and now we have fake res and fake frames
I mean, that was made deliberately to run great on the lowest end PCs, of course you could max it easily and of course it looked like a low end, low fidelity game even in high res.

Basically they adopted the successful WoW approach (even though it wasn't an MMO and came out 7-8 years later). The equivalent of random shooters looking like Fortnite now.
 
Last edited:
If you believe that a heavy scripted demo is a 'gameplay'... :-D
lmao. You clearly didnt watch the demo then because the last ten minutes showcase another level played 3 different ways to showcase just how different each playthrough can be.

i find it hilarious that the devs put out a 35 minute gameplay demo with multiple levels, showcasing several different playstyles, not to mention all white leads and people didnt even bother watching. at this point i am convinced that gaming has been hijacked by a bunch of sad angry little men who dont care about visuals, game design, physics, or storytelling and just want to bitch and moan like a woman PMSing all day.
 
Last edited:
1920x1080p is still the most common display resolution on steam surveys. By not making it too heavy on the hardware, more people should be able to get a decent performance > more sales.

IMG-2282.png


 
This has to be the longest "gap" we have ever had when it comes to Bond game releases....if it flops....expect it to go the same way as the Dead Space remake....(ie hits any momentum in reviving the franchise..)
 
this is true. but AMD had affordable 2x better than console GPUs out in 2020. But people wanted nvidia and refused to buy them. Instead going for 8GB gpus that we all said were garbage even back then. now its coming back to bite them.

absolutely not. Nvidia cards are aging like fine wine, while AMD cards are rotting away.
2020 Nvidia cards still outperform 2025 AMD cards in some RT heavy games, while 2020 AMD cards are completely unable to keep up whatsoever.

I honestly feel bad for people who still run an RDNA2 GPU in their PC. I'd rather have a high end RTX20 series card from 2018 than an RDNA2 card.
 
absolutely not. Nvidia cards are aging like fine wine, while AMD cards are rotting away.
2020 Nvidia cards still outperform 2025 AMD cards in some RT heavy games, while 2020 AMD cards are completely unable to keep up whatsoever.

I honestly feel bad for people who still run an RDNA2 GPU in their PC. I'd rather have a high end RTX20 series card from 2018 than an RDNA2 card.
Depends on the games. Only some RT games favor Nvidia GPUs. UE5 hardware RT, Anvil Hardware RT, Dunia RT, RE Engine RT and Snowdrop hardware RT perform roughly the same on AMD and Nvidia GPUs. Maybe a 10% improvement in some engines? Definitely not the 80-100% we saw in early ray tracing titles. Games like Cyberpunk, Control, Alan Wake were sponsored and had a lot of their RT tech implemented with the help of nvidia engineers who clearly didnt bother optimizing for AMD GPUs, but other engines show that AMD cards are just fine at doing hardware RT. Especially since Ubisoft, Epic and Capcom designed their RT implementations around console tech.

Path tracing obviously is impossible on these cards, but lets face it, no one is doing path tracing on anything below a 3090 when it comes to 2020 era GPUs. i was straight up locked out of PT on my 10 GB 3080 and could only do 30 or so FPS in Doom. hell i just bought a 5080, and i have to drop down to 1440p and use DLSS just to get close to 60 fps in Doom and Indy, and even then in some areas thats not enough.

Meanwhile, the VRAM issue impacted A LOT more games early on and many users on this very board complained about having to reduce textures and details even in cross gen games on the 8GB 3070s. Forget the 3060 or 4060. Even i ran into a vram issue in RE4 and Gotham Knights with RT on. RE4 would literally crash while GK would crawl to 2-5 fps. leading me to turn off ray tracing altogether.
 
i said this in the other thread, but that writer shouldve never opened his mouth and said what he said. you could do everything right in today's society, but if you say one mildly concerning thing, the wolves come for you. And sadly, the sheep follow.

This game has a straight white male lead, a straight white attractive female bond girl, shows Bond as a playboy ogling bikini clad women on beaches, hanging out on yachts with scantily clad women, ray traced graphics, incredible physics and destruction finally using the next gen CPUs, great new melee combat animations, and all of that is ignored because the writer hinted at making the game for a modern audience.

Devs need to understand that the very gamers they are trying to sell this game to are extremely sensitive and insecure, extremely jaded, ready to pounce on every word, ready to dismiss every game if it does not appeals to their own sensibilities. We have to treat them with kids gloves and give them absolutely fuck all to complain about because they do make up the vast majority of gamers. This game just might end up flopping because of that interview, and all of the incredible work done by engineers, artists and level designers will have gone to waste. Everything even remotely negative about this game will now be viewed with that lens.
Gamers are by far the most sensitive and pathetic group of fans on earth out of every entertainment medium I would say.

Let a game have the slightest little thing that isn't meticulously tailored to them and their exact tastes and it's utter bitching and moaning, review bombing, death threats, dev hate, fake outrage, crybaby nonsense for the games entire life span.

It's no wonder no one takes games seriously other than gamers. 50 year old hobby and still viewed as a kids thing by most.
Devs trying to make graphics impress by turning up chromatic aberration

Crank It Up GIF by Zypto


Not impressed
Show me one game in the same genre that looks better.
 
Depends on the games. Only some RT games favor Nvidia GPUs. UE5 hardware RT, Anvil Hardware RT, Dunia RT, RE Engine RT and Snowdrop hardware RT perform roughly the same on AMD and Nvidia GPUs. Maybe a 10% improvement in some engines? Definitely not the 80-100% we saw in early ray tracing titles. Games like Cyberpunk, Control, Alan Wake were sponsored and had a lot of their RT tech implemented with the help of nvidia engineers who clearly didnt bother optimizing for AMD GPUs, but other engines show that AMD cards are just fine at doing hardware RT. Especially since Ubisoft, Epic and Capcom designed their RT implementations around console tech.

Path tracing obviously is impossible on these cards, but lets face it, no one is doing path tracing on anything below a 3090 when it comes to 2020 era GPUs. i was straight up locked out of PT on my 10 GB 3080 and could only do 30 or so FPS in Doom. hell i just bought a 5080, and i have to drop down to 1440p and use DLSS just to get close to 60 fps in Doom and Indy, and even then in some areas thats not enough.

Meanwhile, the VRAM issue impacted A LOT more games early on and many users on this very board complained about having to reduce textures and details even in cross gen games on the 8GB 3070s. Forget the 3060 or 4060. Even i ran into a vram issue in RE4 and Gotham Knights with RT on. RE4 would literally crash while GK would crawl to 2-5 fps. leading me to turn off ray tracing altogether.

your inability to correctly set your settings is not an issue of the card.

RE4 runs absolutely fine with 8GB and RT. you just have to correctly set the texture pool setting.
lowering it will only marginally affect the actual quality of textures, and will maybe result in some texture pop-in, but that is usually very hard to notice.

as long as you use any variant of the "High" setting, you'll probably not notice any real difference unless you look at it under a magnifying glass.

so on an 8GB card, go down to "High 0.5GB"... it will maybe use 4GB of memory total at that point. giving you lots of headroom to turn on RT.
 
Last edited:
your inability to correctly set your settings is not an issue of the card.

RE4 runs absolutely fine with 8GB and RT. you just have to correctly set the texture pool setting.
lowering it will only marginally affect the actual quality of textures, and will maybe result in some texture pop-in, but that is usually very hard to notice.

as long as you use any variant of the "High" setting, you'll probably not notice any real difference unless you look at it under a magnifying glass.

so on an 8GB card, go down to "High 0.5GB"... it will maybe use 4GB of memory total at that point. giving you lots of headroom to turn on RT
Tried that. I thought the difference in textures was too high when dropping the settings.
 
Tried that. I thought the difference in textures was too high when dropping the settings.

as long as you are on High the game will load the same quality of textures on any of the High settings. it will simply not pre-cache them and that can result in pop-in or far away textures lookin slightly worse.

there's no way it looked bad enough to not consider it.


and as to "being locked out of path tracing"... dude... You're locked out of path tracing on all AMD cards if you want playable framerates.

sure it sucks that some GPUs have not enough memory to enable path tracing in some games... but that's every AMD card! even the 9000 series cards get stramrolled by mid range Nvidia cards from 2020 when enabling path tracing.
 
Last edited:
End of the day, I can't take this avatar seriously as 007. I don't want to control that character for a game, pretending he's supposed to be James Bond. Suspension of disbelief will be nonexistent – they've made this game for the Tom Holland generation but it holds no appeal to me.

Should have just called it Kingsman instead, I'd have loved that tbh.
And Tom Holland gen couldn't give two shits about 007... the concept of target audience/demographic its truly lost
 
as long as you are on High the game will load the same quality of textures on any of the High settings. it will simply not pre-cache them and that can result in pop-in or far away textures lookin slightly worse.
Yes now that you mention it, it was the LOD pop-in that made me switch to the higher res textures. much more noticeable than some ray traced reflections.
there's no way it looked bad enough to not consider it.


and as to "being locked out of path tracing"... dude... You're locked out of path tracing on all AMD cards if you want playable framerates.

sure it sucks that some GPUs have not enough memory to enable path tracing in some games... but that's every AMD card! even the 9000 series cards get stramrolled by mid range Nvidia cards from 2020 when enabling path tracing

no, its not just vram. these 2020 GPUs just dont have the raw horsepower to run path tracing at playable framerates. i just told you that i dropped settings on my 5080 and there were some levels that were not hitting 60 fps in doom and indy even using dlss at 1440p. i am sure if i went down to medium or low settings, i might be able to run them at 1440p dlss performance or 720p but i didnt spend over a thousand bucks upgrading from a $850 GPU just so i can play games at low to medium settings with path tracing.

cyberpunk is the exception. it ran well at 1440p dlss performance even on my 3080, but 3060 or 3070? highly doubt it. especially since its very heavy on the CPU and most of those gamers paired up those GPUs with zen 2 CPUs that just cant handle path tracing or anything remotely CPU heavy.

point is that it doesnt really matter if AMD GPUs cant do path tracing when almost all of the Nvidia GPUs just dont have the raw horsepower to run them. Especially the ones from 2020 that people refuse to upgrade.
 
RIP consoles
RIP those user reviews. Alienating consoles is already a slippery slope. But demanding those specs to pc users who we already find out are still using older graphics cards and setups is dumb as well. Another incoming game not optimized will hopefully lead to less developers doing this. Doubtful. But we can hope, lol.
 
absolutely not. Nvidia cards are aging like fine wine, while AMD cards are rotting away.
2020 Nvidia cards still outperform 2025 AMD cards in some RT heavy games, while 2020 AMD cards are completely unable to keep up whatsoever.

I honestly feel bad for people who still run an RDNA2 GPU in their PC. I'd rather have a high end RTX20 series card from 2018 than an RDNA2 card.

Raster performance and VRAM of RDNA2 is absolutely fine but AMD wanted to drop diver optimizations for it two months ago (or so). AMD fanboys always talked about fine wine but that was true for ~8 years of GCN. Now AMD doesn't give a fuck anymore...
 
The recommended specs are essentially console-equivalent GPUs from 2020.

If your PC is below the capabilities of a PS5, I don't think you're looking at this game.
 
These dumpster fire devs are going to have to start optimizing ram usage again because they have got super fucking lazy in the last 5 years.
 
Looks great to me. Hitman engine is pretty lean. 007 is surely gonna have fuck tons of assets loaded for the entire runtime. 1660 gang is back in the game.
 
Top Bottom