KaiserBecks
Member
Ryzen 7 3700X, 32gb ram, 3080ti, installed on a regular ssd
3440x1440. ultra settings
Game looks amazing and runs like a charm
3440x1440. ultra settings
Game looks amazing and runs like a charm
What frames are you getting? I'm thinking about not even playing it on mine if it's as bad as the benchmarksRyzen 7 3700X, 32gb ram, 3080ti, installed on a regular ssd
3440x1440. ultra settings
Game looks amazing and runs like a charm
Yep. People were mocking me when I said I was worried about how well it would run on PCThis game really lacks optimization.
Graphics don't look nowhere near good enough to justify these performance levels.
I would say that Bethesda has a long way to fix the performance on this game, but they are so lazy they will just leave it to modders to fix it.
Yeah that feels a little strange to me, your CPU is on the older side but it's high end for that gen and should be able to keep up.Hmm, not getting great performance on my system at 3440x1440p.
Averaging in the 40-60's outside (can dip down to 30 rarely). Runs a bit higher when indoors but not great. Also doesn't feel smooth at times even when it is running at higher FPS, like the frame pacing is off.
AMD Ryzen 3900x
ASUS Strix OC 3090 24GB
32GB DDR4 3600 Mhz
I know my CPU is getting on the older side but was hoping for better performance.
I have tried the DLSS mod and some of the suggested performance tweaks but am not seeing a whole lot of difference.
Apparently not. While Starfield favors AMD GPUs, it seems to like the current gen Intel CPUs over the current Ryzen line up. Also memory speed seems to matter a lot with this one.For Starfield, go with AMD. Gaming wise 7800x3d is best bet.
I thought I was getting crazy, but yeah, this game is stupid demanding for how it looks. I am getting with 2560x1440 through parsec around 60 FPS, maybe little bit more. I thought my HW 7950X3D and 4090 would smash this shit, but apparently not. Sucks because I have here Mac Pro and it allows 120FPS.Yep. People were mocking me when I said I was worried about how well it would run on PC
“Bethesda games run on a potato”
“My 1060 will get 60 fps”
Yeah…you mutherfuckers were dead wrong. I knew once it was limited to 30 fps on XSX that it would struggle on PC.
I have a 7800X3D/4090 and it performs fine with me, especially with the DLSS mod installed at 70% render resolution and I get around 70-90 fps, but I am not remotely close to most users and that sucks.
I haven't checked but it looks and plays much, much better than on Series X, that much I can say.What frames are you getting? I'm thinking about not even playing it on mine if it's as bad as the benchmarks
Interesting, saw some early reports that larger cache makes a big difference. I guess you never damn know on PC these days, lol.Apparently not. While Starfield favors AMD GPUs, it seems to like the current gen Intel CPUs over the current Ryzen line up. Also memory speed seems to matter a lot with this one.
I7 13700k
4070
32 gigs ddr5
No idea on how it runs as the thing is STILL downloading on windows store. Saying it now has just over 300gb to download!
Anyone with similar specs please let me know how it runs. Started it on my series x as I was away but home today and downloading now.
The game either have useless fuckign low or useless fucking ultra .
LMAO Starfield doesn't even use RT and it is outperformed on an RDNA2 AMD card with poor RT support by all these games which have RT what is this version of Creation Engine doing??
Best thing to do in these situations is just go to Youtube and search 'Starfield RTX2080'. You dont always get super accurate results as a lot of people dont know how to look after their pc's, but you will get an idea.How do you guys think an i79700k, 32 gbs of DDR4, and an RTX 2080 would handle it? I'm certain I could run it, but would it be a good experience?
well they have to compare it to other demanding games .Isn't it a bit odd to compare games that have been out for a while and have received numerous patches to a game in early access?
Well if you take something like The Last of Us on PC today compared to lauch, its practically like a different game in terms of performance.well they have to compare it to other demanding games .
you think the day 1 patch will improve performance for starfield?
well they have to compare it to other demanding games .
you think the day 1 patch will improve performance for starfield?
How do you guys think an i79700k, 32 gbs of DDR4, and an RTX 2080 would handle it? I'm certain I could run it, but would it be a good experience?
you might want to double check it's actually running at the advertised speed because a lot of the time unless you configure an XMP profile in the BIOS it will run at the default non XMP speed.Glad I opted for 6000mhz DDR5 with my new rig
you might want to double check it's actually running at the advertised speed because a lot of the time unless you configure an XMP profile in the BIOS it will run at the default non XMP speed.
in my cases with my past 2 asus motherboards i ran into issues with XMP where it would blue screen windows when i tried to run the ram at full speed even with an XMP profile enabled. iirc i had to manually configure XMP profiles with the voltage and clock speeds instead of using the presets.
"wmic memorychip getspeed" is a handy command you can use to have windows report the actual speed your ram is running. i actually forgot i fixed my XMP issue and was running 5600Mhz but checked the other day and I'm actually running 6400Mhz.
it was supported as it's pretty generic (g skill trident rgb 6400) and a widely used motherboard (rog z790-f) but the default motherboard ram profile didn't work and either blue screened windows or did work, but wasn't running the speed it was supposed to. only when i manually edited the voltage and clock speed values to match those listed n the ram sticks did it work.Thanks for the heads up but yes, XMP (actually for me it's EXPO as I'm running an AMD chip) is enabled and the RAM is running at full speed.
It's important to check your motherboard model's QVL to see a list of officially supported RAM for XMP/EXPO usage before buying RAM. I'm guessing that's why you ran into issues with XMP in the past; the memory you purchased probably wasn't on your motherboard model's QVL.
it was supported as it's pretty generic (g skill trident rgb 6400) and a widely used motherboard (rog z790-f) but the default motherboard ram profile didn't work and either blue screened windows or did work, but wasn't running the speed it was supposed to. only when i manually edited the voltage and clock speed values to match those listed n the ram sticks did it work.
the reason i bring it up is the situation where it didn't blue screen and was running a slower speed unbeknownst to me. it's possible that just being listed as compatible isn't enough and you can verify that with the wmic command
VRR/G-Sync helps a lot. That way frame dips don't look as stuttery. The game swings between 60 and 100 fps for me and the difference between those is noticeable but both are smooth and those swings aren't abrupt, it's more like the difference between open and indoor areas.I decided to lock it to 30fps with motion blur on medium (i feel like i'm back in my console days lol.
I honestly dont know how people prefer flucturating framerate over a consistent one,
That is really badWas watching the HW Unboxed benchmark, in 1080p ultra preset they get 48fps with my 3080 Ti.
Fired up Cyberpunk 2077 to see how it performs in comparison, in most demanding downtown area of Night City:
1080p ultra no RT - 110fps - yes, over twice as fast than Starfield
1080P ultra RT Ultra - 65fps - still 20 fps over Starfield, with RT Ultra!
1080p ultra RT Pathtracing - 35fps - just 13 fps less than Starfield
This is hilarious. And of course Cyberpunk looks much better graphically, even without RT, and takes place in city that's what, 20 times the size of New Atlantis in Starfield?
Since everything in Starfield is segmented into cells separated by loading screens, how the fuck and why the fuck is it so demanding? Keeping count of dropped sandwiches can't be that difficult.
Maybe at Medium settings? Watch that Hardware Unboxed video. You will have dips under 60 for sure.Any chance I can run this at 60fps? I have an
i7 4790k
2070 super
32gb ddr3
and an SSD.
I play at 1080p. Hoping there's no CPU bottleneck
Running at 67% resolution scale and I think Profile C.Yeah that feels a little strange to me, your CPU is on the older side but it's high end for that gen and should be able to keep up.
When you use the DLSS mod, what resolution scale are you using?
60fps in some "dungeon" or barren planet maybe , outside of new atlantis 40-50fps at low crowd density . Yea , cpu and ram bottleneck .Any chance I can run this at 60fps? I have an
i7 4790k
2070 super
32gb ddr3
and an SSD.
I play at 1080p. Hoping there's no CPU bottleneck
Thanks! I'll hold off until my next upgrade(s).60fps in some "dungeon" or barren planet maybe , outside of new atlantis 40-50fps at low crowd density . Yea , cpu and ram bottleneck .
LMAO Starfield doesn't even use RT and it is outperformed on an RDNA2 AMD card with poor RT support by all these games which have RT what is this version of Creation Engine doing??
Any chance I can run this at 60fps? I have an
i7 4790k
2070 super
32gb ddr3
and an SSD.
Hard CPU bottleneck. You will not get a stable 60, if you can get a stable 30 you you should lock your framerate and count it as a win, since you are way below min spec CPU.Any chance I can run this at 60fps? I have an
i7 4790k
2070 super
32gb ddr3
and an SSD.
I play at 1080p. Hoping there's no CPU bottleneck
so ive got an i5 6600k running at 4.4ghz, which is below minimum.
however I have more than enough RAM and a 3060 RTX.
will my proc. be too much of a bottleneck?
you will most likely have <30 FPS drops in cities. but most of the time it should to okay with a 30 or 40 FPS lock. practically like Steam Deck (deck users falsely thinks game runs fine in empty regions or spaceship)so ive got an i5 6600k running at 4.4ghz, which is below minimum.
however I have more than enough RAM and a 3060 RTX.
will my proc. be too much of a bottleneck?
so ive got an i5 6600k running at 4.4ghz, which is below minimum.
however I have more than enough RAM and a 3060 RTX.
will my proc. be too much of a bottleneck?
Bethesda lists a 6800K as a minimum. But consider that the 6800K is a Broadwell CPU. This is basically a Haswell architecture with a 14nm process node. Yours is a Skylake.
That CPU has 2 more cores and 4 more threads than yours. But yours has higher clock speed and better IPC.
On a negative side, yours uses slower DDR4, unless you OC.
Your CPU is probably the bare minimum to get around 30 fps.
The i5-6600K only has four threads, no hyperthreading which at the time was restricted to i7. The 6800K is 6C/12T.
I played about 5hrs and it worked fine for a few hours, but the more I started fast traveling and going in between loading screens it started to freeze the game and require me to crash out to desktop. I had a 5 freezes and crashes. Disabling the mod either in the menu or using the '*' key allowed me to proceed past those parts, but I can't keep playing like that. DLSS3 also causes issues with distance star clusters you can see from a planet's surface. Especially if they a lot of space gas around the cluster/galaxy. It will flicker a black stencil, maybe from an alpha mask the code is seeing and having trouble dealing with it. I don't know. Unfortunately, I had to give up on the mod. In it's current state it's broken.Tried the DLSS 3 injector and wow that guy deserves his $5. Using DLAA so much better than native quality and hitting 120fps minimum and mostly maxed at the 138fps Reflex caps you at.
Fake frames or not I can't tell thanks to the injected Nvidia Reflex.
Maybe there are some horrible issues in other parts of the game but been playing for hours and had no problems.
Toggling between FSR and DLAA in real time and seeing how much better DLAA looks is honestly embarrassing for AMD.
Oh bother. Well that sucks.I played about 5hrs and it worked fine for a few hours, but the more I started fast traveling and going in between loading screens it started to freeze the game and require me to crash out to desktop. I had a 5 freezes and crashes. Disabling the mod either in the menu or using the '*' key allowed me to proceed past those parts, but I can't keep playing like that. DLSS3 also causes issues with distance star clusters you can see from a planet's surface. Especially if they a lot of space gas around the cluster/galaxy. It will flicker a black stencil, maybe from an alpha mask the code is seeing and having trouble dealing with it. I don't know. Unfortunately, I had to give up on the mod. In it's current state it's broken.
It was great for me up until that started happening. When I used it at first I wasn't doing a bunch of gravity jumps and fast travels all over the place. I would say to keep using it until it crashes. It's possible it will work fine for you. I'm not the only person to experience it. There's a bunch of other users on the Patreon reporting it. It just came out like 24hrs ago, to be fair, but to be fair I also paid 24hrs ago. Maybe on my next playthrough it will be stable. It would be even better if Bethesda implemented it natively, but I don't believe that will ever happen.Oh bother. Well that sucks.