G-Sync is the god-level gaming upgrade.

I have no stuttering when using Nvidia Shield streaming from my PC with a Gsync screen at 144 Hz

It goes 1440p 144hz -> 1080p 60hz

It somehow, works. I was expecting hell to be honestly.
 
I usually don't have any stuttering to my Steam link with my monitor set at 165Hz. but now and again, one game will be a pain.
 
By the way everyone:

The sweet spot for Gsync monitors are still ACER and ASUS offerings with the 1440p, IPS, 144hz+ right?

Going 4K, 21:9 or VA still brings drawbacks in certain areas right?
 
To what degree is G-sync worth it? I'm in the market for a new a monitor (for a new PC) and I'm debating between the C49HG90 which is a Samsung 49" super ultra wide monitor (32:9), 3840x1080, HDR and 144hz refresh rate, or one of the announced 35" G-Sync HDR 200hz 3440x1440 monitors from Acer/Asus that are coming out later in the year. I reckon both are bound to cost the same but I really like the form factor of the Samsung one and the familiarity that having 2 seamless 1920x1080 brings (I already have a 34" 3440x1440 display, but with no bells or whistles). One the major differences is G-Sync support with the Samsung doesn't have.

The computer that this will be paired with is a 7700K, 32 GB DDR4, 1080 Ti machine.
 
To what degree is G-sync worth it? I'm in the market for a new a monitor (for a new PC) and I'm debating between the C49HG90 which is a Samsung 49" super ultra wide monitor (32:9), 3840x1080, HDR and 144hz refresh rate, or one of the announced 35" G-Sync HDR 200hz 3440x1440 monitors from Acer/Asus that are coming out later in the year. I reckon both are bound to cost the same but I really like the form factor of the Samsung one and the familiarity that having 2 seamless 1920x1080 brings (I already have a 34" 3440x1440 display, but with no bells or whistles). One the major differences is G-Sync support with the Samsung doesn't have.

The computer that this will be paired with is a 7700K, 32 GB DDR4, 1080 Ti machine.

Get a monitor with Gsync. It really removes a burden of the user by making you not care about stuttering due vsync, tearing, and fluctuating framerates (if it's an high refresh screen) when between 60-120+hz

It's hard to explain without showing it hands on. The difference between a screen with it and without is big in many games. But it's not a visual experience, it's more of a responsiveness, feeling type of experience.
 
Get a monitor with Gsync. It really removes a burden of the user by making you not care about stuttering due vsync, tearing, and fluctuating framerates (if it's an high refresh screen) when between 60-120+hz

It's hard to explain without showing it hands on. The difference between a screen with it and without is big in many games. But it's not a visual experience, it's more of a responsiveness, feeling type of experience.

So is my understanding flawed in that the main benefit of G-Sync is between 40~ FPS and 60 FPS?
 
By the way everyone:

The sweet spot for Gsync monitors are still ACER and ASUS offerings with the 1440p, IPS, 144hz+ right?

Going 4K, 21:9 or VA still brings drawbacks in certain areas right?

Ehhh? Well, I don't know what any of that means. The only drawback to 21:9 I guess is if you just don't like 21:9. And 100hz < 144hz, I suppose.

So is my understanding flawed in that the main benefit of G-Sync is between 40~ FPS and 60 FPS?

Depends on the refresh rate too, doesn't it? I think the gsync range for the 100hz ultrawides is 30-100, which is fantastic. Personally, I'm going to love being able to uncap my framerate once I snag one of these monitors tonight (finally).
 
So is my understanding flawed in that the main benefit of G-Sync is between 40~ FPS and 60 FPS?

Yes. The benefit of gsync goes beyond those framerates. If you have a higher refresh rate monitor (as most gsync monitors are), you'll avoid the tearing between 60-144hz that you'd normally get if you couldn't achieve a locked vsynced 144 fps. It creates a beautiful buttery smooth highly responsive image at these framerates. 100-140hz is where these monitors really shine I think.
 
Guys I'm upgrading from a GTX 760 on a 32" Samsung Smart TV from like 2012 to a GTX 1080 with a G-Sync AOC Agon AG352UCG. Just waiting on everything to come in the mail.

Is my brain gonna explode
 
Yes. The benefit of gsync goes beyond those framerates. If you have a higher refresh rate monitor (as most gsync monitors are), you'll avoid the tearing between 60-144hz that you'd normally get if you couldn't achieve a locked vsynced 144 fps. It creates a beautiful buttery smooth highly responsive image at these framerates. 100-140hz is where these monitors really shine I think.

Exactly.
I find the range of 60-140 more relevant for Gsync than 30-60.
Because between 30-60 I can see when the framerate changes, even if it doesn't stutter
 
Anyone know what causes this? Only happens in dark scenes with black colours

20642186_662607453935676_33102994_o.jpg


20641482_662607457269009_1173059968_o.jpg


20629308_662607463935675_881074581_o.jpg
 
Anyone know what causes this? Only happens in dark scenes with black colours

Insufficient bit depth in the display chain at some point if it's not inherent in the source material. Most TN monitors are 6 bit with FRC if you have a TN panel which could be a culprit.
 
Insufficient bit depth in the display chain at some point if it's not inherent in the source material. Most TN monitors are 6 bit with FRC if you have a TN panel which could be a culprit.
It's the ASUS PQ248Q so it's TH I think
 
It's the ASUS PQ248Q so it's TH I think
If the options are the same as a PG348Q make sure it's in the Racing or sRGB modes. (Racing = sRGB with the picture controls unlocked)
The Scenery or Cinema modes use a low gamma which brightens things up and makes banding or compression artifacts noticeable.
Don't adjust the "desktop color settings" in the NVIDIA Control Panel or load an ICC profile.
I don't know if that will fix it, but it should give you the least amount of banding possible for that panel.
 
PG248Q is a 6-bit TN monitor. Unfortunately I think that's how the image is going to look. You might be able to improve the problem some by adjusting the picture controls but it's a known weakness of 6-bit panels.
 
If the options are the same as a PG348Q make sure it's in the Racing or sRGB modes. (Racing = sRGB with the picture controls unlocked)
The Scenery or Cinema modes use a low gamma which brightens things up and makes banding or compression artifacts noticeable.
Don't adjust the "desktop color settings" in the NVIDIA Control Panel or load an ICC profile.
I don't know if that will fix it, but it should give you the least amount of banding possible for that panel.
It's on racing mode. Have to use Nvidia colour settings though or the picture looks washed out
 
Got the Asus PG348Q.

Feels nice to finally not worry about tearing. Also, I purposefully cranked the res scaling in Battlefield 1 to try and get the FPS to dip below 60 and I didn't even notice it at all when that happened. And when the FPS does go into the 80-90 range, it feels freakin great. Not to mention way smoother in Windows.

Other than that, I think I actually liked the color calibration of the Dell U3415W better out of the box. Also, I'm not sure if all that is worth dropping $1000+ on a monitor. But I just couldn't deal with not having gsync anymore and I already spoiled myself with a 3440x1440 ultrawide.
 
What are you changing to make it not "washed out"?
Do you still get the banding if you reset the NVIDIA color controls?
Doesn't happen without NVIDIA colour settings or at least not as bad

But without using those colour settings it's too bright and washed out and thats with the monitor setup properly following a guide on tomshardware iirc
 
Got the Asus PG348Q.
Feels nice to finally not worry about tearing. Also, I purposefully cranked the res scaling in Battlefield 1 to try and get the FPS to dip below 60 and I didn't even notice it at all when that happened. And when the FPS does go into the 80-90 range, it feels freakin great. Not to mention way smoother in Windows.
Other than that, I think I actually liked the color calibration of the Dell U3415W better out of the box. Also, I'm not sure if all that is worth dropping $1000+ on a monitor. But I just couldn't deal with not having gsync anymore and I already spoiled myself with a 3440x1440 ultrawide.
The stock calibration for my PG348Q is fairly decent in the sRGB or Racing modes.
The only thing I had to do was make a minor adjustment to the white balance as it was slightly too red. (94 / 98 / 100 RGB to hit D65)
Gamma is a little low, measuring 2.1, so I do wish there were more calibration controls, but that seems to be the main thing that G-Sync displays are lacking.
Newer monitors seem to include gamma presets, but no actual 10pt gamma controls or LUT adjustments.

Yikes, well those settings are definitely causing banding.
I'm just confused as to why you would need settings like that. The reviews I've checked have all said that gamma is too high out of the box on this monitor, making things too dark - yet you're still making the image even darker with those adjustments.

According to the Tom's Hardware review:
http://www.tomshardware.com/reviews/asus-rog-pg248q-24-inch-180hz-g-sync-gaming-monitor said:
gamma1ojsaf.jpg

The initial gamma measurement indicates a contrast control that is set too high. It clips highlight information and makes darker content look too bright. This has a deleterious effect on perceived contrast. TN and IPS don't have the greatest black levels in general, but getting gamma right can go a long way towards creating the deep blacks and rich shadow detail we all crave.
gamma2eisd8.jpg

Changing the preset to Cinema fixes the shadow detail issue, but now the mid and bright tones are too dark and murky. Calibrating only the RGB sliders fixes the grayscale nicely but the trace is still much too dark, running at around a 2.7 average.
gamma30tstp.jpg

Thankfully Asus has included gamma presets. Selecting 1.8 will bring tracking down near 2.2 where it belongs. It's a pretty major shift in accuracy and one that has a significant effect on color saturation and luminance, which we'll show you below. There is excellent performance baked into the PG248Q. You just have to know how to unlock it.
gamma4vhsuy.jpg
Their recommendation is to use Racing Mode with contrast reduced to 42, and gamma set to 1.8
I'd also suggest adjusting the monitor tilt if you're viewing the screen at an angle; e.g. looking down on it.

EDIT: You replied as I was typing that up.
Doesn't happen without NVIDIA colour settings or at least not as bad
But without using those colour settings it's too bright and washed out and thats with the monitor setup properly following a guide on tomshardware iirc
What display are you coming from?
It sounds like you may not be used to a calibrated image, if the Tom's Hardware measurements are accurate.

If you want a "darker" image, use their settings but increase the gamma control on the display from 1.8 to 2.2 or 2.5. Don't mess with the NVIDIA color controls.
 
The stock calibration for my PG348Q is fairly decent in the sRGB or Racing modes.
The only thing I had to do was make a minor adjustment to the white balance as it was slightly too red. (94 / 98 / 100 RGB to hit D65)

I'm not saying the calibration is bad, just that the U3415W requires fewer tweaks to get it just right.

I just used the windows color calibration to create a slightly higher gamma profile.
 
As someone that's only planning on console gaming, should I invest in a G-sync monitor? I would be upgrading from a BenQ RL2455HT.
 
As someone that's only planning on console gaming, should I invest in a G-sync monitor? I would be upgrading from a BenQ RL2455HT.

Nope. Wait for next generation of TVs that should support HDMI 2.1 and variable refresh rate. Might be patched into consoles to work over HDMI 2.0.
 
So, I recently did a clean install of Windows 10 after the last upgrade killed my PC.

I've installed the NVidia drivers, but my ROG PG279 won't overclock properly! Whenever I set Overclocking to On all my desktop Windows disappear, and the monitor shows that it's at 60Hz instead of 165Hz. If I restart Windows, the monitor won't show anything, just black.

If I then go to the ROG's settings and disable Overclocking, it works. Once it's back to 144Hz (or lower) everything's fine.

I have an EVGA GTX 1080ti SC2. Overclocking worked properly before the re-install. &#129300;
 
So, I recently did a clean install of Windows 10 after the last upgrade killed my PC.

I've installed the NVidia drivers, but my ROG PG279 won't overclock properly! Whenever I set Overclocking to On all my desktop Windows disappear, and the monitor shows that it's at 60Hz instead of 165Hz. If I restart Windows, the monitor won't show anything, just black.

If I then go to the ROG's settings and disable Overclocking, it works. Once it's back to 144Hz (or lower) everything's fine.

I have an EVGA GTX 1080ti SC2. Overclocking worked properly before the re-install. &#55358;&#56596;

Though It's not a great answer, I can't tell the difference between 144hz and 165hz at all.
 
Though It's not a great answer, I can't tell the difference between 144hz and 165hz at all.
It turned out that as soon as I set the monitor to Overclock, windows would set it as the secondary display. I have an HDTV connected through HDMI, but it was off, so I didn't notice.

Changed it back to primary.
 
Ehhh? Well, I don't know what any of that means. The only drawback to 21:9 I guess is if you just don't like 21:9. And 100hz < 144hz, I suppose.

Problem is that reaching 100+ fps for most modern games at 3440x1440 requires a 1080gtx and above. For HQ settings at least. 200fps are out of the question,unless I opt for SLI
 
Ended up picking up the d2716dg from Micro Center this weekend. It's unbelievable.

Coming from a 1080p 120Hz benq non g-sync monitor, I honestly didn't think I would really notice much of a difference... I just thought it would be like 120HZ with 120FPS I was wrong... I was so wrong. The resolution bump to 1440 is a lot more noticeable than I thought. It is much easier for me to spot enemies in games like PUBG, and multiplayer games like battlefield 1 and overwatch are so crisp and smooth now as well even when dipping FPS. I was never bothered by tearing in games before, but now having seen no tearing with g-sync... what the hell, how did I ever notice? It's kind of weird... but in a really awesome way. I'm sold on g-sync... there is no going back. Every monitor purchase I make will not have to include the g-sync chip... no doubt

Here are my settings:

NVCP:
- GSYNC: On
- VSYNC: Force On
- Max Pre-Rendered Frames: app setting (some say it should be 1, but I don't want to mess with this)

Nvidia Inspector:
- Frame Limit: 143FPS

In-Game:
- Vsync: Off

I also noticed a small stuck pixel (it's blue) and I can't seem to get rid of it. So I could try and replace the monitor for another or just deal with it... I can't really see this pixel unless I am standing above the monitor.

thanks to this text i just bought the same monitor, rip wallet
 
A G-Sync monitor is still out of my budget, but I'm still kind of in the dark about what benefits it holds over a 144 Hz monitor with Lightboost on, Fast Sync enabled in Nvidia Control Panel, and a GPU that can run almost all games in 1920x1080 well above 60 FPS and a lot of them over 100 FPS. If I'm already not seeing screen tearing or stuttering, what's the benefit of G-Sync at this point?
 
I have a 16GB i7 1080Ti PC, and I finally saved up for a 144hz/1440p gsync monitor Been using it for about a month now and the topic title is absolutely true.

(TO BE CLEAR I am platform agnostic. Use AMD+freesync if you want. This tech is just... this is nearly incredible.)

The best way I can explain it is that even with vsync/60fps and a 980ti, Titanfall 2 looked like things... didn't flow smoothly like they do now when I would turn quickly. I turn off blur, bloom, etc, and can't stand tearing of any sort. I always had vsync on, and even tried the nvidia control panel vsync extra settings.

Some games' menus don't seem to respect the settings in the nvidia control panel, so weird flickering happens, but in my experience so far every game benefits from gsync.
 
Guys, I need help...

I have the Dell S2417DG and a Zotac 970. I started dicking around in the settings for the monitor and overclocked it to 165. I booted up Overwatch, the screen flashed, and then moved everything over to my secondary monitor and went black. I can't get the monitor working over Display Port at all! I tested the HDMI connection and that works, but when I switch to DP, the monitor shows up in Nvidia Control Panel, but I can't select it and turn it on.

So what's the most likely culprit? Software setting (please please please)? Bad DP cable? Bad DP port on the 970 (bad)? Or bad DP port on the monitor (very bad).

This sucks! Anyone run into anything like this before?
 
Why force (NVCP) V-Sync? I leave it off and turn it off in every game. You don't need it.


There is FreeSync, but not one talks about it. Wish NVidia would stop being shitty and support it.

If you turn-off v-sync you'll get tearing when the frame rate is over the refresh rate of your monitor. Turning it on and also having g-sync enabled means that when the frame rate is over the refresh rate of your monitor v-sync will be engaged and you will not get tearing.

Personally I just lock my frame rate globally at 142 with RTSS.
 
Why force (NVCP) V-Sync? I leave it off and turn it off in every game. You don't need it.


There is FreeSync, but not one talks about it. Wish NVidia would stop being shitty and support it.

And more
Vsync ON with Gsync ON on NVCP also acts as a frame smoother.

The BlurBusters article explains it in detail. Leave it on for the best experience.
 
A G-Sync monitor is still out of my budget, but I'm still kind of in the dark about what benefits it holds over a 144 Hz monitor with Lightboost on, Fast Sync enabled in Nvidia Control Panel, and a GPU that can run almost all games in 1920x1080 well above 60 FPS and a lot of them over 100 FPS. If I'm already not seeing screen tearing or stuttering, what's the benefit of G-Sync at this point?
LightBoost/ULMB have strict requirements for the framerate matching the refresh rate similar to VR's requirement of never dropping below 90 FPS - but VR can fall back on reprojection while LB/ULMB cannot. If it does not, you get very bad stutter and double images.
Fast Sync causes "microstutter" and requires that the framerate is at least 2x the refresh rate for a meaningful reduction in latency. It has to be much higher than that to match G-Sync.
The best option for reducing latency is to limit the framerate to 3 FPS below your refresh rate - but without G-Sync this will cause a noticeable stutter every second.
If you care about smoothness, the best you can do is use V-Sync and set the maximum pre-rendered frames option to 1.

Why force (NVCP) V-Sync? I leave it off and turn it off in every game. You don't need it.
"V-Sync" in the NVIDIA Control Panel acts as frame-time compensation when G-Sync is active.
If you have V-Sync disabled in the NVCP and in-game, you can get stuttering and tearing when the framerate is very close to the maximum refresh rate due to slight variances in frame-time.
You don't need to have V-Sync enabled in games, but it should be enabled in the NVIDIA Control Panel.

Since you want to prevent the game exiting the G-Sync range and switching over to regular V-Sync behavior, you should also limit the framerate to 3 FPS below your display's maximum refresh rate.
Without V-Sync enabled in the NVCP, this has to be more like 15-20 FPS below the maximum refresh rate.

In-game limiters are typically the lowest latency option, but not always, and not all games have the option.
RTSS is the next-lowest latency and it works nearly universally.
Limiting framerate via NVIDIA Profile Inspector adds as much latency as V-Sync, so there's no point to it.

Guys, I need help...
I have the Dell S2417DG and a Zotac 970. I started dicking around in the settings for the monitor and overclocked it to 165. I booted up Overwatch, the screen flashed, and then moved everything over to my secondary monitor and went black. I can't get the monitor working over Display Port at all! I tested the HDMI connection and that works, but when I switch to DP, the monitor shows up in Nvidia Control Panel, but I can't select it and turn it on.
So what's the most likely culprit? Software setting (please please please)? Bad DP cable? Bad DP port on the 970 (bad)? Or bad DP port on the monitor (very bad).
This sucks! Anyone run into anything like this before?
You should be able to bring up the menus and disable the overclock even without an input to the monitor.
I'd also try another DisplayPort output on the GPU and see if that works.
I suspect that the issue here is that the cable can't handle the extra bandwidth, as my understanding was that the "overclock" here was simply that NVIDIA GPUs and G-Sync displays were built to go beyond the DisplayPort 1.2 spec, not that anything is actually being "overclocked" here.
 
When you have it connected to DP, even if it doesn't get a signal, can you access the monitor's settings and disable the overclock?

edit: beat
 
Top Bottom