If you’ve ever tinkered with the video settings of a game, you have likely encountered the option to turn VSync on or off.
In the majority of cases, this is not accompanied by any explanation as to what VSync is, and the name itself does little to clear up the confusion.
VSync is short for Vertical Synchronization, and it is used to combat the issue of screen tearing.
In essence, this issue occurs when your graphics card produces more frames than your monitor is capable of displaying on your screen.
For example, if your GPU can output 80 frames per second and your monitor has a 60hz refresh rate, the monitor will attempt to squeeze those 80 frames into its 60 refresh cycles. This causes some parts of the screen to be displayed independently, resulting in screen tearing.
Table of ContentsShow
What Is VSync?
Although PC gaming has now existed for a few decades, it wasn’t until the rapid technological advancement of graphics cards in the 2000s that we encountered the issue of screen tearing. This is because GPU technology quickly moved forward while monitor technology progressed far more slowly.
Screen tearing was accepted by some gamers as simply playing a waiting game until monitors with faster refresh rates were developed. Fortunately, some innovative people had other ideas.
VSync is the initial solution developed to address the screen tearing issue. It involves restricting the GPU output software to match the monitor’s maximum refresh rate. Theoretically, this sounds like an ideal solution, but there are several concerns and these are the reason why VSync is largely considered outdated in 2025.
One of these issues occurs when the GPU is unable to perform up to the monitor’s refresh rate. As a result, the monitor will leave the previous image on display until the next one is ready, which causes perceptible stuttering.
Fortunately, there are now technologies far better suited to handle this issue, but more on those later.
Likely the biggest issue with VSync is its input lag. This is particularly frustrating in games where quick reactions are necessary, such as a shooter game. It’s an even greater issue if the game in question is a multiplayer game, where your opponent could end up beating you simply because they use a different syncing solution.
Should You Use VSync?
From what has been said so far, the simplest answer would be yes, you should use VSync. The advantages are clear, while the drawbacks are less likely to hinder you.
However, the more nuanced answer is that you should only use VSync if it’s absolutely essential. The reason for this is simple: there are superior alternatives.
As mentioned previously, VSync technology is quite uncomplicated and it didn’t take long for other GPU giants to provide their own solutions.
VSync Alternatives
First and foremost, we must acknowledge the deserving credit; VSync was the initial solution. It was an appropriate solution back then and grew to be a fundamental aspect of graphics settings in the following decade.
While we can praise VSync for its achievement, we have to admit that its time has largely passed and that there are now superior alternatives available. Let’s take a look at which one is currently the best.
AdaptiveSync
First up, we have AdaptiveSync, which is separate from NVIDIA’s Adaptive VSync (notice the extra ‘V’). AdaptiveSync is the sole technology mentioned here that wasn’t developed by either AMD or NVIDIA. It was developed by VESA, the organization responsible for the DisplayPort standard, which is commonly used today.
To avoid confusion, both AMD and NVIDIA are members of VESA, but they weren’t part of the development process. AdaptiveSync is also a supplementary standard, which means any member of VESA can use it. Both AMD and NVIDIA used it to develop their own brand-specific screen-tearing solutions.
Probably the best thing about this standard is the way it smooths out stuttering when the FPS drops below the monitor’s refresh rate. It’s so good that it can feel like there are more frames than there actually are.
AdaptiveSync accomplishes this by enabling an image to be displayed once it’s completely rendered while maintaining the preceding image up in the meantime. How does that differ from VSync?
It’s quite simple, really. AdaptiveSync modifies the refresh rate of the monitor and forces it to wait until the frame is ready before loading it up. You can take a look at the image below to better visualize this.
FastSync
FastSync is NVIDIA’s version of AdaptiveSync and an upgrade of its own Adaptive VSync, which was considered a bit of a shambles. As the industry pioneer that it is, NVIDIA quickly responded to its own failure with FastSync.
FastSync attempts to achieve the same thing as the AdaptiveSync standard but runs into some issues where stuttering and jerking are more apparent. It does its job, but we can only recommend using it if you’re gaming online.
It’s worth pointing out that, much like the tech it’s trying to emulate, it’s still a superior option than VSync.
Enhanced Sync
While the previous entry was NVIDIA’s endeavor at fixing the vertical synchronization issues, this is AMD’s crack at it.
There isn’t much more to say about the way AMD addressed this. Enhanced Sync can chiefly be considered the same as FastSync, but with AMD’s distinctive branding all over it.
The same issue that affects FastSync, sporadic stuttering, is also noticeable here. Although both technologies were developed to eliminate stuttering, neither succeeded entirely. However, they were both great attempts, considering their respective release dates. All in all, Enhanced Sync does manage to decrease the stuttering level of VSync.
G-Sync
We’re doing well now. G-Sync is NVIDIA’s answer to screen tearing and it works great.
For NVIDIA, being the initial company to release a groundbreaking product has its advantages. NVIDIA was determined to take full advantage of this and sought to leave no money on the table.
That’s right, as much as G-Sync is the perfect solution for screen tearing, it also comes at a price.
For NVIDIA to be able to solve the problem of adjusting the monitor’s refresh rate, they had to invent a module that is implemented directly into the monitors themselves. This module, in conjunction with the software solution, is able to dynamically modify the monitor’s refresh rate to match the frames produced by the GPU.
This excellent and innovative solution requires the installation of the module by monitor manufacturers. As such, NVIDIA chose to charge money for it, hence the additional cost for consumers.
The key reason why G-Sync is a superior technology than VSync is that it eliminates the input delay. The principal drawback since the inception of VSync is finally a thing of the past.
G-Sync Ultimate
This is the latest addition to the G-Sync brand and can finally somewhat justify the additional cost. NVIDIA made sure that this was noticeable and required a thorough and meticulous inspection to be passed by monitor manufacturers in order to implement their G-Sync module.
The reason for this is straightforward: G-Sync Ultimate brings a lot to the table. NVIDIA managed to add 152 backlight zones, which means the IPS panel is capable of producing HDR images with enhanced precision.
FreeSync
As you might have guessed, this is AMD’s screen-tearing solution. However, AMD decided to allow their module to be completely free, hence the name FreeSync.
Its reasoning is probably twofold. NVIDIA was the first company to reach the market and AMD lagged behind. NVIDIA also chose to charge money for their product, so AMD making its solution free could give it a strategic edge.
To avoid any uncertainty, we have to admit that FreeSync is equally as good as G-Sync and the fact that it’s free makes it a more attractive option. However, things aren’t always that simple and that’s also the case here.
One of the reasons why NVIDIA charges monitor manufacturers for the use of its G-Sync module is to ensure the uniqueness of its cards. G-Sync monitors can only work with NVIDIA cards.
One might think that AMD would do the same but the reality is that FreeSync monitors can be used with both AMD and NVIDIA GPUs. There is also a caveat here, as not all FreeSync monitors are compatible with NVIDIA GPUs.
In 2019, NVIDIA shocked the world when they announced that their G-Sync cards would have support for FreeSync monitors. Unsurprisingly, there are rigorous standards to be met, but the fact remains that the best option is currently a FreeSync monitor.
Much like G-Sync technology, FreeSync reduces the exasperating input lag. Something else that AMD can boast about is FreeSync’s reduced flicker, which really comes in handy in those extended gaming sessions.
FreeSync Premium
AMD attempted to improve upon their FreeSync model with FreeSync Premium, and we can say that they did a really good job.
Premium brings a host of enhancements, with its most significant being the Low Framerate Compensation. This innovation responds to the frame rate dropping below the monitor’s range by doubling the FPS number and using it as the monitor’s refresh rate. This eliminates screen tearing and, although the frame rate will be lower, it will at least be consistent.
FreeSync Premium Pro
As its name suggests, Premium Pro is another upgrade of FreeSync technology. In this case, it aims to preserve equilibrium as AMD’s ray-tracing-enabled cards are released.
Much like G-Sync Ultimate, Premium Pro seeks to maintain the appropriate refresh rate of the monitor to display HDR images that are produced by the latest graphics cards.
Why You Shouldn’t Use VSync
As suggested, VSync simply isn’t an optimal solution. The worst part is that it causes input lag, which no gamer wants to deal with.
If available, we would recommend either G-Sync or FreeSync, depending on which graphics card you have. As mentioned earlier, FreeSync appears to be a preferable option at this point with even NVIDIA adapting to it.
It’s important to point out that, if your monitor is neither G-Sync nor FreeSync compatible, you’ll have to go for VSync, which should eliminate screen tearing anyway.