Am I alone and being firmly against upscaling, never tried it, and everything is played or native resolution only? There have been games that only used 20% of GPU so I doubled game reaultion, but never upscaling.
PC Gaming
For PC gaming news and discussion. PCGamingWiki
Rules:
- Be Respectful.
- No Spam or Porn.
- No Advertising.
- No Memes.
- No Tech Support.
- No questions about buying/building computers.
- No game suggestions, friend requests, surveys, or begging.
- No Let's Plays, streams, highlight reels/montages, random videos or shorts.
- No off-topic posts/comments, within reason.
- Use the original source, no clickbait titles, no duplicates. (Submissions should be from the original source if possible, unless from paywalled or non-english sources. If the title is clickbait or lacks context you may lightly edit the title.)
For some games I'm happy to turn it on as it drops power/temps and provides virtually identical output (as far as i can tell anyway) to native, but my fans don't need to go into overdrive mode.
I may even put on frame gen too if I just want to bump a stable-ish 80-90fps to stable 120fps, and again drops power and temps slightly. That sometimes does cause smearing but for the most part I don't notice enough to be annoyed. Without them I would probably be running with more power draw and higher temps, and possibly still not even hitting lower resolutions at 120, some games as you say though can hit 120 no problem even without and the gpu won't be stressed.
want to bump a stable-ish 80-90fps to stable 120fps, and again drops power and temps slightly. That sometimes does cause smearing but for the most part I don't notice enough to be annoyed.
So why do you want high fps? The issue with low fps is, that it causes smearing.
I don't notice smearing much and run my games/display at 30 fps.
And then there are games with motion blur effect, silly, right?
My first priority is picture quality. Play on max/ultra everything but disable upscaling
I'm sure you're not alone but can I ask why? Some games use it better than others, and the tech has come a long way. If I have to choose between native 4k and 40fps, or 80 or even more fps with tiny artifacts that I only really notice when I'm actively pixel peeping... I mean, I'll take the latter, personally.
But with those options, your latency is still equivalent to 40fps so it won’t feel any snappier and now you’re introducing artifacts so the image quality is worse. I get upscaling. I don’t get why anyone would ever enable frame gen.
Neither I or the person I was replying to were talking about frame gen.
As for the latency, not in a way that I notice. I'd rather play 80 or more fps upscaled with minute artifacts that I don't notice, than 40 fps native 4k, just as I said. It feels much better to me, even if it's placebo.
I view upscaling as giving a developers a cop out for not having to optimize games, as well as an admission of noting capable of enginner gaming GPU's that can do 4K ultra path tracing 165fps.
I'm also getting suspiciois that AAA+ games are dismissing original wring and story development to replace with flashy graphics and then sell it on how good the game looks. Indie studies don't have the budget for upscalling, no 6 year old games from anybody has upscaling, I find it to be more of a gimmick than a solution. Nobody plays a game 4 years after release because of how good the game looks for grahics.