You basically just said that people are better off with a 3080-3090 for 1080p as it will never even use close to it's full capacity.
No, not even close. I was getting 50-80% on my gtx970, and even when moving upto my current 2070Super I'm still getting the same rough results from the same games, since I'm also using a stronger cpu which provides higher fps, so a higher workload comparatively.
1% usage or 99% usage is the exact same thing, there's no difference to the output to the screen. The difference is when the frame changes and demands more resources are used. If usage is lower, there's more room for expansion. If you are at 98% usage, an intense frame or frames can and will put that to 100% and then there's nothing left. No more resources. You are capped. So if you are at 98% and the bomb goes off and creates a giant fireball and gpu usage hits 100%, you'll get stutters, frame drops, freezes. You just ran out of cores, vram etc they can do no more, accept no more.
It's no different than trying to run gta5 on a 4690k vs a 4790k, same cpu except the i5 will see 100% usage and lower fps, but the additional threads on the i7 will lower usage and you don't get bottlenecks as a result.
Op was saying he was disgruntled because no matter what he did to the pc he was getting 90% gpu usage, whereas others were hitting 98%. He assumed that the others were doing better because of higher usage. I'm saying that's backwards, He was in a better position as he still has 10% available resources whereas the others were almost bottlenecked and most likely did have patches of 100% usage and were bottlenecking.
Usage isn't using a % of the clock speeds or memory speeds or timings, they run at 100%, usage is the amount of cores required to complete the frame, the amount of memory required to hold the frame etc. Just like you will use every muscle to swing the hammer, but don't need to use the entire strength of all those muscles to do it. You use all 5 fingers on one hand to type with, but you don't need to jam the keys down as hard as you can, just takes a light touch. Same motions, same fingers, same results, different usage in the amount of finger strength applied.
Cpu sends a frame to the gpu. Gpu takes that data and creates a wire frame first, placing every object provided at the addresses provided using the dimensions provided. When everything is placed, the gpu fills in colors, overlays affects from lighting, shadows, rtx, whatever. All that takes a certain amount of time. Usage of 1% or 99% is irrelevant, it is whatever the gpu needed to use to render that frame. Doesn't affect the amount of time it took at all.
The difference between using a 3080 and op's 2070S is in the core clocks, memory clocks, IPC etc. A 3080 takes less time to render the frame than a 2070S does, so can complete more frames per second, higher possible fps. Usage doesn't affect that.
If a gpu has to use 1000 cores and 1Gb of vram to complete a frame, that might be 60% usage for a 2070Super and only 40% usage for a 3080, but that usage doesn't matter at all. What matters is the faster clocks, faster vram, higher IPC of the 3080 dumping out that frame 2x as fast as the 2070Super, so doubling fps.