Please remember that any mention of competitors, hinting at competitors or offering to provide details of competitors will result in an account suspension. The full rules can be found under the 'Terms and Rules' link in the bottom right corner of your screen. Just don't mention competitors in any way, shape or form and you'll be OK.
He uses a camera to film the screen, rather than capture on the pc directly.
All I see is the ATI crowd getting very defensive and rather confused, it's not micro stutter lol.
Unigine Engine kissing NVIDIA
No game uses that much tesselation anyway so it's a pretty pointless benchmark. I doubt any game will use that much tesselation for a long long time.
It's called DX11 tessellation, you may have heard of it? Nvidia are rather good at it.
Shouldn't this thread be closed as clearly the OP's video is not a real world example of the differences between both cards and all we are doing is dredging up benchmarks to try and proove which card plays smoother, plus a lot of posturing seems to be going on.
Personally I have never noticed an issue with microstutter on my 5970 cards or GTX 480's before them. I guess now that I know what microstutter is I could probably find it if I wanted to look for it. More importantly I have found that there is very little in it between the cards when playing actual games regardless of what benchmarks show. I've used GTX 480's in SLI and Tri Sli, and the same sort of multi gpu setup with ATI cards. Some people have said from their experience the 480 cards feel smoother when playing games with higher minimum framerates. Maybe it's true, maybe it's a placebo effect. I'm too busy actually playing games to notice this. I don't see the point of all this.
"microstutter" would generally not show on screen like that... to notice it you really need to be playing the game, theres a slight "rubber band" effect to your input its very slight but noticeable especially in fps games - otherwise as a casual observer its quite hard to notice microstutter.
I love how people come out in the defense of crossfire despite the fact its patently crap at low framerates... a large number of mostly unbias people who've moved from a crossfire setup to even a single ATI GPU and especially an nVidia setup will often comment on how much more consistant it feels.
And also slightly off topic, isn't it true that Nvidia dedicates shaders to tessellation, which if it did in a real game (assuming a real game comes out with tesselation which I doubt given the state of console ports) would result in performance hits elsewhere.
"microstutter" would generally not show on screen like that... to notice it you really need to be playing the game, theres a slight "rubber band" effect to your input its very slight but noticeable especially in fps games - otherwise as a casual observer its quite hard to notice microstutter.
I love how people come out in the defense of crossfire despite the fact its patently crap at low framerates... a large number of mostly unbias people who've moved from a crossfire setup to even a single ATI GPU and especially an nVidia setup will often comment on how much more consistant it feels.
Not true, compared my 5870 to the 480 in DX11 titles with tessellation, I disabled tessellation and enabled tessellation at certain areas in metro 2033 and Dirt2 where tessellation was prominent, there was big performance hit with the 5870 but a marginal hit on the 480.
Wouldn't change my 5970 for anything else. Certainly was a misleading thread title! The OP obviously has no idea what "real world" means.