I mean performance-per-Watt for the entire card. I'm expecting it to be similar... It would be nice to see an improvement, and it may well happen, but I'm not expecting it to be significant. I agree that the 2Gb 5870 would be the right Cypress card to compare to for this measure.
We've been down this road before so no need to rehash old arguments, but since the proportion of transistors given over to control logic within each SP group has increased (20% reduction in shaders for a 10% reduction in area), I'm expecting the extra power required to run these to roughly balance out the increase in efficiency that we will see from a 4D shader architecture.
It would be nice to see a small improvement though, and it's entirely realistic that we will. I wouldn't necessarily consider the architecture to be "a failure" if it doesn't show global performance-per-Watt improvements though, as the improved scalability will provide a platform for future generations.
edit:
Okay sure
Pretty damn sure that my money is safe though!
The only problem is the slides say logic was simplified and made smaller
The shader size is reduced by 10% for the same performance, it doesn't mention the logic but in other slides it says both the thread dispatcher and core logic was simplified(read reduced) because its a far more even and simple setup so core logic has also been reduced, which is, iirc, what I argued before
The reason its not 20% smaller for 20% less shaders(in a given space) is they've not moved from 4 simple + one complex shader, to 4 simple shaders, but to 4 "medium" shaders that can do a little more than the previous shaders. The new shaders are bigger than the 4 simple ones, but smaller than the complex ones.
The whole core right the way through is much more simplified with basically identical shaders throughout, theres no chance in hell it will end up the same performance per W as Cypress, but I expect only a small boost over Barts(because it has a lot of improvements already. Most likely we're talking about probably a 180W Cypress vs a probably 230-240W 6970, thats likely to be really 40% faster. It would really need to be pushing 250W with the same amount of memory to not improve performance/watt over Cypress.
As for another poster asking why we care about TDP, in this case its because the 6990/dual GF110(still really questionable if that will happen) will almost certainly WANT to fit within the 300W bracket otherwise LOTS of companies won't put it into their computers for sale. Yes, we on OCUK who might buy one(ok those who post here but by somewhere cheap
) don't care if it uses 600W in a single slot, but Dell sales vs OCUK sales, AMD/Nvidia care about Dell and OCUK sales numbers don't effect them in the slightest.
AS for overclocking, we've seen it pretty much confirmed there will be a TDP overclocking tool, infact I wonder if TDP will replace clock speed entirely, or both will work together, but it seems very likely we'll see a TDP limit either in CCC or AMD overdrives tool(the software for CPU overclocking) updated to offer the option.
Which will, actually poop on my consistant advice for no one to bother using furmark as no other games will load it as much.
For instance I normally argue that if you can overclock to say 900Mhz in Furmark and no further stable, thats got no direct effect on your maximum stable say Crysis overclock because a 580GTX for instance will pull 300W in Furmark and 250W in games, so a 580gtx might be stable 75Mhz higher in Crysis(or 2Mhz, basically Furmark tells you nothing).
However, if the GPU will auto up clocks to fit in a TDP limit you tell it to, knowing where you're stable at temps/clock wise for 300W will be very useful finally.