Please remember that any mention of competitors, hinting at competitors or offering to provide details of competitors will result in an account suspension. The full rules can be found under the 'Terms and Rules' link in the bottom right corner of your screen. Just don't mention competitors in any way, shape or form and you'll be OK.
I don't know about you guys in the Polaris and Pascal threads but I am going to go and use my cards for a bit of gaming.![]()
Hilarious. That's exactly what Nvidia want. They deliberately break games so AMD cards perform poorly on them. It might fool you but not me. The very reason they do this is why I will NOT be buying a 1080 or indeed any Nvidia card in the near future. Call me an AMD fan boy if you will. But at least I'm doing it because I value their business practices, not get suckered in for shiny marketing. (not to mention my card is faster and cheaper than a 980). That's just a bonus for me. I would rather support an underdog, honest company than a greasy, underhand tactic dog like Nvidia.
So sure, call yourself a hero for sticking with AMD in the current climate if you want, meanwhile I'll be enjoying much higher performance in many game titles.
One of the major benefits for Nvidia's Pascal architecture is that they can achieve really high clock frequencies. The 16nm Pascal based GeForce GTX 1080 boosts over 1800 MHz. AMD will bake Polaris at 14nm and as some information found in the SiSoft Sandra database, they reach 1.27 GHz with Polaris 10.
The SiSoftware Sandra 2015 OpenGL benchmark entry (look it up here) was entered 3 days ago and allegedly shows an entry of a Polaris 10 sample at 1266 MHz. The chip is identified by its Hardware ID, 67DF: C7 and 67DF: C4. These are both Polaris 10 ASICs, one Pro and one XT model.
The entry denotes once again 36 compute cores (Nvidia call theirs SMs), 36 x 64 shader processors per cluster would mean 2304 active shader processors. The entry also reveals 8 GB memory being clocke at 7.6 GHz running over a 256-bit, so that would be 243 GB/s It however is unclear if AMD opted for GDDR5 or GDDR5X.
Obviously this might be just a reference sample, AMD can still increase the clock frequency at launch, with that in mind it remains to be seen what we finally going to see from the red camp.
nVidia's tactic seems to be working whatever it is, I was very surprised to see the likes of humbug and tommybhoy buy 970's considering how much they rip into their practices.
I still think the 2304 shader is the desktop pro version.
Developers will always prioritize the higher market share vendor over the smaller one.
Gameworks is so 2015!!
Dat' Gameworks effect is having less and less of an issue - you only have to look at games like The Division.
All the Gameworks defenders are very silent about that game and it is one of the main Nvidia sponsored Gameworks titles this year.
People are getting so suckered into the marketing not realising that Gameworks is just a rebranded modern PhysX.
But looking at the improvements that are being made to the way Polaris handles tesellation,I expect less and less of an impact Gameworks will have,since many of the more intensive effects use tessellation.
Thats over a 20% frequency increase over a R9 390 or R9 390X. Even if the uarch has not changed that would mean similar performance to an R9 390X for the 2304 shader part.
Much higher performance in which games? Afaik AMD does pretty well in most new games and are not that far behind in Gameworks games either.
It depends on your setup. Even Nvidia's GTX 750 (not even the Ti) pulls ahead of AMD's Fury on some systems.
beany_bot talks about the 390 being faster and cheaper than a 970 and so therefore being the best value, but that's an incredibly blinkered view. If you have to upgrade your PSU to run a 390 but not a 970, the 970 is clearly better value. If you're putting together an ITX build and a 390 literally won't fit (never mind heat/power issues), the 970 will naturally be a better choice - and with mini-ITX, there's no choice at all. Using Linux or have any interest in SteamOS? Again, you'd have to be ignorant of the issues to go with the 390. No-one sane would put an AMD GPU in a Steam Machine.
These are the issues I hope AMD finally addresses with Polaris, because for me, AMD's Windows performance is not its problem (despite seemingly everyone here focusing on nothing else). I'm sure people here will tell me that it's all just around the corner, but we've been hearing these promises for nine years now, so let's wait for something tangible rather than being suckered in by marketing, shall we?
Lol, gaming on Linux (at this point in time, the future may change my opinion)
Regarding those SiSoft Sandra numbers; the P10 card shown there also has 36 Compute Units, while the GTX 1080 only has 20.
http://www.guru3d.com/news_story/amd_polaris_10_samples_work_at_1_27_ghz.html
Also if you the SISoft Sandra scores mean anything the P10 card on it scored more than a GTX 980Ti. Mostly because it does significantly better in Double and Quad float GP Compute.
Polaris 10 = 1674.64 Mpix/s
http://ranker.sisoftware.net/show_r...efdce4dde4d5e5c3b18cbc9aff9aa797b1c2ffc7&l=en
GTX 980Ti = 1256.49 Mpix/s
http://ranker.sisoftware.net/show_r...efdce5d5e6d4e7c1b38ebe98fd98a595b3c0fdc5&l=en
So from that benchmark Polaris has some amazing:
Double compute 527.88 vs 175.32
Quad compute 29.17 vs 7.92
That is all great, but will any of it translate into gaming performance? Hopefully yes![]()