Associate
- Joined
- 5 Jul 2011
- Posts
- 82
My brain just melted from reading the posts in this thread.
I need pictures!
I need pictures!
Please remember that any mention of competitors, hinting at competitors or offering to provide details of competitors will result in an account suspension. The full rules can be found under the 'Terms and Rules' link in the bottom right corner of your screen. Just don't mention competitors in any way, shape or form and you'll be OK.
What happens when we get to 1nm?
So to get more refined we need to move away from silicon?
I can just imagine the threads when we get to hydrogen
"my GTX 1780 just flew away"![]()
I thought its only electrons that have quantum tunneling, not protons or neutrons. Or am I being daft?
Below is a promising material, maybe they can work out the kinks and find a way for it to be suitable for chips like CPU's etc.
http://www.pcworld.com/article/188656/ibm_details_worlds_fastest_graphene_transistor.html
Are you suggesting it's not possible to adjust how double precision processing is handled?Apart from that suggestion being a bit... odd... we've already seen several modifications to the design. As an example, the GTX280 had a single dedicated 64-bit processing unit in each SM (leading to one double precision unit for every 8 SP cores). The GTX480, on the other hand, performs both single- and double-precision computations using the same "CUDA cores", taking two clock cycles to perform each DP computation, and one clock cycle per SP computation.
Be that as it may, power consumption is notably lower when operating in double precision mode - presumably due to only a small subset of the GPU being active during alternate clock cycles (i.e. the "second bite" of the 64-bit cherry).
We did some tests on matrix-matrix multiplication using a Tesla C2050 (we were testing performance rather than power draw but still...). With large matrices, single-precision mode would top out at around 88% fanspeed whereas double precision would top out in the mid-low 70s (using the same fan profile and target temperature).
For what it's worth runtime was pretty much as expected (double precision mode takes very close to twice as long), but there was a difference in heat output (= power draw).
I'll make it simple for you:
* Power draw = heat output
* More heat output => better cooler required
* Better cooler => extra R+D, and/or a bigger and noisier fan
* Hot space-heater GPUs put customers off, as do noisy fans (hence the importance given to noise and power draw in reviews).
... and apart from all this, the more energy you put through the GPU the more difficulties you will have in maintaining stability.
Maximum power draw is becoming an ever more important metric (hence why nvidia and AMD are investing heavily in driver-level power containment systems). For a given power-draw cap, power efficiency determines performance... And performance is good.
What exactly are you trying to say? That 28nm will be MORE leaky than 40nm, but that this will somehow translate into improved per-unit-area power efficiency?
As a general rule of thumb, the smaller you make the manufacturing process the more difficult it is to prevent current leakage, since you need greater relative precision in manufacturing. I see no reason (yet) to convince me that will change with 28nm.
Wall of blather.
What are SP's and DP's?
Wow. I'm not even going to bother to respond to that. In fact, I don't see a single point in there - all I see is a guy standing on a soapbox, shouting "YOU'RE WRONG" at the top of his voice in the hope that will somehow make it true
If you want to discuss something related to GPU design then great, do so. Otherwise, please stop hounding people and being so aggressive. It's tiresome. I post in this forum to discuss new GPU technology, and on occasion, to help others understand the various features, when I can. I don't want to engage in trivial arguments and pass veiled insults around - if I did, then I'd call my ex.
Honestly, sometimes I don't know what your issue is...
AMD having been smart enough to pre-empt the diminishing returns they are running into with their architecture and move towards making pipelines more efficent, potentially they could end up similiar to nVidia (tho very different in implementation) shader wise with a more complex and higher clocked shader domain on 28nm giving them some pretty impressive theoretical peak performance.
Meanwhile we get to see the revised Fermi architecture on a process thats a better fit, the design itself is very strong despite what some people claim so we could see another leap like the 8800GTX here.
At least you speak with an unbiased opinion, very interesting/informative also![]()