I'm trying to found out a bit more info on the memory architecture of the latest Tesla cards.
The previous series 8 had say; 128 processors, grouped in blocks of 16, with each block having access to on chip high speed memory and all blocks having access to shared global memory. Hence to maximise computation speed up, data would need to be constrained to on chip shared memory as much as possible as read/writing to global mem had large overheads compared to on chip mem.
All I can find on the Tesla C1060 is that it has 240 processors with access to 4GB shared memory - is it as simple as that? 240 processors each with rapid access to the full 4GB??
Cheers
The previous series 8 had say; 128 processors, grouped in blocks of 16, with each block having access to on chip high speed memory and all blocks having access to shared global memory. Hence to maximise computation speed up, data would need to be constrained to on chip shared memory as much as possible as read/writing to global mem had large overheads compared to on chip mem.
All I can find on the Tesla C1060 is that it has 240 processors with access to 4GB shared memory - is it as simple as that? 240 processors each with rapid access to the full 4GB??
Cheers
Last edited: