The Arcturus / CDNA 1.0 based Instinct MI100 has 120 Compute Units, running at a TDP of 200w, according to a leaked BIOS, in Feb 2020, link here:
https://www.techpowerup.com/263743/...i100-arcturus-hits-the-radar-we-have-its-bios
AMD seem to have achieved a very low TDP of 200w by clocking the GPU at just 1334 Mhz, leading to a GPU capable of 20.49 TFlops. Previous Vega based Instinct GPUs like the MI60 ran at 300w, and was capable of 14.75 TFlops.
Apparently, the MI100 has no 3D engine, which could explain how they were able to increase the CU count to 120 (due to reduced die size).
All this makes me wonder, for the top RDNA 2 GPU, couldnt AMD reduce the CU count to ~96 (cheaper to produce), then increase the GPU clock upto Navi10 levels? That should keep tdp at 300w or less. AMD wouldn't need to use the HBM2 VRAM or 4096 bit memory bus, so power savings could be made there.
I guess this comes from the idea that the CDNA and RDNA 2 architectures might be very similar, in some ways. The CDNA architecture is very power efficient it seems, it would be a waste not to carry these improvements across to next gen GPUs.
https://www.techpowerup.com/263743/...i100-arcturus-hits-the-radar-we-have-its-bios
AMD seem to have achieved a very low TDP of 200w by clocking the GPU at just 1334 Mhz, leading to a GPU capable of 20.49 TFlops. Previous Vega based Instinct GPUs like the MI60 ran at 300w, and was capable of 14.75 TFlops.
Apparently, the MI100 has no 3D engine, which could explain how they were able to increase the CU count to 120 (due to reduced die size).
All this makes me wonder, for the top RDNA 2 GPU, couldnt AMD reduce the CU count to ~96 (cheaper to produce), then increase the GPU clock upto Navi10 levels? That should keep tdp at 300w or less. AMD wouldn't need to use the HBM2 VRAM or 4096 bit memory bus, so power savings could be made there.
I guess this comes from the idea that the CDNA and RDNA 2 architectures might be very similar, in some ways. The CDNA architecture is very power efficient it seems, it would be a waste not to carry these improvements across to next gen GPUs.
Last edited: