I just can't stand the additional latency of mGPU setups now, whereas before it never used to bother me. Weird.
Single powerful card + nice Gsync/Freesync monitor all the way imo, provides a much richer all-round experience![]()
Agree
Please remember that any mention of competitors, hinting at competitors or offering to provide details of competitors will result in an account suspension. The full rules can be found under the 'Terms and Rules' link in the bottom right corner of your screen. Just don't mention competitors in any way, shape or form and you'll be OK.
I just can't stand the additional latency of mGPU setups now, whereas before it never used to bother me. Weird.
Single powerful card + nice Gsync/Freesync monitor all the way imo, provides a much richer all-round experience![]()
I hope that Vega brings some new universal technology that is usable across all games that significantly or at least decently improves gaming performance.
This new HBCC maybe useful, but only if it's usable by devs with little effort.
It's hardware controlled so it needs only minimum effort in low level APIs (few lines of code to enable), and works by default in lower dx versions as far as i know.
8gb should be enough for a good couple of years anyway imo. Until PS5 and an even newer Xbox comes out. By the time 8gb is not enough we will be on a new gpu.
I dunno, I'd bet money on Nvidia getting involved with a game soon after Vega releases and trying to get the memory usage pushed beyond 8gb's so the Ti and TXp's can be seen to stand above every other card just like they did with the Fury and ROTTR.
I think the point of HBCC is being missed, no game requires 8gb to render one frame. Its not a simultaneous requirement, AMD is proposing or stating they will juggle textures before they are ever required. Instead of brute force they declare they have improved the intelligence of design and operation of a graphics card not to require the loading of every part of the game textures into immediate GPU memory.
Maybe it'll screw up, maybe it'll be bad 1% of the time and people will avoid it like the plague as its a nightmare for min frame rates. Or it could be perfect, do you believe in this guy :
![]()
If they're bigging up the HBCC as a way to reduce mem usage, I'd bet that only the top Vega card will have >4GB. I reckon the 1070/1080 class cards will have 4GB only, to keep costs down. And that will certainly put many off.8gb should be enough for a good couple of years anyway imo. Until PS5 and an even newer Xbox comes out. By the time 8gb is not enough we will be on a new gpu.
I made my own face mask with it and now I stare at the mirror and stroke myselftbh any true Vega faithful should be printing off said picture and sticking it on the side of the monitor for daily adoration
Hopefully its a really unpopular idea and people dont get it, any squeeze on stock initially might mean the 4gb version is the one to get. After a while it'll sink in, presuming Raja Koduri was correct in his estimationsGoing to take a lot of convincing, 8GB being fine, because of the new HBCC.
Am really excited to see what HBCC brings.
At its Capsaicin & Cream event today, AMD announced that its High Bandwidth Cache Controller (HBCC), a feature introduced by its "Vega" GPU architecture to improve memory management, will increase game performance tangibly. The company did a side-by-side comparison between two sessions of "Deus Ex: Mankind Divided," in which a HBCC-aware machine purportedly presented 2x better minimum FPS, and 1.5x better average FPS scores, than a non-HBCC-aware system (though the old, trusty frame-rate counter was conspicuously absent from both demos).
AMD also went on to show how HBCC seemingly halves memory requirements, by deliberately capping the amount of addressable memory on the HBCC-aware system to only 2 GB - half of the 4 GB addressable by the non-HBCC-aware system, while claiming that even so, the HBCC-enabled system still showed "the same or better performance" through its better memory management and bandwidth speeds. If these results do hold up to scrutiny, this should benefit implementations of "Vega" with lower amounts of video memory, while simultaneously reducing production costs and overall end-user pricing, since smaller memory pools would be needed for the same effect.
https://www.techpowerup.com/231093/...e-controller-improves-minimum-and-average-fps
To me these technologies sound like they are aiming at the 4k market.
Technically I would say they demonstrated "slowdown avoidance" (in the absence of enough RAM) as opposed to performance improvement. Like you said in the second paragraph the only reason this happened is that they ran the game with 2GB of memory so that game data does not fit in VRAM. This forced the card to move data back and forth from system RAM, which the HBCC powers.
In reality this is a non-issue as cards come with enough VRAM for games to work properly. Nowadays 4GB is more than enough for 1080p gaming and 8GB is more than enough for 1440p. Cards that are capable of 4K like the 1080ti have 11GB.
As long as you have enough to store all game data, you don't need the HBCC, the game will run at full speed and the HBCC cannot "improve" the speed. Only in cases where you don't have enough VRAM would the HBCC become handy and that would just prevent the slowdown.
On the other hand, it makes for future-proofing your card. If you buy a 4GB Vega that is fast enough for 4K and a game with 7GB of data sets comes along, you should be able to play it at full speed (without slowdown) because of the HBCC.
The HBCC is really a feature for professional use where cards doing data crunching need to process huge data sets.