• Competitor rules

    Please remember that any mention of competitors, hinting at competitors or offering to provide details of competitors will result in an account suspension. The full rules can be found under the 'Terms and Rules' link in the bottom right corner of your screen. Just don't mention competitors in any way, shape or form and you'll be OK.

10GB vram enough for the 3080? Discuss..

Status
Not open for further replies.
'Adjust settings as needed' on a brand-new, £700 flagship GPU because it doesn't have the VRAM to run Ultra on new game(s) that it would otherwise have enough power to run? Yeah, that sounds like a road to buyer satisfaction on a card you bought less than a month or two ago.

Hear hear. When I said this only a month or so ago some made it sound like I was being a crazy person and I wouldn't need to anyway but no way am I spending £700 on a card that has to compromise with settings after less than a year or maybe even 2.

If I was buying a card with detail compromises in mind I would just buy a 2080 or a 5700XT second hand.
 
Flagship card Jensen special, 10gb ram is a NO if you have a brain.
Just buy the better 6800xt 16gb ram december 8.
You are full of brains with a post like that though? :p

Launch date for the 6800XT is November 18th.
 
They're not going to do that. It's 12GB or 24GB, or the performance will be terrible.

There are massive gains from increasing the memory bus width. If Nvidia reduced it to 256bit, they'd probably lose 20-30% performance

They did 8GB on the RTX 3070 did they not? They did 10GB on the 3080... Both are faster transfer wise than AMD.
 
Last edited:
It's worth noting that some developers are just inept themselves and will do something like measure their own vRAM usage in their engine with tools which tell them vRAM allocated, rather than what is used. And they'll make estimates for vRAM usage in the video settings menu which are way out. One notable example of this is Resident Evil 3, which when I maxed out in real world testing claims 12.5Gb of vRAM needed in the menu, and has actual real world allocation of about 7.5Gb. People were claiming it was "bumping up against 8Gb limit because of this, the the REAL usage in game was like about 5.5Gb on average.

The simple fact is the entire industry is measuring this badly, ever single mainstream reviewer without exception that I have read doing any kind of vRAM analysis is telling you what is allocated, not what is used. Same goes for people who do benchmarking or other analysis. There's only a small number of people that understand this technical difference, the people doing low level game modding using a tool called Special K which hooks games exes and allows 3rd party modding of low level things, it also has the ability to inspect vRAM usage vs allocated. A handful of people had been using that prior to Afterburner beta to do this kind of analysis, but it's such a niche thing no one really knows about it.

There's good evidence that the Nvidia engineers do know about this difference, first of all they're technical experts so you'd expect them to know. Second of all they made claims that modern AAA games at Ultra 4k were basically 4-6Gb of usage. And interestingly if you make this claim on these forums like I have a number of times you're barraged by responses denying this and claims that game need way more. Then when you actually just take a little while to sit down and inspect those claims, you realise that Nvidia are essentially right.

Given how many times I've seen this I will just hold off judgement until we can really tell what is going on.
 
I will just hold off judgement until we can really tell what is going on.
Yep. That is always my position.

That is why even though I was annoyed at 3080 only having 10gb and not a fan of it as I wanted to see see it have more. In the end after looking into it properly I realised everything works perfectly fine with it now and chances of more than a handful of games needing more between now and launch of next gen cards that I will want to play is small.
 
Yep. That is always my position.

That is why even though I was annoyed at 3080 only having 10gb and not a fan of it as I wanted to see see it have more. In the end after looking into it properly I realised everything works perfectly fine with it now and chances of more than a handful of games needing more between now and launch of next gen cards that I will want to play is small.

The topic of the thread is "is 10gb of vram enough on the 3080". The answer is no as there should not even be a question about whether a handful of games will need more before your next upgrade which maybe inside a year away.

I know why you went 3080 though so all good and i would want gsync myself at 4k.

When paying out around £700 or more it shouldn't even be a small thought in the back of your head that for some games you might be short on vram end of thread really. I think the last time we had these discussions was Fury x and we know Fury ran into memory issues over time, i think those that keep the 3080 long enough will run into similar problems.
 
Given how many times I've seen this I will just hold off judgement until we can really tell what is going on.
Yes, it is clear to see from your 500 posts and 15000 words in this thread that 'witholding judgement' is certainly a speciality of yours.

Just... lol...

You and Ta couple of others have been unequivocal throughout this thread about how 10GB is enough for the entirety of this generation and people are essentially stupid for daring to think otherwise... and now the moment there is some actual doubt you are suddenly "reserving your judgement"?

Don't make us laugh. :D
 
Last edited:
Yes, it is clear to see from your 500 posts and 15000 words in this thread that 'witholding judgement' is certainly a speciality of yours.

Just... lol...

You and TNA have been unequivocal throughout this thread about how 10GB is enough for the entirety of this generation and people are essentially stupid for daring to think otherwise... and now the moment there is some actual doubt you are suddenly both "reserving your judgement"?

Don't make us laugh. :D

I thought TNA was Total NonStop Action and not Tiny Nippers A***.

Edit ahh you changed it before I quoted you

I am never going to see those 3 letters now without thinking about that.
 
I thought TNA was Total NonStop Action and not Tiny Nippers ****.

Edit ahh you changed it before I quoted you

I am never going to see those 3 letters now without thinking about that.
As much as it gave me a brief light-hearted chuckle, I decided to keep it more mod friendly. :p
 
Last edited by a moderator:
Yes, it is clear to see from your 500 posts and 15000 words in this thread that 'witholding judgement' is certainly a speciality of yours.

Just... lol...

You and Ta couple of others have been unequivocal throughout this thread about how 10GB is enough for the entirety of this generation and people are essentially stupid for daring to think otherwise... and now the moment there is some actual doubt you are suddenly "reserving your judgement"?

Don't make us laugh. :D

Dunno what you're on about mate. I've had loads of people volunteer up games/examples that apparently disprove what I'm saying and I've withheld responding to those people until either I could replicate the circumstance myself through testing, or I've found evidence that others have posted from their own testing. And once I've tested I actually post evidence of the real usage in the way of screenshots and youtube videos recording game play. The sum total of all that evidence to date supports what I said.

This game may be an exception to what I've said, but I'll reserve judgement on that until we can actually test it, which is what I've done with literally every other game and is clearly documented across these threads with all the relevant evidence.
 
Dunno what you're on about mate. I've had loads of people volunteer up games/examples that apparently disprove what I'm saying and I've withheld responding to those people until either I could replicate the circumstance myself through testing, or I've found evidence that others have posted from their own testing. And once I've tested I actually post evidence of the real usage in the way of screenshots and youtube videos recording game play. The sum total of all that evidence to date supports what I said.

This game may be an exception to what I've said, but I'll reserve judgement on that until we can actually test it, which is what I've done with literally every other game and is clearly documented across these threads with all the relevant evidence.

Richdog is talking out of his arse. It's what he does.
 
The topic of the thread is "is 10gb of vram enough on the 3080". The answer is no as there should not even be a question about whether a handful of games will need more before your next upgrade which maybe inside a year away.

I know why you went 3080 though so all good and i would want gsync myself at 4k.

When paying out around £700 or more it shouldn't even be a small thought in the back of your head that for some games you might be short on vram end of thread really. I think the last time we had these discussions was Fury x and we know Fury ran into memory issues over time, i think those that keep the 3080 long enough will run into similar problems.

Yeah, those who keep it for longer than one gen may run into a problem. I will be building a family member a PC and that includes monitor and for that build I will be going with 6800XT as I know he will likely be keeping that card for a long time. Even if he was not I would still go with that as it will go nicely with 5900X.

I paid £650 for my 3080 and for my needs I know 10gb will not be an issue until next gen. I am happy to turn down texture setting one notch if needs be in a handful of future titles. No big deal for me. The main games I am getting this card for will all likely not need more than 10gb, those are Cyberpunk 2077, Dying Light 2 and Bloodlines 2.

I try my best to say it how I see it without letting bias get in the way. That is why even though I am getting a 3080 I am still very excited about AMD’s offerings and think they have done very well. That said having a OLED that is G-sync and a monitor that is g-sync and the top 3 games I am looking forward to having RT/DLSS means I choose the 3080 for myself. I will not cut of my nose to spite my face. Even though I am no fan of nvidia or the greedy leather jacket man.


I thought TNA was Total NonStop Action and not Tiny Nippers A***.

Edit ahh you changed it before I quoted you

I am never going to see those 3 letters now without thinking about that.
That guy lost all credibility. I have put him in his place more than once so he stuck me on ignore. Now he takes every opportunity to take a dig at me he can. Anyone who does not agree with him is apparently a mug.

He makes stuff up to to. Like where have I said anyone is stupid for daring to think 10gb is not enough? Sad person, makes me laugh :D
 
I have an RTX 2080 8GB and VRAM was never an issue till the last month at 1440p. Ordered an FE RTX 3080 as RTX and DLSS are must-haves for me over higher res textures which I can knock down 1 notch if needed with . The Godfall is an AMD sponsored title and it wouldn't surprise me if they intentionally souped up the textures with no discernible visual difference to have the NVIDIA cards at a disadvantage to market Navi like NVIDIA did with tesselation in Crysis 2
 
Yeah, those who keep it for longer than one gen may run into a problem. I will be building a family member a PC and that includes monitor and for that build I will be going with 6800XT as I know he will likely be keeping that card for a long time. Even if he was not I would still go with that as it will go nicely with 5900X.

I paid £650 for my 3080 and for my needs I know 10gb will not be an issue until next gen. I am happy to turn down texture setting one notch if needs be in a handful of future titles. No big deal for me. The main games I am getting this card for will all likely not need more than 10gb, those are Cyberpunk 2077, Dying Light 2 and Bloodlines 2.

I try my best to say it how I see it without letting bias get in the way. That is why even though I am getting a 3080 I am still very excited about AMD’s offerings and think they have done very well. That said having a OLED that is G-sync and a monitor that is g-sync and the top 3 games I am looking forward to having RT/DLSS means I choose the 3080 for myself. I will not cut of my nose to spite my face. Even though I am no fan of nvidia or the greedy leather jacket man.
Fair enough, most people won't care about turning down the odd setting (texture quality for example) to remove any potential hitching or other issues associated with memory saturation.

I think given your display, it does make more sense to get the 3080 so can see why you are going that route.

Video memory debates always run high with passion, regardless of which camp you are in (camp 8GB is more than enough or camp 16GB is the minimum you'd want :p).
 
Fair enough, most people won't care about turning down the odd setting (texture quality for example) to remove any potential hitching or other issues associated with memory saturation.

I think given your display, it does make more sense to get the 3080 so can see why you are going that route.

Video memory debates always run high with passion, regardless of which camp you are in (camp 8GB is more than enough or camp 16GB is the minimum you'd want :p).
You and TheRealDeal can at least understand my choice on this occasion and not resort to lying and name calling like Richdog.

You know me Matt, purchased a few AMD cards from you over the years, I am not a fanboy and have had more AMD cards than Nvidia. Just happens to be the case the 3080 suits me better on this occasion. I never keep cards for too long, so won't be long until my main rig is rocking an AMD card again. I have more than one PC in the house and the other monitor does both freesync and is g-sync compatible. That has a AMD card in it ;)


Richdog is talking out of his arse. It's what he does.
Indeed.
 
You and TheRealDeal can at least understand my choice on this occasion and not resort to lying and name calling like Richdog.

You know me Matt, purchased a few AMD cards from you over the years, I am not fanboy and have had more AMD cards than Nvidia. Just happens to be the case the 3080 suits me better on this occasion. I never keep cards for too long, so won't be long until my main rig is rocking an AMD card again. I have more than one PC in the house and the other monitor does both freesync and is g-sync compatible. That has a AMD card in it ;)
I know, i used to get PMs from you asking if i was selling any GPUs in the next few weeks on more than one occasion. :D
 
The 3080 already cant hit a 60fps average on a number of titles at 4k if you max everything out, and out of all of those we've seen so far, none of them have been a result of having insufficient vram, not even watchdogs. people buying a new card expecting everything to run flawlessly using ultra over-the-top settings need a reality check i think.

I was just think that a 3080 won't be able to run my favorite sim at max settings and it has nothing to do with vram. The game is just too demanding. The 3090 won't be able to do the job either.

If "flagship" means the ability to run every title in 4k at max settings or no buy, then they can't buy *any* graphics card on the market.

-At all.
 
Status
Not open for further replies.
Back
Top Bottom