It seems one of the biggest hurdles for many people wanting to use AI models is vram. I know you can use vram & system ram with ollama, but performance takes a big ding, at least in my experience. I know in games the vram is fairly closely tied to graphics processor capabilities, or at least that’s my thought.
With Nvidia’s latest card releases I thought the 5080 would have 20 to 24GB of vram but obviously not which gets me thinking they’re really pushing AI and their high end towards 5090. Also does that mean the 5080 doesn’t really have enough vram for the gpu or is right on the edge at least for games that could fully utilize it?
Would it benefit AMD to release cards with large amounts of VRAM as an alternative? I mean if AMD released cards with much higher vram, maybe at the low end you have 16GB and maybe even up to 48 or 64GB on the high end. I get it that games most likely wont use that much vram but AI would and there could be a market for people in machine learning that need the ram but don’t have the budget to drop $2K on a 32GB card.
Feel free to tell me why I’m way off base and correct my thinking.
Is vram just so expensive it would make no sense?