Speculation and Rumors: Nvidia Blackwell ...

Where can I read more about Nvidia's effort to push AI users to workstation cards? I'm a bit skeptical about this, since they market their Geforce GPU's for LLM and image generation use.

Also, 32GB isn't alot for AI use. I bet you won't see the 5090 being recommended on r/localllama for the same reason that no one recommends Nvidia workstation cards either, they're just way too expensive for the amount of VRAM they offer. A system with 2x3090 will be a lot cheaper and has +50% VRAM capacity. It can for example run Llama 3 70B at 4 bits per weight whereas a single 5090 won't be able to.
It should be self evident from the pricing and the fact that they gimp the drivers on gaming cards. They’re obviously aware the x90 cards are used by creators and for AI, but they clearly prefer you to buy the cards with higher markup. Why else would they have dropped the NVLink?
 
It should be self evident from the pricing and the fact that they gimp the drivers on gaming cards. They’re obviously aware the x90 cards are used by creators and for AI, but they clearly prefer you to buy the cards with higher markup. Why else would they have dropped the NVLink?
Lol ... not sure where you get your facts but there are not any gimped drivers. If fact if you choose you can install the Studio drivers if you decide though the difference is the amount of QA involved for stability. Same games, same applications.
 
It should be self evident from the pricing and the fact that they gimp the drivers on gaming cards. They’re obviously aware the x90 cards are used by creators and for AI, but they clearly prefer you to buy the cards with higher markup. Why else would they have dropped the NVLink?
The only artificial driver level limitation on a 4090 that I'm aware of is that P2P (GPU-to-GPU memory access through PCIe) isn't supported. This and NVLink have to do with multi-GPU setups only and even there, not having these features doesn't limit or slow down stuff like multi-GPU LLM inference.

The direct GPU-to-GPU linking is more helpful with AI training however. So if there is a push to drive "AI users" towards workstation GPU's, it's aimed at professionals.

Also, doesn't the pricing make it pretty clear that Nvidia isn't pushing your average Stable Diffusion enjoyer towards a workstation card? I mean, what kind of a hobbyist AI user can afford them?

Getting back to Blackwell, I'll point out that the way AI models are advancing means that during it's lifespan, the 32GB will be much more limiting to the 5090 than the 24GB is/has been to the 4090. The first Stable Diffusion model, released two years ago, was 2GB in size. The current sota T2I model that can be run on a home computer, Flux.1, is 22GB in size.

I would be surprised if Nvidia finds it necessary to gimp the 5090 AI performance in any other way than keeping the current multi-GPU limitations in effect (if even that), the 32GB memory capacity is gimped enough already.
 
Back
Top