Nvidia Shows Signs in [2022]

Status
Not open for further replies.
NVIDIA "unlaunches" RTX 4080 12 GB due the name
If it was due to the name, they would rename it, not cancel it?
So would could be the real reason?

Edit: I found it. The guy running DSOG called them out on phone and said he thinks the name is confusing, so they cancelled it because they are so nice. At least that's what he thinks.

Well, fine. So we have one option less.
And there is enough room for appropriate 3070 pricing :O

Seriously, maybe 3070 gets 16GB, but earlier was planned to have less. So 3080/12 is redundant.
Kudos to Intel! ; )
 
Last edited:
If it was due to the name, they would rename it, not cancel it?
So would could be the real reason?
It takes time to rename something when you've made tons of cards with old names. If it was just stickers it would be quick, but they need to reflash every card too.
At least I got the impression it will be relaunched under different model number later (likely 4070 or 4070 Ti)
 
Nvidia's phrasing is "tens of thousands NVIDIA GPUs including A100 and upcoming H100 accelerators". So unless Tom's hardware has some further information on this, it's not necessarily tens of thousands of A100 and H100 as some of the hardware could be something lower end.


If it truly is tens of thousands of those highest end accelerators it would be a huge amount. For context:

FfVf3o3VQAA_gsk.jpg
 
Nvidia's phrasing is "tens of thousands NVIDIA GPUs including A100 and upcoming H100 accelerators". So unless Tom's hardware has some further information on this, it's not necessarily tens of thousands of A100 and H100 as some of the hardware could be something lower end.


If it truly is tens of thousands of those highest end accelerators it would be a huge amount. For context:

Tom's article is just plain wrong. Both NVIDIA and Oracle say "tens of thousands GPUs including A100 and H100", not "tens of thousands of A100 and H100 GPUs"
 
The partnership, which builds on earlier deployments, sets up Nvidia with the kind of infrastructure it requires to expand on a long-term goal to become a software powerhouse. It also gives Oracle’s cloud service the plug-and-play hardware capacity and software framework to easily deploy AI software.
...
The companies are “looking at the full stack so not just the GPUs and infrastructure but getting into the software layer, getting into the service layer,” Leung said.

Nvidia is known as a graphics chip company, but is betting its future on generating more revenue from software and services. The company is looking at a Netflix style subscription business model and charging customers when its software and hardware are used to create products.
...
The AI Enterprise offerings from Nvidia have so far been limited to a handful of virtual machine interfaces on Google Cloud, Microsoft Azure and Amazon Web Services, which have their own AI software offerings that are largely based on open-source tools. But Nvidia has found a full-stack partner in Oracle, which is willing to take on the graphics chip maker’s proprietary software stack for its cloud service.
...
Oracle customers can currently get clusters of 512 GPUs, and is adding tens of thousands of GPU capacity, Leung said. The GPUs and AI Enterprise software stack will sit on top of the core Oracle Cloud infrastructure, which includes bare metal compute, storage and networking hardware.
...
Nvidia’s aiming to provide software services as a subscription model, and the chipmaker declined to comment on whether it’ll get a cut from GPU instances on the Oracle Cloud.
 
Last edited:
Tom's article is just plain wrong. Both NVIDIA and Oracle say "tens of thousands GPUs including A100 and H100", not "tens of thousands of A100 and H100 GPUs"
They are not that wrong, the only AI data center GPU missing from that statement is the A30, so the correct statement should be A30, A100 and H100. What a massive difference!

The other data center GPUs, the A40, A10, A16 and A2 are based on the GA102 and GA107 dies, and they are for visual computing and video processing not AI.
 
Last edited:
They are not that wrong, the only AI data center GPU missing from that statement is the A30, so the correct statement should be A30, A100 and H100. What a massive difference!

The other data center GPUs, the A40, A10, A16 and A2 are based on the GA102 and GA107 dies, and they are for visual computing and video processing not AI.
It's not just for AI, it's "Accelerated computing and AI". Cloud services offer usually wide variety of different hardware configurations for their clients to serve different needs efficiently (even in this day of GPUs supporting several clients at once), why would Oracle be any different?
 
It's not just for AI, it's "Accelerated computing and AI". Cloud services offer usually wide variety of different hardware configurations for their clients to serve different needs efficiently (even in this day of GPUs supporting several clients at once), why would Oracle be any different?
Their main press release focused on AI, they integrated the whole NVIDIA AI software stack, AI hardware platforms, the whole sha·bang, I don't see any mention of visual computing and video processing in this release, and even if it was, these products are low volume anyway. The bulk of GPUs is most definitely going to be the A30, A100 and H100, you know the products directly aimed at serving AI. They even stated they are using A100 for data processing.
With the full NVIDIA AI platforms available on OCI instances, the extended partnership is designed to accelerate AI-powered innovation for a broad range of industries to better serve customers
Data processing is one of the top cloud computing workloads. To support this demand, OCI Data Science plans to offer support for OCI bare metal shapes, including BM.GPU.GM4.8 with NVIDIA A100 Tensor Core GPUs across managed notebook sessions, jobs, and model deployment.
 
Oracle is adding tens of thousands of Nvidia “Ampere” A100 and “Hopper” H100 GPU accelerators to its infrastructure and is also licensing the complete Nvidia AI Enterprise stack so its database and application customers – and there are a lot of them as you see above – can seamlessly access AI training and inference if they move their applications to OCI. At the moment OCI GPU clusters still top out at 512 GPUs, according to Leo Leung, who is vice president of products and strategy for OCI and who has very long experience in cloud stuff with Oracle, Oxygen Cloud, Scality, and EMC.

To us, it looks like Oracle is adding capacity, meaning more GPU clusters, not scaling out its GPU clusters to have thousands or tens of thousands of GPUs in a single instance for running absurdly large workloads with hundreds of billions of parameters. Leung says that the typical OCI customers are still only wrangling tens of billions of parameters, so the scale OCI is offering is probably sufficient.

Leung was mum about when – or if – the NeMo LLM large language model service that Nvidia just announced at the fall GTC 2022 conference last month might be integrated into OCI, but we reckon that Oracle would rather not have a service running on AWS or Google Cloud or Microsoft Azure (where presumably these cloud LLMs run) linked to services running on OCI. And that means Oracle will eventually have to have enough GPU cluster scale to run the NeMo Megatron 530B model internally. That right there is 10,000 or more GPUs. So Oracle saying it is adding “tens of thousands” more GPUs is, well, a good start.
 

Gaming revenue half of same quarter last year, inventory up 100% from last year to $4.5B. But they are saved from posting an outright terrible result by growing data center revenue 31% YoY, to $3.83B, or now 2.4 times as much as their entire gaming revenue. Total revenue is down 17% YoY, income down 72% YoY.

Their balance sheet is worrying, though, and appears to prove MLID right on at least something. They have both $4.5B of inventory, and a total of about ~$4B of pre-paid assets (... where they hid the non-current part under "other non-current assets". Not illegal, but given they have not done that before, also not a good look. Rather drawing the eye where they don't want you to look.) Given how inventory is on balance sheet on basis of cost, not value, they are probably sitting on more than a year's sales worth of inventory + pre-paid fab allocations. And that's at normal sales numbers. If you assume that past quarter was representative of gaming sales for the near future, they will probably still be selling GA102-based cards next summer.
 
Status
Not open for further replies.
Back
Top