NVIDIA Maxwell Speculation Thread

Arun

Unknown.
Moderator
Legend
I know it is a bit early for this (ETA 2H13 on TSMC 20HP), but I figured this part of the forum hasn't been very active in the last few days, so might as well start it now rather than later.

We know the first (or one of the first) chips will include ARM-based Project Denver processor cores, but we don't know if the rest of the line-up will and so whether they will fundamentally be part of the programming model for next-gen CUDA. Some of the best info we have that *might* be about Project Denver is from Bill Dally's Echelon presentation, but that's aimed at a specific 2018 project which would imply it's talking about a 2015-2017 chip on 14nm or 10nm. And they've said it's clearly subject to change, so it's probably something mid-way between Maxwell and what they'll actually deliver for Echelon if they get the design win.

So yeah... speculation/expectations/random thoughts go here!
 
http://venturebeat.com/2011/02/01/m...-capable-processor-code-named-project-denver/
MR: So we’ve been working on it for years(presumably on Denver). We’ve been in collaboration with ARM for a shorter time because it uses their future generation chip architecture. We’ve had hundreds of people working on it for a very long time. It’s been rumored to be everything, like an Intel-compatible product. We thought the best thing is to tell people what it was.
So it's not so new after all
 
ARM has support for 40 bit addressing so the most immediate concern is alleviated.

What other major advantages are for a 64-bit ISA except full speed "double" operations? Now, out of those, how many of them would benefit a desktop CPU?

(I'm assuming that the first Denver iteration won't dream to target the traditional server market )
 
Well addressing single objects larger than 4 GB for instance. And are you suggesting that being forced to use paging methods would be OK from year 2014 onwards?
 
ARM is lowballing the issue as they are about to release a 32bit core and don't want it to look outdated ; that article basically said "we don't wish to talk about 64bit at that momen for market reasons".

remember Intel, they basically released the 64bit pentium 4 with the feature turned off and dealt with it as a little dirty secret.
maybe ARM wishes for the turnkey NAS for home and small business short term, where Atom with 1GB ram seems to dominate, but surely they need 64bit after that.

memory is amazingly cheap anyone can build a computer with 16GB well under 500 euros.
 
Agreed. And two more points:
- Fermi *already* has 64-bit addressing
- Maxwell and Project Denver are supposed to be used in supercomputers

So I would expect Maxwell ARM core to be at least 64-bit.
 
And are you suggesting that being forced to use paging methods would be OK from year 2014 onwards?


If they're fast enough, why not? ;) (TLBs FTW?). The compilers will generate the code, not you.

Of course, I do agree with you all that this is just a stop gap. It's just that I don't see it as a major issue as apparently you do; as long as nV uses this solution for the first iteration only and they will sell them (just) as co-processors in the high performance server markets.

Ninja-edit:

Actually I didn't have any big expectations about Denver when I first heard the announcement especially since ARM already announced 2-3 months ago that they don't see a large-scale ARM adoption in the server market (meaning: no, we don't have a 64 bit ISA planned). I personally view Denver as a sort of a first try from Nv. So in this context, only 32 bit ISA is no major drawback.

If I'm wrong however.. well then you're right ;). It's just that I fail too see how Nv could be that stupid to try to go against Xeons and Opterons with Maxwell
 
Last edited by a moderator:
I am quite sure Project Denver is armv8, i.e. 64bit.
IMHO the last high-performance armv7 "mainstream" core will be cortex A15, slated for 2012. ARM might downplay 64bit now, but my interpretation of that xbitlabs quote is pretty much that they will move to 64bit after cortex A15.
But yes, this would likely make Maxwell one of the first armv8 chips out in the wild.
 
ARM did say NV licensed their next-generation ISA so that's definitively ARMv8 which is nearly certainly 64-bit. It's a good point that Maxwell might beat ARM to it though.
 
A speculation thread for a product whose predecessor is at least 6 months away :oops:

That's kinda jumping the gun, isn't it arun? :)

Anyways, the real question is what will be the level of integration of the cpu and gpu isa? Will nv expose gpu as a co-processor, sorta like 8086 and 8087, thereby integrating at the instruction stream level? That will be totally awesome, but would crucially depend on what kind of co-processor mechanisms arm will expose to it's licensees. BTW, does ARM allow making such extensions today to it's arch licensees? If not, then what is the mechanism that ARM allows for attaching accelerators?

The staid option is that the integration is restricted to LLC level and a driver controls scheduling of thread blocks.
 
I think more of the Cell ; for both the Cell or current and past designs for ARM licensees, you've got a general purpose core and an on-chip bus to talk to the rest of the chip. here the core seems pretty separated.
i.e., ARM sells a core typically to be integrated in a micro-controller or SoC. I guess that beyond the core (including maybe NEON) and its buses, what happens is no longer ARM's business. ARM cores talk to various DSPs, GPU and accelerators already and various kind of controllers and hardware.

I don't know anything about ARM and coherent links, shared memory or address space etc. though.

what made me think of the Cell is nvidia's emphasis on the memory hierarchy and memory locality. This is what tells apart Cell from classic CPUs, the local storage for the SPE.
 
I think more of the Cell ; for both the Cell or current and past designs for ARM licensees, you've got a general purpose core and an on-chip bus to talk to the rest of the chip. here the core seems pretty separated.
i.e., ARM sells a core typically to be integrated in a micro-controller or SoC. I guess that beyond the core (including maybe NEON) and its buses, what happens is no longer ARM's business. ARM cores talk to various DSPs, GPU and accelerators already and various kind of controllers and hardware.

I don't know anything about ARM and coherent links, shared memory or address space etc. though.

what made me think of the Cell is nvidia's emphasis on the memory hierarchy and memory locality. This is what tells apart Cell from classic CPUs, the local storage for the SPE.

I'd like a larrabee-esque integration.

Once again, what kind/level of integration does arm allow in the name of coherent accelerators?
 
I admit my curiousity made me click on that link.
Is it just me, or does the silicon underneath that overlay look a lot like a zoomed-in mismash of a GF100 die shot?

The silicon doesn't seem to match the description for things like the L1.
 
yes the areas feel labelled at random, and now I can see the symetrical features that cry "nvidia SP".
the colored rectangles are totally non-sensical.

this is an utterly stupid picture - and I fell for it :D
 
I admit my curiousity made me click on that link.
Is it just me, or does the silicon underneath that overlay look a lot like a zoomed-in mismash of a GF100 die shot?

The silicon doesn't seem to match the description for things like the L1.

And I thought they were putting up die shots of their first silicon. :rolleyes:
 
It's not even a "patchwork", it's a crop-resize-job from Fermi:

youhavebeentrolled.jpg


Just splash some random coloured rectangles over it, add some boundaries of functional parts... :)
 
Back
Top