A minor note: according this thread on Ace's hardware forums Larrabee vector ISA is called LRBni, and it's different from AVX, but I guess we already knew that.
What you say is all true. I just think a design from the early 90's would be more at home in a museum than tomorrow's multi-teraflop C-GPU
Half of that 30M transistor budget should be cache..
LOL. I was hinting to the fact that caches are more dense that logic, that's it!How useful would it be without that cache?
LOL. I was hinting to the fact that caches are more dense that logic, that's it!
Agreed. 32+ cores is more likelyTrue, but I was basing my numbers off transistor count, not area. So I still think the 350 cores would be a little useless, unless we say everything but the cores takes 0 transistors...
Going by 1.4 billion for Nvidia's chip, and sandpile.org's listing of 4 million for the P54, 350 original Pentiums would fit.
They wouldn't be too useful, since they lack the other 26 million transistors Larrabee's cores have, and there's no other logic or interconnect to actually talk to them.
They wouldn't be too useful, since they lack the other 26 million transistors Larrabee's cores have, and there's no other logic or interconnect to actually talk to them.
The hints of Larrabee's vector functionality and scatter/gather memory access capability seem to indicate its vector functionality exceeds the limitations of current x86 SSE, so why expect current x86 tools to do it justice?
They do mention Ct, but they don't mention working on an implementation if that's what you meant.Hope Intel guys are working on a Larrabee implementation of the Ct language, unfortunately they don't mention it at all on their siggraph paper.
Besides high throughput application programming, we anticipate
that developers will also use Larrabee Native to implement higher
level programming models that may automate some aspects of
parallel programming or provide domain focus. Examples include
Ct style programming models [Ghuloum et al. 2007], high level
library APIs such as Intel® Math Kernel Library (Intel® MKL)
[Chuvelev et al. 2007], and physics APIs. Existing GPGPU
programming models can also be re-implemented via Larrabee
Native if so desired [Buck et al. 2004; Nickolls et al. 2008].
Larrabee cores are 30M? Where did you hear this?
Oh! Hm. Anyone know the transistor count for 256KB of cache, PCI-E bus and some memory controllers? Time to bust out the calculator..
edit: this doesn't count any cache tags, just data arrays