Definitely, which points to 4 setup pipes and 64 ROPs in Hawaii.Hawaii = 2x Pitcairn-on-a-die? Looks more and more likely.
Definitely, which points to 4 setup pipes and 64 ROPs in Hawaii.Hawaii = 2x Pitcairn-on-a-die? Looks more and more likely.
You're missing the point. Just because a competitor can nearly match you with 256-bit doesn't mean that 384-bit or 512-bit is useless. They would be even faster with it. It's a matter of cost/benefit.yet,if your competitor achives nearly your speed with just 256bit and beats you by 30% with the same 384bit, you might consider it's not the bandwith that causes it.
That's not what I'm saying. The cost savings of a smaller bus aren't just in the GPU. You also have fewer chips, fewer traces, simpler PCB, etc. But if the standard becomes 4-8GB, some of those savings are gone. Last gen NVidia probably got substantial mariginal savings from having 256-bit in its GTX 680 instead of 384-bit or 512-bit, but a 4-8GB 256-bit card is probably not much cheaper than a 4-8GB 512-bit card.Designing a 512bit controller to get to get more memory sounds somehow wrong,
I believe they traded some of the ROPs for advanced GPGPU/HPC features.
my point was, that having a competitor for comparison that shows you can achieve 30% more with 384bit doesn't make it sound wise to increase the width by 33% if you look at the die shots, which show the controller being a huge part and does not benefit from shrinks (if they'd want to migrade to 20nm in an tick-tock fashion as intel does), the efficiency doesn't scale linearly.You're missing the point. Just because a competitor can nearly match you with 256-bit doesn't mean that 384-bit or 512-bit is useless. They would be even faster with it. It's a matter of cost/benefit.
I'd think the cheapest solution is probably right in the middle, 6GB with 384bit.That's not what I'm saying. The cost savings of a smaller bus aren't just in the GPU. You also have fewer chips, fewer traces, simpler PCB, etc. But if the standard becomes 4-8GB, some of those savings are gone. Last gen NVidia probably got substantial mariginal savings from having 256-bit in its GTX 680 instead of 384-bit or 512-bit, but a 4-8GB 256-bit card is probably not much cheaper than a 4-8GB 512-bit card.
exactly, and that's why it's surprising if they'd increase the controller to 512bit. adding some transistors to rise efficiency is probably cheaper than increasing the controller. the controller needs to be aligned to the external power demands.For example: 10% performance boost for board cost going from $200 to $250 probably doesn't make sense, but 10% boost for board cost going from $300 to $330 probably does.
exactly, and that's why it's surprising if they'd increase the controller to 512bit. adding some transistors to rise efficiency is probably cheaper than increasing the controller. the controller needs to be aligned to the external power demands.
I'm not an expert of that area, but I was told that the IO part of a chip is the one that barely reduces in die size, power consumption -> cost, that's why we had 384bit GPUs when the last console generation arrived and that's why today's fastest GPUs have 384bit (which may change in a few days, of course).
Well Dave was pretty strongly hinting that per per 128bit section a memory controller with slower mem speed will be quite a bit smaller than a fast one. Someone calculated that Tahiti's mem controller is bigger than 2 X Pitcairns so going with a slower 512bit can save you die space compared to a fast 384bit controller. With 384bit they would have needed to be at the bleeding edge of mem speeds, possibly having to do a lot extra work on the controller whereas now they can go with slower chip and probably have easier job with the controller.
Per 128b chunk the Pitcairn PHY is about half the size and Tahiti has 3 of them.
my point was, that having a competitor for comparison that shows you can achieve 30% more with 384bit doesn't make it sound wise to increase the width by 33% if you look at the die shots, which show the controller being a huge part and does not benefit from shrinks (if they'd want to migrade to 20nm in an tick-tock fashion as intel does), the efficiency doesn't scale linearly.
my point was, that having a competitor for comparison that shows you can achieve 30% more with 384bit doesn't make it sound wise to increase the width by 33% if you look at the die shots, which show the controller being a huge part and does not benefit from shrinks (if they'd want to migrade to 20nm in an tick-tock fashion as intel does), the efficiency doesn't scale linearly.
I would imagine it is more complicated than just slapping 4x Pit's MCs on there
the power efficiency increases, the real world efficiency reduces, I wonder at what ratio that balances (or rather at what curve).Well Dave was pretty strongly hinting that per 128bit section a memory controller with slower mem speed will be quite a bit smaller than a fast one. Someone calculated that Tahiti's mem controller is bigger than 2 X Pitcairns so going with a slower 512bit can save you die space compared to a fast 384bit controller. With 384bit they would have needed to be at the bleeding edge of mem speeds, possibly having to do a lot extra work on the controller whereas now they can go with slower chips and probably have easier job with the controller.
you're ignoring that they had achived the same speed (in games) with a Die size of 294mm (vs 365mm I think) and 3.54Billion Transisitors (vs 4.3BTransistors).gkar1 said:You really have no point because you're barking up the wrong metric. They achieved 30% more performance but they had to use a die with 60%+ more transistors. That hardly seems worth it if the leaked performance figures are true.
That does not make any sense; they would earn less.I would suggest they go straight to the 500-550$ price tag instead of the suggested 650$ and I would bet then slowly decreasing it in the next weeks.
This would fix the price relatively stable, give some people the choice to buy two of them for ~1000$ as they want, and most importantly- will give a very positive light on the product, better than if it is priced 150$ higher...
Simply the speed. 5GBps vs. 6+GBps makes obviously a hell of a difference. Dave was talking about the PHYs (which take about 20mm² per 128 bit section on Tahiti), these parts don't even know about ECC (that is done higher up in the actual DRAM controller I suppose).if they strip down the memory interface to be Pitcairn alike (tho, I have no idea what compute features beside maybe ECC made the difference),
That does not make any sense; they would earn less.
Also i doubt the price will decrease much for at least the first couple of months.
wasn't that obvious to me, sorry :/. <20% clock gain forced AMD to have a +50% bigger die area? why made that sense in the first place at all?Simply the speed. 5GBps vs. 6+GBps makes obviously a hell of a difference.
I thought the DRAM controller is the actually last stage on a chip between the chip and the memory that costs Die space independently of any process shrink and it was why increasing bit width was rather expensive.Dave was talking about the PHYs (which take about 20mm² per 128 bit section on Tahiti), these parts don't even know about ECC (that is done higher up in the actual DRAM controller I suppose).
yes, it would be more efficient if you'd utilize the address-space equally, but you are also fragmenting the address/memory space. I'd think the granularity of the address space is 4k? (is that hard wired or can AMD polish that with never drivers?)Btw., I would actually expect a 512Bit interface with the same aggregate bandwidth as a faster 384bit interface to be slightly more efficient/higher performance on average, even if it is just the higher number of open pages possible with this setup.
wasn't that obvious to me, sorry :/. <20% clock gain forced AMD to have a +50% bigger die area? why made that sense in the first place at all?
Perhaps the underlying elements of Tahiti's memory controllers were reused from Cayman's design for time-to-market reasons, and thus haven't benefited from advancements to efficiency since then.wasn't that obvious to me, sorry :/. <20% clock gain forced AMD to have a +50% bigger die area? why made that sense in the first place at all?
The actual memory controller should shrink, the PHYs (that's the part usually marked as memory interface on die shots) not. The PHYs are responsible to create the actual signals on the external pins. They need to be able to drive much more current at very high frequencies than the internal on die connections. That's why they need to be relatively large (and tend to grow with higher frequencies). But the PHYs are not en-/decoding the ECC information. As said, they don't really "know" what data or commands they are driving to the pins or are receiving, they just "translate" the signals from the memory controller or the memory between the lower clock speed memory controller and the high speed external interface.I thought the DRAM controller is the actually last stage on a chip between the chip and the memory that costs Die space independently of any process shrink and it was why increasing bit width was rather expensive.
That's what one usually strives for and why the address space is interleaved between the memory channels.yes, it would be more efficient if you'd utilize the address-space equally,
As said, it's interleaved between channels anyway.but you are also fragmenting the address/memory space.
The interleaving granularity is probably much smaller, could be as small as a cache line size (64 byte) or a a very small integer multiple of it.I'd think the granularity of the address space is 4k? (is that hard wired or can AMD polish that with never drivers?)
Nice strawman argument; Anyway, all these people you speak of can just wait 2-3 months until the price comes down.Ok, if I and thousands other people tell you that we are not going to buy anything unless it is at this exact price, how would you feel? Do you think it would impact your sales numbers or you go straight with your horns no matter what the clients demand?