Next-Gen iPhone & iPhone Nano Speculation

It could be that they're getting higher average density due to being proportionately more SRAM instead, ie more than 2MB of L2 cache or some other big SRAM area, but this would seem rather unfitting too. Maybe if for some special purpose.
This would be my bet - the large increase in transistors is due to a larger proportion of intrinsically higher density features.
My bet right now is on TSMC's 28nm simply being that much denser than Samsung's 28nm.
I sincerely doubt this, both for fundamental reasons, and because it isn't supported by any process data I've seen.

Memory, (and to some extent very regular logic) is your best bet.
 
L2 increase from 1MB to 2MB gets you an extra 48M transistors. Where do you think the other ~450M come from?
They might have added more memory to the ISP to support all the new camera features like 120 fps video, burst mode, and constantly caching pictures for comparison to implement the improved image stabilization, auto exposure, auto focus, and auto tone mapping algorithms.

They also said the A7 has secure storage for fingerprint data. No idea if that is a large data set or not.
 
They might have added more memory to the ISP to support all the new camera features like 120 fps video, burst mode, and constantly caching pictures for comparison to implement the improved image stabilization, auto exposure, auto focus, and auto tone mapping algorithms.

They also said the A7 has secure storage for fingerprint data. No idea if that is a large data set or not.

Both good points. I assume the GPU likely has appreciable cache size increases too.
 
L2 increase from 1MB to 2MB gets you an extra 48M transistors. Where do you think the other ~450M come from?
I don't know. That's why I'm so curious about the ChipWorks teardown.

So gate first vs gate last has no appreciable affect on transistor density?
A lot of things affect transistor density, many of them balancing. If the processes are targeted at the same niche, results from different foundries, at similar process nodes, have been surprisingly close considering just how many variables there are.

My prediction is that in all likelyhood the billion transistors in the A7 aren't partitioned just as in the A6, but that a greater portion of transistors are in highly regular (or otherwise intrinsically more dense) structures. Does that seem unlikely?
 
Some people are saying that A7 is a lot denser because it's proportionately more GPU and GPU is denser, but I doubt this because higher end Series 6 is supposed to be more area efficient than Series5XT - which is what you'd expect because they don't eat the area inefficiency cost of stacking together separate cores. Since they "only" get 2x the GPU performance and since I suspect part of that would be due to a clock boost I don't expect there'd be substantially more than 2x the GPU transistors used, even accounting for the extra space needed for new features.

My feeling is that they have dedicated more die area to CPU than GPU. Some are suggesting that they have gone quad core, that would be a doubling right there (discounting the die shrink). Even if not, then they have used 64-bit CPU, which probably comes with an increase size.

The stated graphics performance, whilst clearly acceptable to Apple, implies a quite conservative Rogue implementation, which again might point towards die size limitations.

Any feeling that the high transistor count might be related to a big increase in my chip memory, given that memory is highly dense.
..edit, I see that's already been mentioned.
My bet right now is on TSMC's 28nm simply being that much denser than Samsung's 28nm.

Wasn't the ongoing speculation for some time now that even when Apple start shifting to another fab, that the logistics would be such that they'd have to continue to dual source ?Perhaps my oversimplified conclusion from that was that they'd be getting the same soc from multiple manufacturers. For what people are saying here, it wouldn't be possible for Apple to state an soc size if it was coming from multiple sources.
 
Last edited by a moderator:
Fair enough. One last question. What is the block on the top left of the single core A5 that's duplicated that there's only one of on the dual core A5 in the bottom right?

Is Hydra possibly the "MP" part of a multi-core implementation. I'm certain I've read that SGX544 is different from SGX544MP1 (the implication being that the MP circuitry brings more than just inter core glue).

With regard to blocks that would have to be in the cores, for Apple TV chip it would still need the IMG video decoder block (VXD), but likely the video encoder block, that is present in the iphone and ipad chips, might have been dropped, given there are no cameras.
 
Last edited by a moderator:
They might have added more memory to the ISP to support all the new camera features like 120 fps video, burst mode, and constantly caching pictures for comparison to implement the improved image stabilization, auto exposure, auto focus, and auto tone mapping algorithms.

They also said the A7 has secure storage for fingerprint data. No idea if that is a large data set or not.

Just wondering here. Is it reasonable to think 64-bit has any significant advantages for these new tricks?

Thanks in advance!
 
What exactly was it that Apple said about transistor count?

1 billion transistors in a 102 mm² die on 28 nm isn't that unusual. Just for comparison: X-Box One SoC has ~5 billion transistors in 363~mm², Tahiti has 4.3 billion in 365 mm² and GK104 has 3.5 billion in 294 mm², all on TSMC's 28 nm process (not all the same process I believe). Given those densities and assuming that Samsung's 28 nm process is capable of roughly the same densities then Apple's A7 doesn't seem so unusual.

Maybe the confusion stems from A6 not being very dense? Has it ever been confirmed how many transistors there are in the Apple A6?
 
Just for comparison: X-Box One SoC has ~5 billion transistors in 363~mm²
Xbone APU (not SoC, since it is not a system on a chip) has an awful lot of SRAM on it, which is why it has such a monstrous transistor count. It's not really applicable in a comparison to more conventional processors.
 
Xbone APU (not SoC, since it is not a system on a chip) has an awful lot of SRAM on it, which is why it has such a monstrous transistor count. It's not really applicable in a comparison to more conventional processors.

True, that's why I didn't give it as the only example. It is pretty close to being a SoC though, as it's much more than just a CPU, GPU and SRAM on a single chip. It was all just meant as an example of showing that 1 billion transistors in 102 mm² isn't that unusual. The unusual part is Apple claiming that the A7 SoC has double the transistors of A6, or maybe we misunderstood Apple and they were comparing to something else.
 
You think Samsung's 28nm allowed them to do a chip with nearly twice the density as the 32nm A6? For comparison, scaling the A5 from Samsung's 45nm to their 32nm node yielded a chip about 58% the original size, so this would be an even bigger density improvement from what's only a half node change in designation. And note that Samsung's 28nm is basically a tweak to their 32nm, with the same basic tool flow and all that.

Some people are saying that A7 is a lot denser because it's proportionately more GPU and GPU is denser, but I doubt this because higher end Series 6 is supposed to be more area efficient than Series5XT - which is what you'd expect because they don't eat the area inefficiency cost of stacking together separate cores. Since they "only" get 2x the GPU performance and since I suspect part of that would be due to a clock boost I don't expect there'd be substantially more than 2x the GPU transistors used, even accounting for the extra space needed for new features.

It could be that they're getting higher average density due to being proportionately more SRAM instead, ie more than 2MB of L2 cache or some other big SRAM area, but this would seem rather unfitting too. Maybe if for some special purpose.

My bet right now is on TSMC's 28nm simply being that much denser than Samsung's 28nm.

Doesn't Samsung use a gate first approach on their 28nm node, which they trumpet as allowing higher densities, because of less restrictive design rules, apparently at the expense of yield %, compared to gate last. Given that everybody is moving to gate last @20nm, this supposed density advantage can't be that great vs the other benefits of gate last. Come on Chipworks, this wait is killing me:LOL:
 
A lot of things affect transistor density, many of them balancing. If the processes are targeted at the same niche, results from different foundries, at similar process nodes, have been surprisingly close considering just how many variables there are.

Simply principles of physics creating more common solutions. TSMC doesn't even have options at the 20nm node.

My prediction is that in all likelyhood the billion transistors in the A7 aren't partitioned just as in the A6, but that a greater portion of transistors are in highly regular (or otherwise intrinsically more dense) structures. Does that seem unlikely?

Not at all unlikely. Several members have pointed out some big transistor spending blocks potentially.

My potential counterpoint - did Apple creating the M7 move anything off of the A7 that was in the A6? I've even heard people speculate that the 1 billion is across A7 and M7, even though it was clearly on the A7 slide before they even introduced M7.

My feeling is that they have dedicated more die area to CPU than GPU. Some are suggesting that they have gone quad core, that would be a doubling right there (discounting the die shrink). Even if not, then they have used 64-bit CPU, which probably comes with an increase size.

The stated graphics performance, whilst clearly acceptable to Apple, implies a quite conservative Rogue implementation, which again might point towards die size limitations.

Any feeling that the high transistor count might be related to a big increase in my chip memory, given that memory is highly dense.
..edit, I see that's already been mentioned.

My question is whether or not they did a paradigm shift on logic density/circuit block approach. Lot of people have pointed out to me how much less dense the cortex A9 cores were on the A5 compared to other solutions like Tegra 2.

Wasn't the ongoing speculation for some time now that even when Apple start shifting to another fab, that the logistics would be such that they'd have to continue to dual source ?Perhaps my oversimplified conclusion from that was that they'd be getting the same soc from multiple manufacturers. For what people are saying here, it wouldn't be possible for Apple to state an soc size if it was coming from multiple sources.

TSMC is not part of the GloFo/Samsung/UMC foundry alliance, making me think the processes are less likely directly portable. My guess is that if they are indeed sourcing TSMC, that's why Apple isn't providing 5s pre-orders. They anticipate smaller availability. Of course, maybe it's the new fingerprint sensor or CMOS image sensor causing that too.

Is Hydra possibly the "MP" part of a multi-core implementation. I'm certain I've read that SGX544 is different from SGX544MP1 (the implication being that the MP circuitry brings more than just inter core glue).

After having it explained to me, that's my basic interpretation too.

True, that's why I didn't give it as the only example. It is pretty close to being a SoC though, as it's much more than just a CPU, GPU and SRAM on a single chip. It was all just meant as an example of showing that 1 billion transistors in 102 mm² isn't that unusual. The unusual part is Apple claiming that the A7 SoC has double the transistors of A6, or maybe we misunderstood Apple and they were comparing to something else.

Schiller made two claims. He verbally said twice the transistors. They showed 1 billion on the slide. He's not one to misspeak so I believe what he said.

Doesn't Samsung use a gate first approach on their 28nm node, which they trumpet as allowing higher densities, because of less restrictive design rules, apparently at the expense of yield %, compared to gate last. Given that everybody is moving to gate last @20nm, this supposed density advantage can't be that great vs the other benefits of gate last. Come on Chipworks, this wait is killing me:LOL:

Yes, they do gate first. Interestingly, they're claiming a gate density increase almost on par with the full node transition from 45nm to 32nm in this PDF. Looks like it's almost a 33% density improvement. I did not give enough credit. I thought they were doing 20% at best.

Interestingly, the gate density that accompanied TSMC's 28nm announcement in 2009 is 3900Kgates/mm^2. That's the same claim in the Samsung PDF. TSMC is gate-last.

Maybe the density between foundries really isn't all that different.
 
Yes, they do gate first. Interestingly, they're claiming a gate density increase almost on par with the full node transition from 45nm to 32nm in this PDF. Looks like it's almost a 33% density improvement. I did not give enough credit. I thought they were doing 20% at best.

Interestingly, the gate density that accompanied TSMC's 28nm announcement in 2009 is 3900Kgates/mm^2. That's the same claim in the Samsung PDF. TSMC is gate-last.

Maybe the density between foundries really isn't all that different.

I didn't realise that Samsung 32/28nm process was LP. I know that it's comparing apples with oranges but when I looked at the voltage tables for Krait 300 (LP) vs Krait 400 (HPM), HPM allowed the frequency to scaled much higher for equal or even lower voltage vs LP. If the A7 were fabbed at TSMC on its HPM process, clockspeed could be ramped much higher, without an increase in voltage.
I still think that Samsung is the likely candidate, due to logistical reasons, given the number of high volume 28nm customers already at TSMC, could they reliably supply the large numbers of wafers needed by Apple.
 
Independent of which process at which fab was used, a significant part of the increased density is likely resulting just from how Apple is implementing the cores' design/transistor layout. They're hitting their performance and power consumption/thermal targets without needing to spread the design across so much silicon as before (whether that's due to better design and technology this past year or so and/or them not being as ambitious with their SoCs' performance targets.)

The Apple TV's A5 appeared to be the start of their experimentation at optimizing for area more than before, and I feel their choice to go with a GPU configuration in the A6 which had less cores mixed with a higher clock speed was a continuation down that road. Now, the A7 seems to go even farther in that direction.

I don't think the somewhat underwhelming (at least compared to the perhaps-somewhat-unrealistic expectations I had) graphics performance of the A6 and A7 is a coincidence considering their apparent interest in optimizing more for area.
 
Actually that post seems more about criticizing the 64-bit CPU than whether the A7 has Rogue.

Having a 64-bit CPU along with more than 4 gigabyte of physical memory on a smartphone is practically pointless and will substantially decrease the battery life. Unless all the iOS apps are recompiled and optimized for 64-bit (something that will take years), there is no benefit of having a 64-bit CPU, other than to showoff. When you design a microprocessor, you try your best to cut down on power consumption, rather than making it consume more doing things that aren't even essential. ARM designed the 64-bit architecture for servers (read more here and here), since servers use a lot of physical memory and do things that require 64-bit wide registers but Apple’s move to make it available to smartphones is entirely stupid.
 
Just wondering here. Is it reasonable to think 64-bit has any significant advantages for these new tricks?

Thanks in advance!
ARMv8 mandates IEEE 754 compliance for NEON as well as adding double precision support for NEON. I don't know if Apple actually does image processing with such precision though.

ARMv8 also standardizes optional acceleration for AES encryption/decryption and SHA hashing which Apple may be using in combination with the fingerprint sensor.

Ben Bajarin claims that A7 is dual-core, not quad. I can't get him to elaborate why he's sure.

http://techpinions.com/the-apple-a7-a-is-for-ambition/22934

Another author outright claims A7 is 1.7 GHz but it being Rogue is just rumored. Super confused on that one.

https://medium.com/tech-talk/fb96c0d7fd4e
The 1.7GHz clock speed may be a holdover from the previous 31% faster rumour since that's a 31% higher clock than the 1.3GHz in the A6.

http://cannyvision.com/2013/09/12/the-most-forward-thinking-apple-yet.html

There's some thought that introducing 64-bit support now is less for smartphones than to prep for a true Apple TV/console which would need >4GB of RAM to be competitive with next-gen consoles and would help to have developers already familiar with 64-bit iOS to hit the ground running.
 
Back
Top