Onion+ and the associated cache management stuff may be Sony IP.
The hardware world will always move forward, but it's the software optimization that delivers the final experience.
I'm not saying that the Cell was a CPU + GPU I'm saying that the PS4 SoC with the CPU & GPGPU together is creating a Cell like Processor.
Jaguar Cores as the PPU & the GPGPU CU's as the SPE's.
People are looking right at it but still can't see what's happening.
mmm, yes, but did anybody here really thought Sony wouldn´t put any of its Research and Development IP inside something really transcendental into PS4?.
At least, if you don´t make a better GPU than what is in the market for more than a year give us some virtual reality stuff!.
I mean...isn´t somehow PS4 a little "generic"?. Well, it will be very developer friendly but still, Samsung could launch a similar console in three months with little R+D investment and Android OS taped inside.
mmm, yes, but did anybody here really thought Sony wouldn´t put any of its Research and Development IP inside something really transcendental into PS4?.
At least, if you don´t make a better GPU than what is in the market for more than a year give us some virtual reality stuff!.
I mean...isn´t in some way PS4 a little "generic"?. Well, it will be very developer friendly but still, Samsung could launch a similar console in three months with little R+D investment and Android OS taped inside.
1) The CPU cores can bypass cache and go straight to main memory (or straight to the GPU which is going to be sitting on the same memory bus). Nothing new here, SSE has had instructions to do so since the beginning and CPUs have routinely supported it, they utilize write coalescing buffers for it.
2) L2 cache has some kind of "volatile" setting - my guess is this basically makes a line write through instead of write back, so that cache doesn't have to be manually flushed to be accessible via the non-coherent link. But who really knows...
Without the SPEs, Cell would have been without qualification more than a little terrible.Without the SPE's the PS3 was a little "generic" but when devs put the SPE's to good use the PS3 shined & I think that will be the case with the PS4, when devs use the CU for compute tasks.
Without the SPE's the PS3 was a little "generic" but when devs put the SPE's to good use the PS3 shined & I think that will be the case with the PS4, when devs use the CU for compute tasks.
But if you use CUs for computing... where is the 7850 for rendering?.
But if you use CUs for computing... where is the 7850 for rendering?.
Of course it can do compute and graphics at the same time, like every other desktop GPU that's been on the market for the past few years.. Love_In_Rio is saying that you don't get the full rendering power of a 7850-level GPU if you're using some of the CUs for something else.
But if you use CUs for computing... where is the 7850 for rendering?.
Of course it can do compute and graphics at the same time, like every other desktop GPU that's been on the market for the past few years.. Love_In_Rio is saying that you don't get the full rendering power of a 7850-level GPU if you're using some of the CUs for something else.
& Cerny is saying that it can do both at the same time efficiently.
meaning that you don't have to take away the CUs because the CUs can do Graphics & Computing at the same time efficiently.
So you're suggesting the power of the CU's magically doubles if you're doing both graphics and compute work on them simultaneously since one type of workload has no negative impact on the other?
& Cerny is saying that it can do both at the same time efficiently.
meaning that you don't have to take away the CUs because the CUs can do Graphics & Computing at the same time efficiently.
Without confirmation, unfortunately, I can only point to speculation of Liverpool's CU's being capable of doing double instuction for rendering AND compute simultaneously through the fact that they're unified CU's for both combined with the compute rings doing simulteneous asynchronous compute.
That's called GCN.
The difference for the upcoming designs is that their ability to receive commands has been broadened and made more flexible, but it is functionally the same.
Or more likely, that's where the whole 14+4 thing originates. You can only reliably use 14 CUs for graphics as 4 CUs are somewhat reserved for compute tasks. Hence using those 4 CUs doesn't impact graphics rendering which is reliant on the 14 CUs.
Can't a regular AMD GCN GPU do that ? I meant performing different jobs simultaneously scheduled by different ACEs.
So you are saying they doubled up (duplicated) everything in the CU and then made it so you could only use half of it for GPU and half of it for compute? Because that's the only way what you says works. Otherwise, anytime a CU is being used for a compute workload it can't be used for a graphics workload.
Liverpool has multiple compute rings and pipes that provides fine control over how the system resources are divided amongst all of the application’s GPU workloads. This pipelines can perform simulteneous asynchronous compute.