Yup, thank you for clarifying this for readers who may not understand why we went away from single core high clocks to multi-core.It's a pretty well-investigated tradeoff in the server/datacenter world where power is a significant running cost. For the same performance, wide and slow is dramatically more power efficient than narrow and fast, but costs more. So it poses a really interesting tradeoff between fixed Si cost vs. variable runtime cost, and from what I've seen DC customers are more than happy to pay for Si in order to offset power costs.
In the client space it's not power but thermals that's a bigger issue, so instead of fixed-vs-variable cost tradeoff it's a fixed(Si)-vs-fixed(thermal design) cost tradeoff.
I would probably say that it's a well investigated physical limitation which led to the introduction of multi-core computing.
It really just comes down to voltage being cubic with clockspeed, and the faster you go, eventually your power demands get insanely high, but if you keep clockspeeds down and increase the number of cores, you gain computational power without introducing significant power and heat draw - of course at the cost of being able to probably feed both cores there is likely to be efficiency loss.
But yes, you are correct, in the datacenter world this is actually a thing. I was speaking purely about consumer GPUs, we just have power limitation challenges because we can only clock so fast before cooling is just impossible.
But you are right, when you're building cores that are extremely wide, and very few chips per wafer with a high defect rate, chips nearly cost the same as the wafer itself. I guess we are discussing a design philosophy on keeping costs down, I don't think there's any inherent 'architecture design philosophy' where someone would purposely choose wide and slow because it has benefits outside of power and cooling.