Because after stumbling on this.
I can’t help but think of the possibility that the days of speeding up Blender via Moore’s law may be drawing to a close. Now it’s always subject to change when you consider that the nanometer scales will be going down a bit more over the next several years, but unless Intel decided to increase the focus on more cores and faster transistor/chip/material designs compared to increasing the power of the onboard GPU, there may not be a lot of hope for a major new leap in speed unless we really get technologies like graphite transistors, nano-tubes, and optical computing off the ground and into products. Granted, this is just initial information and it hasn’t been thoroughly tested yet, but I wouldn’t be surprised if Intel’s not far off the mark in speed.
So it looks to me like the primary increases in speed in software over the next few years will depend more on optimization techniques than the number of transistors on the die, especially when looking at the fact that Haswell boasts another major architectural change yet the speed increase is expected to be a relatively small 10 percent or maybe a little more. (and this is the ‘Tock’ part of Intel’s release cycle). Right now it seems like the major thing for speeding up software is newer, faster instruction sets, and to optimize Blender that way will mean having to slowly drop older sets over time much like how Ton made the decision to drop the original SSE sets.