The problem I have with that assessment is that I don't think it justifies Bulldozer being a downgrade from the K7-derived family of cores for so many workloads. I'd think it would have been much cheaper, and lower risk, to continue incrementally improving K7/K8/Family 10h/Husky rather than trying a new design as complex as Bulldozer
It is clear that there were alternatives to Bulldozer inside AMD but somehow the Bulldozer crew won the argument inside the company and got the funds and resources needed to develop their project. This is when they lost it.
With 20/20 hindsight the best alternative would have been to kill Bulldozer in the design stage and come with a more conventional uarch based around phenom or something along Intel lines, beefier cores with SMT.
One of the reasons that I think they didn't go with a phenom die shrink is that they would have lost a lot of work they did for Trinity.
That said, a new die shrink of phenom processors may have been cheaper for AMD, and even improved their competitive position a bit, but it would not solve AMD fundamental problem, which is that Intel is opening a HUGE gap in both IPC and performance per watt on top of the process gap.
I guess what I'm saying is, Bulldozer was too complex in too many dimensions (microarchitecture, custom circuit design, semicustom implementation, etc), and AMD could likely have done better with fewer resources by simply biting off less.
I don't buy this argument that AMD could not correctly predict the workloads that servers and desktops would be running in 2012. Bulldozer isn't really competitive in performance per area and performance per watt with Core in both server and desktop workloads. In servers, AMD cannot even match Intel EP line, let alone the EX line.
Simply put, Bulldozer does not work even in their best case scenario. This does not point out to bad execution, but for a fundamentally flawed concept, much like Netburst was. No matter how Intel tuned Netburst, they could not consistently beat AMD, and Intel engineers had nothing to blame, no bug in the chip, no failure in the process, just the concept.
I think licensing an ARM core saves piles of money. Verifying an x86 core takes huge amounts of manpower and compute infrastructure (they have thousands of machines running tests 24/7 through the whole design process). Buying a known-good core eliminates all of that.
It does, but it hurts your market position.
If I am to buy an ARM server, I can ask prices to AMD and, let's say, generic manufacturer XYZ that also sells vanilla A5x based SoCs. If my design isn't too complex to the point of Freedom Fabric not being the deciding factor but a nice plus, as there isn't any differentiation between the two chips, I can just make the two race to the bottom.
AMD may win the order but at razor-thin margins. This model could work for Seamicro alone, but it does not work for a company that designs its own chips.