So you didn't read what I said. Got it.any time someone says that BD has any weaknesses you link that article to say those weaknesses aren't actually weaknesses at all.
So you didn't read what I said. Got it.any time someone says that BD has any weaknesses you link that article to say those weaknesses aren't actually weaknesses at all.
Well, if we take Rory's statements as being true then he has already written off AMD doing either (a) or (b).
Big cores and leading edge process nodes are both written off as being the old way the old AMD approached the market. It would make him a liar if he tells the analysts that but then internally he keeps spending and prioritizing the development of big cores on leading edge nodes.
Now the reality is Rory had little choice but to make that his strategy because the cost associated with developing big cores is immense, and it gets even more immense as you go to newer nodes. AMD simply doesn't have the cash to do that anymore, its not a matter of willpower and desire.
If you listened to that last conference call that mrmt linked up with Rory talking about where AMD is going I think it is pretty clear that Rory is prioritizing AMD's remaining resources towards furthering the bobcat/jaguar APU lineage going forward.
It will still be competitive, no one else is capable of marrying x86 compatibility to GPU capability like AMD can. But the x86 compatibility puts them into Via territory for niche market spaces (like the PS4) that others can't get into while the GPU IP keeps them out of reach from pretty much all others (they can easily defend themselves from would-be intruders into their niche market spaces once they are established in them).
The question is what sort of TAM does that leave AMD to play in? Is it enough TAM to enable a $6B/yr revenue or is the TAM from all those niche markets only large enough to support a $2B/yr revenue model?
It is obvious that AMD is going to become a smaller fish as they try to find ponds in which they can survive. But how much smaller is that going to be?
As for (a) and (b) above, I personally doubt we'll see AMD finish out Excavator, and I also don't see how AMD could hope to find the cash needed to support the R&D of building next-gen after Steamroller. I think Kaveri is going to be it for big cores from AMD unless someone swoops in and hands them $10B or so.
So you didn't read what I said. Got it.
You said that it was actually primarily only due to three reasons:Darek Mihocka also did some low level analysis of Bulldozer and he gave us a hint on why Bulldozer is so slow:
Then I said:Johan De Galas here at Anandtech found it to be primarily comprised of three things: low clock speeds (relative to the pipeline length), L1 instruction cache is too small, branch misprediction penalty is too large. Cache latency was not a major factor.
Then you have the gall to repeatedly telling me I'm not reading. Maybe the problem is I can't read whatever point it is you're not expressing. If it's that stuff about its problems being more "low level" than L2 latency and about improving L2 latency being "treating the symptom" that's just a load of crap. L2 latency to L1 data cache actually does matter and increasing the size of L1 icache 1.5x or whatever it is won't always reduce the miss rate as much or even at all.mrmt said one person had various criticisms, you replied to say that Johan De Gelas determined that there issues with BD were primarily due to three problems. No where did anyone say anything about desktop or server apps.
With the CPU area of Trinity, at the same node, Intel could fit two cores *and* a GPU.
With the area of Bulldozer, Intel could fit one and a half SNB.
Why would you compare a 3820? How about a LGA 1155 processor which is a much more accurate comparison both in terms of cost and performance which comes in at 216mm2
Only reason to compare a 3820 is to make AMD's die sizes not seem as big as they really are.
Why ?? because that is the only Intel CPU to compare it apples to apples. Almost same transistor count for the CPU without any iGPUs (FX has 1.2B transistors vs 1.27B transistors for the Core i7 3820). Both were made for the Server first and Desktop secondly.
Why ?? because that is the only Intel CPU to compare it apples to apples. Almost same transistor count for the CPU without any iGPUs (FX has 1.2B transistors vs 1.27B transistors for the Core i7 3820). Both were made for the Server first and Desktop secondly.
Important to note that the 3820 has 2 extra memory controllers and 24 more lanes of PCIe.
Right, because you buy your processors based on transistor count right? And all the comparison reviews? They chose their comparison reviews based on transistor count to right?
I'm going to let you in on something. When people buy a processor their main factors are going to be price and performance. NO one I know buys a CPU based on transistor count. Hence, your comparison is anything but apples to apples, and you know it.
The ONLY reason you compared a 2011 processor is to make AMD's die size look more efficient than it really is. There really is no other reason, at least none that makes sense. Transistor count... Jesus.
We are not talking about purchasing the CPU, we are talking about die sizes, transistors and microarchitectures. So we compare at the same transistor count, same die sizes and same node process etc.
Maybe it is you that trying to make the Intel 1155 look more than it is.
Yes, and the FX8350 has 7MB more L2 cache(8+8 vs 1+10) and 4 HyperTransports(two of them are disabled for the desktop CPUs).
Yes, and Intel's comparable SB processor has a die size 216mm2. Period.
You're the only one using LGA2011 as a comparison, that should tell you something, and it isn't that you're right and everyone else is wrong.
Do fused off parts of the die get reported in transistor count?
Too bad that MT performance still doesn't matter for most use cases, especially on low-cost CPUs. People who do 3D rendering, number crunching, video transcoding/editing, etc. will rarely do that on low-end systems. And for most other tasks, single-threaded performance is still far more relevant.AMD Module (excluding L2) at 32nm = 19,42mm2 with 67M tranistors
Intel SandyBridge Core (+HT) (excluding L2) at 32nm = 16,5mm2 with 55M transistors.
The difference in size is only 18% or 2.92mm2, that translates to Trinity being faster in the majority of the MT apps against the Core i3(SB).
Still got it wrong. The "symptom" comment had nothing to do with the L1 I-cache.Then you have the gall to repeatedly telling me I'm not reading. Maybe the problem is I can't read whatever point it is you're not expressing. If it's that stuff about its problems being more "low level" than L2 latency and about improving L2 latency being "treating the symptom" that's just a load of crap. L2 latency to L1 data cache actually does matter and increasing the size of L1 icache 1.5x or whatever it is won't always reduce the miss rate as much or even at all.
AMD has obviously been able to bring out both Llano, Trinity and soon Kaveri despite the economic constraints you mentioned. Why shouldn't they be able to continue like that going forward?
Also, according AtenRa the statement from Rory does not necessarily have to be interpreted as AMD leaving the big core x86 desktop CPU market segment.
Still got it wrong. The "symptom" comment had nothing to do with the L1 I-cache.
Really, you're just putting words in my mouth. Knock it off already.
I'm sure a lower latency L2 would negate the issue, but that would be treating the symptom and not the problem in this circumstance
Johan De Galas here at Anandtech found it to be primarily comprised of three things: low clock speeds (relative to the pipeline length), L1 instruction cache is too small, branch misprediction penalty is too large. Cache latency was not a major factor.
AMD use the same transistor count both for the desktop and server parts, so i believe they count them.
Why ?? because that is the only Intel CPU to compare it apples to apples. Almost same transistor count for the CPU without any iGPUs (FX has 1.2B transistors vs 1.27B transistors for the Core i7 3820). Both were made for the Server first and Desktop secondly.
Too bad that MT performance still doesn't matter for most use cases, especially on low-cost CPUs. People who do 3D rendering, number crunching, video transcoding/editing, etc. will rarely do that on low-end systems. And for most other tasks, single-threaded performance is still far more relevant.
Wait, this is not true at all. The FX line is designed for gamers originally, while LGA2011 is for "extreme enthusiasts". If those CPUs were marketed for server, then Intel and AMD wouldn't need their Xeon and Opteron lines. Sandy Bridge i7-2700K has 1.16 billion transistors, which is closer to the 1.2 billion of 8-core FX than the 2011 6-core is. Furthermore, the two are closer in capability--both overclock and have the same number of cores/threads. If anyone is trying to weasel out a non-apples-to-apples comparison, it is you.
This is the same company that couldn't figure out how to count the transistors in Bulldozer.