I agree on the slightly elevated pillars. When I mentioned possibly pressure fused at an elevated temp, the imprecision of just using an word descriptor causes a lot of grief.From my understanding, the TSV pillars are ever so slightly exposed above the silicon in the end, but I don't know how they are doing it with this new direct bonding technique.
For turning on and off effect, AMD said that the extra V-cache is striped with the existing L3 cache so as long as the addressing logic knows if it's on or off, it shouldn't be a problem but again, not a digital guy so I'm just guessing here.
The image you posted does not match with what you are saying. You seem to be looking at it upside down. The image is a diagram of the wafer in its original processing orientation. Device layers on the bottom and metal layers on the top. For TSVs, it would be flipped over and polished down from what is the bottom in the image. The TSVs go all of the way through the silicon to the other side, which I would think most people would say that it goes all of the way through the die. I don’t think it is relevant whether they add more metal later to connect it to a specific layer or even build it up enough to be an IO pad directly. There is polishing steps after each deposition, so that is just extra metal layer down later. It still goes all of the way through the silicon to the opposite side of the wafer after polishing.I thought I had made it pretty clear in my post but I guess not. Some TSVs go through the substrate and attach to (stop at) the FEOL, some go thorough substrate and attach to (stop at) the BEOL, and some go fully through the die including substrate, FEOL, and BEOL. There isn't just one way of doing TSVs though my understanding that attaching to BEOL has been more or less adopted as the industry standard as via first (FEOL stop) forces the TSV material to go through high heat when the FEOL is processed which limits the types of material you can use for the TSV and via last makes it difficult to align with the rest of the circuit but can be easier when you want to go through a full die. In every case the TSV goes through the substrate and is exposed at the bottom of the die but how far up the die it goes can vary.
I really don't think going decent depth with a TSV is a yield issue but I don't have any data on that specifically. I do know that there are papers showing you can do 200 um+ TSVs without issue but papers don't usually have high volume yield in mind and this is most likely more a limitation in relation to TSV diameter/pitch. I'm also sure processing wafers that are thinned to absolute thinnest possible also has introduces its own issues. I'll have to read up on how they handle these issues over the weekend and see what the gives and takes are.
1)Pure Cu is a very "soft" metal, much more ductile than Si. Easily possible
2)If you magnify the image, you'll see the the uniformly cylindrical pillars slightly penetrating the cone shaped top & bottom connectors.
3)The boundary of the bond is definitely not flat. All of them have a saucer shaped depression.
4)Cold welding doesn't have to mean cold temps, just lower temps than the melting/crystallization temps point. Pressure is often used.
4)The sides of the pillars are supported by the surrounding material, preventing compressive buckling failure and distortion.
5) All of the articles do not give details, only giving the illusion of understanding.
I am interested in the actual details of how this is done
However, one is free to believe that you just rest them together and the weld happens. No problem.
The image you posted does not match with what you are saying. You seem to be looking at it upside down. The image is a diagram of the wafer in its original processing orientation. Device layers on the bottom and metal layers on the top. For TSVs, it would be flipped over and polished down from what is the bottom in the image. The TSVs go all of the way through the silicon to the other side, which I would think most people would say that it goes all of the way through the die. I don’t think it is relevant whether they add more metal later to connect it to a specific layer or even build it up enough to be an IO pad directly. There is polishing steps after each deposition, so that is just extra metal layer down later. It still goes all of the way through the silicon to the opposite side of the wafer after polishing.
What does that actually mean though? If anyone has some article that actually describes how it integrates with the existing L3 I'd appreciate it.
Outside of just increasing the number of sets that are available, being able to turn the entirety of that additional cache on or off would require two different sets of logic to handle where anything gets stored in the cache. It would also mean that switching between the two states would require moving all of the cache entries around to where they belong under the other state since there's no guarantee that they wind up in the same place.
If it's just extra sets it turns the L3 from a 16-way associative cache into a 48-way associative cache. That probably has some significant diminishing returns, but maybe it's useful for certain applications like games that can easily use up several GB of memory and would benefit from having some of that data stick around in the cache for much longer.
I am not saying you are confused. I am saying you are using different meanings for words than I or most others would use. I consider all of the forms of TSVs (Through Silicon Vias) to go through the die. If you want to not count it as going “through the die” if it isn’t routed to the top of the metal stack, then so be it.I'm not confused at all. The whole point of that image was to show that there are currently 3 different ways to do TSVs. Two of the ways the TSV is etched and deposited before the wafer is finished. In via first, it is done before the FEOL, in via mid it is done after FEOL but before BEOL. In both of those cases the wafer is then completed after TSV formed, flipped, and thinned to expose the TSVs on the backside. In the last way (via last), the FEOL and BEOL are both finished, the wafer is flipped, then thinned, and then the TSV is made. Only in the via last method can you take the TSV through the entire die because you bring it in from the backside and can then go through as much of the die as you want. You are building the TSV from the bottom up. For via first and via mid, you can't do this because you build from the top down and are starting in the first layer or two. You cannot take via first or via mid all the way through the die.
Can you link timestamp where that was said ? Thanks 🙂On this video, AMD says that the was a regression on performance on the two chiplets using 3D V Cache. From what we have seen I can see that too. On the 5900X3D Prototype was clocked at 4 Ghz and the 5900X was also clocked at 4 GHZ, they still showed 15% performance improvement but at ISO Speed, the performance would have been close to 12% the 5900X3D was clocked at 4.5 and the 5900X remain with stock boost of 4.8 Ghz. This is the case because games can't take advantage of the added L3 Cache that is on another chiplet.
![]()
The Full Nerd Special Episode: AMD talks V-Cache, Ryzen 6000, AM5, and laptops
Will we see a Ryzen 6000 laptop with the best GeForce GPUs this year? And why did the company ditch its beloved AM4 pins?!www.pcworld.com
Don't you mean interlaced?AMD said that the extra V-cache is striped with the existing L3 cache
Sorry, I can't. Just watch the video from the 4 minute markCan you link timestamp where that was said ? Thanks 🙂
As for pricing I expect AMD to stay true to its price/perf ratio which at the regularly cited ~15% perf improvement would amount to ~$517.5. So an MSRP outside $499 to $519 would be a slight surprise to me.
I'd also expect 5800X3D to be a good overclocker, stock specs playing it safe.
Don't you mean interlaced?
On this video, AMD says that the was a regression on performance on the two chiplets using 3D V Cache. From what we have seen I can see that too. On the 5900X3D Prototype was clocked at 4 Ghz and the 5900X was also clocked at 4 GHZ, they still showed 15% performance improvement but at ISO Speed, the performance would have been close to 12% the 5900X3D was clocked at 4.5 and the 5900X remain with stock boost of 4.8 Ghz. This is the case because games can't take advantage of the added L3 Cache that is on another chiplet.
![]()
The Full Nerd Special Episode: AMD talks V-Cache, Ryzen 6000, AM5, and laptops
Will we see a Ryzen 6000 laptop with the best GeForce GPUs this year? And why did the company ditch its beloved AM4 pins?!www.pcworld.com
What he says is that if you had 1 CCD with V-cache and 1 without, you might have a performance reduction due to the asynchronous caches. He comments later on if they both have V-cache but doesn't talk about performance regression. Honestly, the answer seemed more like a marketing response just beating around the bush trying to come up with some reason why they only have a 5800x3d because I don't think his comments about the synchronous cache having extra latency is even accurate or makes sense.
Hey,
Can you post this as a reply to your comment. I'm banned from Amd and /Hardware for leaking info I wasn't supposed to let ou
I know there are working samples of 5800Xs using a single 8 core CCX made on the 6nm node. 5Ghz+ AC OCs are easily attainable but it remains to be seen when/if that launches.
Even without Vcache, the move to 6nm and the more refined layout and I/O are allowing IF clocks up to 4200 running memory 1:1. With tuned B die this allows for near parity with the 12600K. This is quite impressive given it does not have the 3D cache.
Keep your fingers crossed😉
It isn't a straight shrink, it is a library swap.Supposedly, a straight shrink to N6 is supposed to be a minimal effort process.
What he says is that if you had 1 CCD with V-cache and 1 without, you might have a performance reduction due to the asynchronous caches. He comments later on if they both have V-cache but doesn't talk about performance regression. Honestly, the answer seemed more like a marketing response just beating around the bush trying to come up with some reason why they only have a 5800x3d because I don't think his comments about the synchronous cache having extra latency is even accurate or makes sense.
I made a comment over on /r/Amd in regards to V-Cache sku's before the announcement. Shortly after I received a private message. Only just saw it now, but the info is 8 days old. Haven't posted it at Reddit, but thought that a few here might find it interesting
Can you link timestamp where that was said ? Thanks 🙂
Yeah, I don't fully buy the "you might actually lose a little performance" when going with dual chiplets, the same logic could have been applied to 5800X vs. 5900X at the time of launch, and the 5900X won or traded blows with 5800X in gaming. That being said, the extra cache on both parts would probably help the single die product more, essentially matching the dual die product in most games.Ah nevermind, i found the comment