My question here is : What in the world is this use case that everyone's talking about that's causing seg faults? What is being compiled and how many times? Is the use case a kernel compilation done over and over again w/o stop? Why the creation of a ramdisk? Is memory usage ballooning over time?
This thread is for people that are BUILDING TR systems, not whiners about it. And I have 2 Ryzen systems that have never failed me, and have run 24/7@100% load doing DC.This is getting OT for this thread. If you're cool using an unreliable processor (you ran the test, it failed, your machine is affected) then have at it. You'll probably see random processes die at random times so infrequently that you'll put it down to something else. How many people got random and annoying crashes when overclocking before someone thought of running software that made things crash faster to highlight instability? The looped compile is a stability test. It doesn't mean the system is stable unless you are compiling large software. Compiling software just makes it happen quicker to make a nice easy to reproduce test. The ram disk is just to stop it hammering your SSD while running the tests.
I caught it originally when compiling a new kernel for the machine and it bombed. I've had a windows VM die after 6 weeks and I've had CUPS die twice in 2 weeks. Stuff shouldn't just die, so I did some searching and found a whole lot of other people complaining about a similar thing. Apparently AMD says it affects a tiny proportion of Linux users. It probably does, but at the moment that's you and I. You may not care, but that doesn't mean it doesn't affect your system.
This'll be my last post on it in this thread. I'm happy to elaborate further if you want to create a thread on it (or better still, just read the AMD forum thread I linked originally).
I forgot which video or article I read talking about it, but the active dies are caddy corner from each other. This helps to spread the heat to prevent hotspots. This means the cooler will cover most (if not all) of the dies while cooling a large portion of the rest of the ihs. Remember, heat seeks cold. So I'd expect it to be fairly effective. Keep an eye on core temp variance when setting clocks and voltages and you should be more than fine.TR including Asetek brackets for the ability to mount existing AIO's is fine, but some of these coolers have a circular contact plate, and with the massive size of the heatspreader, how well would it actually cool?
This is getting OT for this thread. If you're cool using an unreliable processor (you ran the test, it failed, your machine is affected) then have at it. You'll probably see random processes die at random times so infrequently that you'll put it down to something else. How many people got random and annoying crashes when overclocking before someone thought of running software that made things crash faster to highlight instability? The looped compile is a stability test. It doesn't mean the system is stable unless you are compiling large software. Compiling software just makes it happen quicker to make a nice easy to reproduce test. The ram disk is just to stop it hammering your SSD while running the tests.
I caught it originally when compiling a new kernel for the machine and it bombed. I've had a windows VM die after 6 weeks and I've had CUPS die twice in 2 weeks. Stuff shouldn't just die, so I did some searching and found a whole lot of other people complaining about a similar thing. Apparently AMD says it affects a tiny proportion of Linux users. It probably does, but at the moment that's you and I. You may not care, but that doesn't mean it doesn't affect your system.
This'll be my last post on it in this thread. I'm happy to elaborate further if you want to create a thread on it (or better still, just read the AMD forum thread I linked originally).
True enough. Look at even a stock Ryzen wraith spire... The Contact plate is circular.I forgot which video or article I read talking about it, but the active dies are caddy corner from each other. This helps to spread the heat to prevent hotspots. This means the cooler will cover most (if not all) of the dies while cooling a large portion of the rest of the ihs. Remember, heat seeks cold. So I'd expect it to be fairly effective. Keep an eye on core temp variance when setting clocks and voltages and you should be more than fine.
Take a look at their TDP guide: the new socket TDP values are not listed, but you can check AM3 ratings since that socket featured 220W CPUs.So, i might end up w/ a noctua cpu cooler while throwing their ugly arse poo colored fan in the trash. Thing is if they are even rated to handle 180W. I went to the page and all of them indicate 100W ratings
I forgot which video or article I read talking about it, but the active dies are caddy corner from each other. This helps to spread the heat to prevent hotspots. This means the cooler will cover most (if not all) of the dies while cooling a large portion of the rest of the ihs. Remember, heat seeks cold. So I'd expect it to be fairly effective. Keep an eye on core temp variance when setting clocks and voltages and you should be more than fine.
I still have my doubts, but we'll see. Circular contact plates isn't a problem for AM4 because the heatspreader is square and of a much smaller size, plus the die is bang in the middle, so everything works out in terms of symmetry. For reference, the TR4 edition of the Noctua air coolers have a contact plate of dimension 70x56mm, which is what you'd really want if these schematics are to be believed.True enough. Look at even a stock Ryzen wraith spire... The Contact plate is circular.
That PDF looks interesting - will have a look at it. Thanks.My last response.
The processor is stable as proofed by its ability to handle sustained executional benchmarks.
It is proofed by the sheer number of people running the processor w/o issue. Beyond that is software, corner cases, and stateful data issues .. aka caching subsystems.
The issue seems to be segmented to a specific compilation use case which makes it a corner case with assigned internal priority. I went through the thread found here : https://community.amd.com/thread/215773?start=525&tstart=0 (90% of the people seem to be quite uninformed on how processors function). They hit a corner case issue and are screaming bloody murder. The main execution pipeline seems fine. Where the problem seems to be centered is on OP Caching and other supportive structures...
http://www.anandtech.com/show/10578...lers-micro-op-cache-memory-hierarchy-revealed
Bugs in such optimization subsystems get worked out as they for all processors.
https://github.com/xoreaxeaxeax/sandsifter/blob/master/references/domas_breaking_the_x86_isa_wp.pdf
This is likely why the problem goes away when you disable OP Caching...
I really should get paid to debug issues like this. Who knows, maybe there is something in store for this bug 😉
and no AMD isn't going to disclose anything of note about what the issue is if they've found it and are working on it. This section of the processor doesn't get detailed much. As for other random issues you have (6 WEEKS UPTIME), it's software ... Congrats on 6 weeks uptime.
If by that you mean which company's board aesthetic is the least offensive, then AsRock has you covered - they're the only ones providing dual-Intel GbE as well, at a reasonable price. Otherwise, if you meant something akin to the WS lineup from Asus, I'm afraid it'll have to wait.Has there been any news of workstation motherboards for TR (not just "gamer" models)?
https://www.amd.com/en/thermal-solutions-threadripperQuestion :
Do AIO water coolers come w/ a damage warranty w.r.t warrantee/guarantee against component damage in case these things leak?
*Taking question over to :
https://forums.anandtech.com/forums/cases-cooling.13/
Will drop inquiry details here once I get them. Trying to resolve going AIO or air cooling for threadripper. The nightmare scenario of a leak on such an expensive system + expensive GPUs and subsequent potential data loss doesn't seem worth it 🙁
Cringe : https://forums.anandtech.com/thread...-trashed-could-i-have-prevented-this.2512319/
So, i might end up w/ a noctua cpu cooler while throwing their ugly arse poo colored fan in the trash. Thing is if they are even rated to handle 180W. I went to the page and all of them indicate 100W ratings
So, you will pay $400 for 3200MT/s ram, when 4133 costs $440-450 for 4x8GB. Considering the Zenith and MSI boards now supposedly run 3600, but the QVL lists none, it means it runs it, but not at the rated speed (XMP profile). So, if you are willing to manually set the ram OC, you get ram that will last you longer and run faster at a small premium. Or maybe kill thinking about this wrong...![]()
Will be cooled by my EVGA CLC 280mm.
Haven't had an AMD Processor since the AMD 64 3200+ back in 2005.
I added a 2nd 1TB 960 Pro NVMe drive to this purchase. Going completely NVMe with this build. No SATA cables and cutting back on power cables too. Will be transferring over my 512 GB 950 Pro NVMe drive from my current build.
So, you will pay $400 for 3200MT/s ram, when 4133 costs $440-450 for 4x8GB. Considering the Zenith and MSI boards now supposedly run 3600, but the QVL lists none, it means it runs it, but not at the rated speed (XMP profile). So, if you are willing to manually set the ram OC, you get ram that will last you longer and run faster at a small premium. Or maybe kill thinking about this wrong...
Yes and no. The website still says 3200 on the Zenith and MSI boards, but the bulk of news coverage since QVLs at the different companies were published went from as marked to Zenith and MSI supporting 3600.It's confirmed 3600 works? Also I doubt theres 3600 ram with CAS 14 timings.

Thanks, I just caught that. They seemingly need to update this with the threadripper product's specs. Indeed U14S is good to 220W. U12S isn't spec'd beyond 140W besides mention and U9S is hard stopped at 125W which is way below 180W. Thus, they need to update the spec w/ what these things can actually achiveTake a look at their TDP guide: the new socket TDP values are not listed, but you can check AM3 ratings since that socket featured 220W CPUs.
The U14S is good for 220W.
The U12S goes beyond 140W but does not reach 220W.
The U9S is only 125W, but the new TR4 model comes with one more heatpipe and an additional fan for push-pull config.
I expect all of them to handle 180W, while U14 will go beyond stock easily.
Just did some back hand analysis found here :I still have my doubts, but we'll see. Circular contact plates isn't a problem for AM4 because the heatspreader is square and of a much smaller size, plus the die is bang in the middle, so everything works out in terms of symmetry. For reference, the TR4 edition of the Noctua air coolers have a contact plate of dimension 70x56mm, which is what you'd really want if these schematics are to be believed.
Yep. Confirms exactly what I said. However, I went a step further and specified the bare minimum diameter needed for actual coverage of the dies by the contact surface. The diameter has to be 51.9mm. This is backed up by :Gamers Nexus on this topic:
But has anyone actually given a thought as to why these CPU's are so heavy? Yes, they are big, but not that big (~130 grams). My bet is that the IHS is much thicker than the ones on most the other CPU's and therefore spreads the heat across itself much more effectively. So its not even essential to have 100% of the die area to be covered by thermal paste/cooler base. Quess we'll find out soon. What do you guys think?