WildCat Lake with 2P and 4LPE CoresWCL? It’s abbreviation overload when searching for that, and I assume you don’t mean World Cricket League.![]()
WildCat Lake with 2P and 4LPE CoresWCL? It’s abbreviation overload when searching for that, and I assume you don’t mean World Cricket League.![]()
Since the context was video transcoding etc I searched for ”wcl video”, ”wcl computer”, and similar. None showed WildCat Lake.Come on it’s easy to search, “wcl intel”
First thing that pops up is Wildcat Lake
That would explain itSince the context was video transcoding etc I searched for ”wcl video”, ”wcl computer”, and similar. None showed WildCat Lake.
Isn't handbrake known to crap out at around 32t before you get diminishing returns?CPU encoding. Loves lots of cores.
So this is where fast cores are better than more cores?Isn't handbrake known to crap out at around 32t before you get diminishing returns?
![]()
Testing x265 encoder scaling on a 128 core Azure VM for 4K HDR
Video encoding has always been a fairly compute intensive task. With every new generation of video encoders, more computational power has been required to achieve the compression benefits. HEVC/H.265 is the latest state of the art video compression standard with x265 being the most popular...singhkays.com
![]()
edit: looking at some reviews of 9900X and 9950X, it looks like x265 and (even moreso) x264 are not scaling that well past 12c/24t. Basically getting at best ~20% scaling from 33% extra cores, and it only gets worse from there.
That’s probably 4K conversion as well. If you do 1080p or lower res, core scaling gets even worse.Isn't handbrake known to crap out at around 32t before you get diminishing returns?
![]()
Testing x265 encoder scaling on a 128 core Azure VM for 4K HDR
Video encoding has always been a fairly compute intensive task. With every new generation of video encoders, more computational power has been required to achieve the compression benefits. HEVC/H.265 is the latest state of the art video compression standard with x265 being the most popular...singhkays.com
![]()
edit: looking at some reviews of 9900X and 9950X, it looks like x265 and (even moreso) x264 are not scaling that well past 12c/24t. Basically getting at best ~20% scaling from 33% extra cores, and it only gets worse from there.
@DrMrLordX:
The first test was from 2018. So the MT implementation in Handbrake and the underlying codec libraries may have improved since then.
Then regarding 9950X vs 9900X (and for the first test too), could the power constraint also not explain why performance does not scale linearly with core count in this case? I’m thinking that the 9950X cores will run at lower frequency than on 9900X, due to same TDP constraint and the former CPU having more cores so less power/core.
Hmm… What would be the reason for that? Some resolution dependent bottleneck?That’s probably 4K conversion as well. If you do 1080p or lower res, core scaling gets even worse.
I wonder what would happen to the total processing time needed if you split the movie into two equal sized parts. Then run two Handbrake instances in parallel each transcoding one of the parts. And finally merge the two parts in the end.The 9900x has a much lower power limit than the 9950x and is just as power constrained. Video encoding just quickly hits diminishing returns once you get past 24t or so and that’s with high res (4K), modern formats. The vast majority of people will see even less benefit because they aren’t encoding that high res and are probably still using x264.
Once you get past the 24ish thread count I mentioned, then yes, at least generally speaking. Just look at where the 5950x sits in these results. Twice the cores/threads of the 7700x but still loses to it.So this is where fast cores are better than more cores?


Total time would go up because you’d still need to export the merged video at the end, so you’d just be wasting cycles.I wonder what would happen to the total processing time needed if you split the movie into two equal sized parts. Then run two Handbrake instances in parallel each transcoding one of the parts. And finally merge the two parts in the end.
Or perhaps even simpler: Run two (or more!) Handbrake instances in parallel, each transcoding a separate movie. Then on a 52C NVL-S there’ll be 26C per Handbrake instance. So the CPU core count scaling issue should have much less impact.
Not sure what you mean. Merging two MKV parts does not require transcoding. So it’s a much quicker operation.Total time would go up because you’d still need to export the merged video at the end, so you’d just be wasting cycles.
I would say it’s quite a normal scenario, but it depends on the user and use case. E.g. you could rip or download some 4K movies and then want to transcode them to reduce size. So you might as well transcode several of those movies in parallel.Edit: You can do multiple encodes at a time, but the vast majority of people aren’t producing multiple videos simultaneously and those that are typically use GPU encoding to massively speed up encoding time. The few real professionals who may do multiple videos and want to only use CPU encoding are pretty much all production studios who have server or Threadripper work stations to get the job done, they don’t touch consumer hardware.
Not sure what you mean. Merging two MKV parts does not require transcoding. So it’s a much quicker operation.
I would say it’s quite a normal scenario, but it depends on the user and use case. E.g. you could rip or download some 4K movies and then want to transcode them to reduce size. So you might as well transcode several of those movies in parallel.
Well, we’re talking about the subsection of the market that does video transcoding. That subsection may be small. But within that subsection I think transcoding several movies should be quite common. So then it might as well be done for several movies in parallel, especially going forward now if we’re going to see substantial speedups if doing it that way.This is not “normal” at all. The amount of consumers who do any kind of encoding is small to begin with. Those who want to do large batch processing without acceleration (i.e., only on the CPU) is nearly non-existent. This is coming from one of the very few people who have done it.
For the five people in the world who do this, sure.Well, we’re talking about the subsection of the market that does video transcoding. That subsection may be small. But within that subsection I think transcoding several movies should be quite common. So then it might as well be done for several movies in parallel, especially going forward now if we’re going to see substantial speedups if doing it that way.
If it really would be that few people who do video transcoding, then it”s strange that it’s a quite common benchmark to include in CPU reviews.For the five people in the world who do this, sure.
Could it be that such video encoding is inherently MP limited (I have no knowledge of the underlying algorithms so this might just be plain stupid)?Once you get past the 24ish thread count I mentioned, then yes, at least generally speaking. Just look at where the 5950x sits in these results. Twice the cores/threads of the 7700x but still loses to it.
View attachment 134583
If you bump up to 4K, things are different, but it still loses to a lower core count Zen 4 CPU. You can get different results with different settings and source videos, but the diminishing returns with large core counts holds true in pretty much every case.
View attachment 134584
No. I mean you should in theory just be able to split the movie into X parts and then transcode each of those parts in parallel completely independently, and join/merge the output of each part.Could it be that such video encoding is inherently MP limited
Not a fallacy if doing it like I described in my posts above.Anyway this shows expecting nice MT speedups even for a task that looks highly parallel is a fallacy.
I have tried to be very clear about it, but I’ll try again, there is a huge difference in the amount of people doing video transcoding and those doing it strictly on the CPU with parallel encodes. The latter is basically nonexistent. Even with single encode jobs, just because reviewers use it only means it’s an easy to run and repeatable test, it doesn’t mean it’s actually representative of what is done outside of reviews.If it really would be that few people who do video transcoding, then it”s strange that it’s a quite common benchmark to include in CPU reviews.
better yields always help, but yea smaller dies mitigate itThe dies are 55mm2 for Clearwater forest so yields shouldn't be a problem
The design cost difference should be mid/low double digits, and then you have to consider Intel foundries wafer cost prices solely from the process POV vs TSMC's (should be higher).It is possible that it is cheaper to produce than if we count margin stacking
They don't list it for their comparison to SRFIntel usually base it on SPEC In server like AMD does
You can always check Intel.com/performanceindex
Maybe, as David C1 pointed out Skymont got like half of the IPC uplift in DC as they did in clientand I feel like Intel also botched the L3 on Clearwater Forest hence the issue(based on Years of Intel Botching L3 in Client and Server) the fabric for sure is bad only 35 GB/s.
Well, we’re talking about the subsection of the market that does video transcoding. That subsection may be small. But within that subsection I think transcoding several movies should be quite common. So then it might as well be done for several movies in parallel, especially going forward now if we’re going to see substantial speedups if doing it that way.
