64 Cores of Rendering Madness: The AMD Threadripper Pro 3995WX Reviewby Dr. Ian Cutress on February 9, 2021 9:00 AM EST
- Posted in
- Threadripper Pro
Knowing your market is a key fundamental of product planning, marketing, and distribution. There’s no point creating a product with no market, or finding you have something amazing but offer it to the wrong sort of customers. When AMD started offering high-core count Threadripper processors, the one market that took as many as they could get was the graphics design business – visual effects companies and those focused on rendering loved the core count, the memory support, all the PCIe lanes, and the price. But if there’s one thing more performance brings, it’s the desire for even more performance. Enter Threadripper Pro.
computational graphics goes brrrrrrr
There are a number of industries that, when looking from the outside, an enthusiast might assume that using a CPU is probably old fashioned – the question is asked as to why hasn’t that industry moved fully to using GPU accelerators? One of the big ones is machine learning – despite the push to dedicated machine learning hardware and lots of big businesses doing ML on GPUs, most machine learning today is still done on CPUs. The same is still true with graphics and visual effects.
The reason behind this typically comes down to the software packages in use, and the programmers in charge.
Developing software for CPUs is easy, because that is what most people are trained on. Optimization packages for CPUs are well established, and even for upcoming specialist instructions, these can be developed in simulated environments. A CPU is designed to handle almost anything thrown at it, even super bad code.
By contrast, GPU compute is harder. It isn’t as difficult as it used to be, as there are wide arrays of libraries that enable GPU compilation without having to know too much about how to program for a GPU, however the difficulty lies in architecting the workload to take advantage of what a GPU has to offer. A GPU is a massive engine that performs the same operation to hundreds of parallel threads at the same time – it also has a very small cache and accesses to GPU memory are long, so that latency is hidden by having even more threads in flight at once. If the compute part of the software isn’t amenable to that sort of workload, such as being structurally more linear, then spending 6 months redeveloping for a GPU is a wasted effort. Or even if the math works out better on GPU, trying to rebuild a 20-year old codebase (or older) for GPUs still requires a substantial undertaking by a group of experts.
GPU compute is coming on leaps and bounds ever since I did it in the late 2000s. But the fact remains is that there are still a number of industries that are a mix of CPU/GPU throughput. These include machine learning, oil and gas, financial, medical, and the one we’re focusing on today is visual effects.
A visual effects design and rendering workload is a complex mix of dedicated software platforms and plugins. Software like Cinema4D, Blender, Maya, and others rely on the GPU to showcase a partially rendered scene for these artists to work on in real time, also relying on strong single core performance, but the bulk of compute for the final render will depend on what plugins are being used for that particular product. Some plugins are GPU accelerated, such as Blender Cycles, and the move to more GPU-accelerated workloads is taking its time – ray tracing accelerated design is an area that is getting a lot of GPU attention, for example.
There are always questions as to which method produces the best image – there’s no point using a GPU to accelerate the rendering time if it adds additional noise or reduces the quality. A film studio is more than likely to prioritize a slow higher-quality render on CPUs than a fast noisy one on GPUs, or alternatively, render a lower resolution image and then upscale with trained AI. Based on our conversations with OEMs that supply the industry, we've been told that a number of studios will outright say that rendering their workflow on a CPU is the only way they do it. The other angle is memory, as the right CPU can have 256 GB to 4 TB of DRAM available, whereas the best GPUs can only supply 80 GB (and those are the super expensive ones).
The point I’m making here is that VFX studios still prefer CPU compute, and the more the better. When AMD launched its new Zen-based processors, particularly the 32 and 64 core count models, these were immediately earmarked as potential replacements for the Xeons being used in these VFX studios. AMD’s parts prioritized FP compute, a key element in VFX design, and having double the cores per socket was also a winner, combined with the large amount of cache per core. This latter part meant that even though the first high-core count parts had a non-uniform memory architecture, it wasn’t as much of an issue as with some other compute processes.
A number of VFX companies as far as we understand focused on AMD’s Threadripper platform over the corresponding EPYC. When both of these parts first arrived to market, it was very easy for VFX studios to invest in under-the-desk workstations built on Threadripper, while EPYC was more for the server rack installations and not so much for workstations. Roll around to Threadripper 3000, and EPYC 7002, and now there are 64 cores, 64 PCIe 4.0 lanes, and lots of choice. VFX studios still went for Threadripper, mostly due to offering higher power 280 W in something that could easily be sourced by system integrators like Armari that specialize in high-compute under-desk systems. They also asked AMD for more.
AMD has now rolled out its Threadripper Pro platform, addressing some of these requirements. While VFX is always core compute focused, the TR Pro now gives double the PCIe lanes, double the memory bandwidth, support for up to 2TB of memory, and Pro-level admin support. These PCIe lanes could be extended to local storage (always important in VFX) as well as large RAMDisks, and the admin support through DASH helps keep the company systems managed together appropriately. AMD’s Memory Guard is also in its Pro line of parts, which is designed to enable full memory encryption.
Beyond VFX, AMD has cited world leadership compute with TR Pro for product engineering with Creo, 3D visualization with KeyShot, model design in architecture with Autodesk Revit, and data science, such as oil and gas dataset analysis, where the datasets are growing into the hundreds of GB and require substantial compute support.
Threadripper Pro vs Workstation EPYC (WEPYC)
Looking at the benefits that these new processors provide, it’s clear to see that these are more Workstation-style EPYC parts than ‘enhanced’ Threadrippers. Here’s a breakdown:
|AMD Zen 2 High-End Comparison|
|1P Flagship||TR 3990X||TR Pro 3995WX||EPYC 7702P|
|TDP||280 W||280 W||200 W|
|Base Freq||2900 MHz||2700 MHz||2000 MHz|
|Turbo Freq||4300 MHz||4200 MHz||3350 MHz|
|L3 Cache||256 MB||256 MB||256 MB|
|DRAM||4 x DDR4-3200||8 x DDR4-3200||8 x DDR4-3200|
|DRAM Capacity||256 GB||2 TB, ECC||4 TB, ECC|
|PCIe||4.0 x56 + chipset||4.0 x120 + chipset||4.0 x128|
To get these new parts starting from EPYC, all AMD had to do was raise the TDP to 280 W, and cut the DRAM support. If we start from a Threadripper base, there are 3-4 substantial changes. So why is this called Threadripper Pro, and not Workstation EPYC?
We come back to the VFX studios again. Having already bought in to the Threadripper branding and way of thinking, keeping these parts as Threadripper helps smooth that transition – this vertical had kind of already said they preferred Threadripper over EPYC, from what we are told, and so keeping the naming consistent means that there is no real re-education to do.
The other element is that the EPYC processor line is somewhat fractured: there are standard versions, high performance H models, high frequency F models, and then a series of custom designs under B, V, and others for specific customers. By keeping this new line as Threadripper Pro, it keeps it all under one umbrella.
Threadripper Pro Offerings: 12 core to 64 core
AMD announced these processors in the middle of last year, along with the Lenovo Thinkstation P620 as being the launch platform. From my experience, the Thinkstation line is very well designed, and we’re testing our 3995WX in a P620 today.
|AMD Ryzen Threadripper Pro|
|3995WX||64 / 128||2700||4200||8 + 1||256 MB||280 W||$5490|
|3975WX||32 / 64||3500||4200||4 + 1||128 MB||280 W||$2750|
|3955WX||16 / 32||3900||4300||2 + 1||64 MB||280 W||$1150|
|3945WX||12 / 24||4000||4300||2 + 1||64 MB||280 W||*|
|*Unsure if this is a special OEM model|
When TR Pro was announced with Lenovo, we weren’t sure if any other OEM would have access to Threadripper. When we asked OEMs earlier in that year about it, before we even knew if TR Pro was a real thing, they stated that AMD hadn’t even marked the platform on their roadmap, which we reported at the time. We have since learned that Lenovo had the 6-month exclusive, and information was only supplied to other vendors (ASUS, GIGABYTE, Supermicro) after it had been announced.
To that end, AMD has since announced that Threadripper Pro is coming to retail, both for other OEMs to design systems, or for end-users to build their own. Despite using the same LGA4094 socket as the other Threadripper and EPYC processors, TR Pro will be locked down to WRX80 motherboards. We currently know of three, such as the Supermicro and GIGABYTE models, plus we have also had the ASUS Pro WS WRX80E-SAGE SE Wi-Fi model in house for a short hands-on, although we weren’t able to test it.
Of the four processors listed above, the top three are going on sale. It’s worth noting that only the 64-core comes with 256 MB of L3 cache, while the 32-core comes with 128 MB of L3. AMD has kept that these chiplet designs only use as many chipsets as is absolutely necessary, keeping L3 cache per core consistent as well as the 8-cores per chiplet (the EPYC product line varies this a bit).
The fourth processor, the 12-core, would appear to be an OEM-only specific processor for prebuilt systems.
Threadripper Pro versus The World
These Threadripper Pro offerings are designed to compete against two segments: first is AMD themselves, showcasing anyone who is using a high-end professional system built on first generation Zen hardware that there is a lot of performance to be had. The second is against Intel workstation customers, either using single socket Xeon W (which tops out at 28 cores), or a dual socket Xeon system that costs more or uses a lot more power, just because it is dual socket, but also has a non-uniform memory architecture.
We have almost all these in this test (we don't have the 7702P, but we do have the 7742), and realistically these are the only processors that should be considered if the 3995WX is an option for you:
|3995WX Comparison Offerings|
|TR Pro 3995WX||64C||$5490||1P||280W||2700||4200||8x3200||128x 4.0||2 TB|
|TR 3990X||64C||$3990||1P||280W||2900||4300||4x3200||64x 4.0||¼ TB|
|EPYC 7702P||64C||$4425||1P||200W||2000||3350||8x3200||128x 4.0||4 TB|
|EPYC 7742||64C||$6950||2P||225W||2250||3400||8x3200||128x 4.0||4 TB|
|Xeon 6258R||28C||$3950||2P||205W||2700||4000||6x2933||48x 3.0||1 TB|
|Xeon W-3175X||28C||$2999||1P||255W||3100||4300||6x2933||48x 3.0||½ TB|
Intel tops out at 28 cores, and there is no getting around that. Technically Intel has the AP processor line that goes up to 56 cores, however these are for specialist systems and we haven’t had one physically sent to us for testing. Those are also $20k+ per CPU, and are two CPUs in the same system bolted under one package.
The AMD comparison points are the best Threadripper option and the best available EPYC Processor, albeit the 2P version. The best comparison here would be the 7702P, the single socket variant and much more price competitive, however we haven’t got this in for testing, instead we have AMD's EPYC 7742, which is the dual socket version but slightly higher performance.
|AMD TR Pro||TR Pro
|AMD TR||TR 3990X||MSI
|EPYC 7742||Supermicro H11DSI||BIOS
|Xeon Gold 6258R||ASUS ROG
|GPU||Sapphire RX 460 2GB (CPU Tests)|
|PSU||Various (inc. Corsair AX860i)|
|SSD||Crucial MX500 2TB|
|Silverstone SST-FHP141-VF 173 CFM fans also used. Nice and loud.|
We must thank the following companies for kindly providing hardware for our multiple test beds. Some of this hardware is not in this test bed specifically, but is used in other testing.
|Hardware Providers for CPU and Motherboard Reviews|
RX 460 Nitro
RTX 2080 Ti
|Crucial SSDs||Corsair PSUs|
Users interested in the details of our current CPU benchmark suite can refer to our #CPUOverload article which covers the topics of benchmark automation as well as what our suite runs and why. We also benchmark much more data than is shown in a typical review, all of which you can see in our benchmark database. We call it ‘Bench’, and there’s also a link on the top of the website in case you need it for processor comparison in the future.
Post Your CommentPlease log in or sign up to comment.
View All Comments
kwinz - Wednesday, February 10, 2021 - linkReally? CPUs are in high demand because GPU programming is hard? That's what you're going with?
Gomez Addams - Wednesday, February 10, 2021 - linkGood heavens that was painful to read. Some of the worst writing have had to suffer through in a while. One tip : compute is a verb and not a noun, just as simulation is a noun and simulate is a verb. Just because marketing droids use a term does not mean it is correct.
Spunjji - Thursday, February 11, 2021 - linkSorry pal, language doesn't work that way. You may not *like* it, but that's the way it is!
croc - Wednesday, February 10, 2021 - linkAre we now ignoring the elephant? EPYC was to launch in 2020. Actually, AMD said that Zen 3 would launch in 2020, but there's a weasel in them words... SOME of the Zen 3 cpu's DID launch, mostly looking like paper though. EPYC is sort-of launching as we speak, and Zen 3 Threadripper is a no-show.
I have said this elsewhere, and I will say it here. It would appear that AMD's lack of fab experience is showing, as they seem to be having issues getting their designs to fab properly at 7nm. Low to no yields? And TSMC is having issues of their own with China buying up as much talent as it can, while threatening to just grab it all in a military takeover. TSMC should have already built an advanced fab somewhere in the west, out of China's reach. Europe? Canada? After Trump, I would say to avoid the US as much as it needs to avoid China.
Spunjji - Thursday, February 11, 2021 - linkAMD were hoping to get Milan into production in Q3 2020 and have it shipping to some customers by Q4, which they did. It's not available to OEMs yet, so hardly a fanfare moment, but not the "elephant" you're trying to paint it as either.
Same goes for Zen 3, too - it absolutely wasn't a "paper launch" - but I see you're just here to push FUD rather than discuss *the article*.
Like, what's this "they seem to be having issues getting their designs to fab properly at 7nm" crap? Whose backside are you pulling that out of?
Amazing how many people seem to think these comment sections are the ideal place to grind their own personal political axes.
Oxford Guy - Thursday, February 11, 2021 - link‘Amazing how many people seem to think these comment sections are the ideal place to grind their own personal political axes’
You seem to think this is your personal website.
Qasar - Thursday, February 11, 2021 - linkas do you, point is ?
Oxford Guy - Wednesday, March 10, 2021 - linkIt’s not the tu quoque fallacy.
tygrus - Thursday, February 11, 2021 - linkIf total desktop sales were 20Million in 2020Q4 and if AMD sold 5M with about 0.95M being Zen3 so AMD could have been 19% of desktop CPU sales being Zen3. That's a good start and a lot better than <5% you may think happens for a paper launch. Notebook market adds another 50M/qtr (20% AMD?) and tablets (probably not AMD) on top of that so Zen3 sales would look like ~7% of AMD consumer CPU's sold that qtr.
Not all sales & deliveries are publicised so server sales may have happened already for Zen3. The FAB capacity & yield were more than enough because it was the substrate & final assembly which limited supply.
ipkh - Sunday, February 28, 2021 - linkReally, they have Global Foundries to thank for this. Global Foundries miseead the market and decided to drop highend node production. This left TSMC as the only highend node company left standing (that does 3rd party fab). Global Foundries and Samsung could have had a much better roadmap working together with IBMs researchers. But they didn't and now we see how much it is costing the entire industry. AMD may be forced to use Samsung Foundries if TSMC production gets tied up.