Intel’s new i9 – 10980 XE, debuting on the same day as AMD’s new Threadripper line, occupies a strange market segment: the “budget high-end desktop . ” Its (cores and 36 threads sound pretty exciting compared to Intel’s top-end gaming CPU, the i9 – 9900 KS —But they pale in comparison toThreadripper (x) ‘s (cores and 64 threads. Making things worse, despite having more than double the cores, i9 – 10980 XE has trouble differentiating itself even from the much less expensive i9 – 9900 KS in many benchmarks.
This leaves the new part falling back on what itdoeshave going for it — cost, both initial and operational. If you can’t use the full performance output of a Threadripper, the i9 – 10980 XE will give you roughly half the performance for roughly half of the cost, and it extends that savings into ongoing electrical costs as well.
Our i9 – 10980 XE test rig was a lot easier to share an office with than the competing Threadripper 3970 x rig. Its EVGA X 399 Dark motherboard didn’t make it look like a scene fromPoltergeistwas playing out in the office, and it drew a lot less power and threw off a lot less palpable heat.
To be completely fair, some of the Threadripper rig’s obnoxiousness probably could have been mitigated with motherboard settings — our NZXT Kraken x 62 cooler’s fans were in full-on leafblower mode the entire time the Threadripper was running, even while idling. While the Threadripper systemdididle at an eye-watering 163 W to the i9 – 10980 XE system’s 69 W, that’s not enough to explain the difference in fan RPMs. The i9 – 10980 XE system didn ‘t spin the fans up to obnoxious levels even under its full 257 W benchmarking load.
No matter how many excuses we make for the Threadripper based on its ROG motherboard’s aggressive default fan settings, though, we ‘ re still looking at W vs 163 W system draw at desktop idle and (W vs) W system draw under full load. In the current HEDT segment, the i9 – 10980 XE is definitely the more frugal part to live with — particularly for those of us who live in Southern climes where July starts sometime in April and doesn’t end until late October.
|Specs at a glance: Core i9 – 10980 XE, as tested|
|CPU||3.0GHz 18 – core Intel Core i9 – (XE (4.7GHz boost) with 24. 75 MB smart cache — expected retail $ 1 , (0)|
|RAM||16 GB HyperX Fury DDR4 3200 – $ 88 atAmazon|
|MSI Geforce RTX 2060 Super Ventus – $ 420 atAmazon|
|Samsung 860 Pro 1TB SSD – $ (atAmazon|
|Motherboard||EVGA X 299 Dark – $ 415 at (Amazon)|
|Cooling||NZXT Kraken X 62 fluid cooler with 280 mm radiator – $ 140 atAmazon|
|EVGA 850 GQ Semi Modular PSU – $ 130 atAmazon|
|Praxis Wetbench test chassis – $ 200 atAmazon|
|Price as tested||≈ $ 2, 668|
If you haven’t spent the entirety of 2019 under a rock, it shouldn’t be any surprise that Intel’s i9 – 10980 XE comes up short when compared to AMD’s Threadripper 3970 x, which also released this Monday. The i9 – 10980 XE is an 18 – core, 36 – thread part to the Threadripper’s 32 cores and 64 threads, and AMD’s 7nm process has ended any offsetting performance-per-core advantages Intel once had.
What’s more surprising is that the i9 – 10980 XE fell generally short of its elder brother, last year i9 – 9980 XE. Its best general-purpose benchmarking result — Cinebench R 20 – placed it within the margin of error of the older part. Worse, both single-threaded and multi-threaded Passmark ratings strongly favored the older chip.
We also compared the i9 – 10980 XE to its gaming sibling, the i9 – 9900 KS. Although the i9 – 10980 XE scored significantly higher in Cinebench R 20, it didn’t score significantly better in the more general-purpose Passmark test, despite having more than twice the cores — and costing twice as much. Meanwhile, in single-threaded Passmark, the 9900 KS was 39% faster — and once again, the i9 – 10980 XE fell behind its older sibling, the i9 – 9980 XE.
Then again — following a pattern we’re all learning to recognize — the i9 – 10980 XE is half the cost of the i9 – (XE, which sits in $ 2, 00 0 territory alongside the much more powerful Threadripper 3970 x. We feel sorry for any retailers sitting on existing purchased i9 – 9980 XE stock, because while the older part is slightly more performant for general-purpose workloads, the difference certainly isn’t worth double the cost.
AI inference workloads
Artificial Intelligence workloads are the one place the i9 – 10980 XE shines in its own right. Intel has been investing a lot of engineering effort into AI workload optimization, and the i9 – 10980 XE features its Deep Learning Boost x 86 extension instructions. With compilers that can take advantage of the new instruction set, Intel has beentellingus that AI workloads can easily double their throughput.
We ranAIXPRT‘s eference benchmarking workloads to test this claim. Sure enough, the i9 – 10980 more than doubled the otherwise more-performant i9 – 9980 ‘s throughput in OpenVINO image recognition work. It’s worth noting that OpenVINO itself is an Intel initiative, and it’s unlikely to be as optimized for AMD processors as Intel. This lack of optimization can explain some of the performance delta between the Threadripper 3970 x and the i9 – 10980 XE, but it’s difficult to come up with anythingbutDeep Learning Boost as the rationale for blowing the doors off Intel’s own i9 – 9980 XE.
And yes, inference workload performance on CPUs really does matter. To explain why, let’s first make sure we’re clear on some basic terminology. A neural network can be operated in one of two modes — training or inference. In training mode, the neural network is essentially doing a drunkard’s walk through a problem space, adjusting values and weights until it has “learned” the best way to navigate that problem space. Inference mode is much lighter weight; instead of having to traverse the entire space repeatedly, the neural network just examines one problem within the space and gives you its best answer, based on what it has learned during training.
When doing significant amounts of training, you want a high-end GPU (or a bank of high-end GPUs) , full stop. They can easily outperform general-purpose CPUs by an order of magnitude or more, saving both training time and operational power. Inference, however, is a different story — although the GPU may still significantly outperform a CPU, if the CPU’s operational throughput and latency suffice for real-time interaction, it’s much more convenient to let it do the work.
Running inference workloads on general-purpose CPUs makes it possible to widely deploy them as tools that can be used in the field , on relatively generic hardware, and without the need for an always-on internet connection. This can also alleviate both privacy and latency concerns involved in sending data off to the cloud for remote inference processing.
Some examples of real-world, modern AI inference processing include voice recognition, image recognition, and pattern analysis. While much of this is currently being done in the cloud, we expect increasing demand for local processing capabilities. Digital personal assistants such as Cortana are one obvious application, but AI can go even further than that.
Right now, the Insider edition of Office 365 allows you to type “what was the highest selling product in Q4?” into a text box and immediately have the appropriate chart produced for you … but that’s only if you’re willing to offload your data to the cloud. Without local inference processing chops, working with confidential data that can’t leave the site where it’s produced will feel more and more restrictive. Intel’s betting a lot of its future on recognizing that fact — and being at the head of the pack when the rest of us do, too.
Listing image by Jim Salter