Once we enable INT8, the $2500 Titan RTX is no less than 3 times faster than a pair of $10k Xeons 8280s

anandtech.com/show/14466/intel-xeon-cascade-lake-vs-nvidia-turing

>Once we enable INT8, the $2500 Titan RTX is no less than 3 times faster than a pair of $10k Xeons 8280s.

OH NO NO NO NO NO NO NO NO NO NO NO

AHAHAHAHAHAHAHAHAHAHAHAHAHAHA

THIS KILLS THE INTEL

Attached: 111682inf.png (650x525, 20K)

Other urls found in this thread:

arxiv.org/pdf/1811.09886.pdf
anandtech.com/show/14305/intel-xeon-phi-knights-mill-now-eol
itpeernetwork.intel.com/unleashing-high-performance-computing/
twitter.com/NSFWRedditVideo

DELID DIS

Attached: 1544667387109.png (813x1402, 324K)

First it was single socket Epyc, now even graphics cards obliterate Xeons. Intel just can't get a break.

Attached: AMD-EPYC-Nobody-Ever-Got-Fired-for-Buying-Xeon.jpg (943x537, 54K)

>int 8
>prone to massive amounts of compacted round off errors
Pretty useless magic trick

YOU CAN'T USE GPU EVERYWHERE MOST OF THE TIME THE CPU NEEDS TO DO JOB

>Doing Inference on CPU

why would you do that?

I hate Intel and all, but that's just a stupid comparison. It's like testing how a game runs on a GPU vs CPU software renderer.

>GPU is better at a GPU workload than a CPU!
All of those intel CPU +Nvidia GPU supercomputer clusters are surely going to be thrown out upon this information being brought to light.

>game
>CPU software renderer
Today you don't even have the option to do that.
They just don't make games like they used to

Attached: 1525388794348.png (380x349, 77K)

yeah let's ignore it's almost twice as fast in fp32

>comparing GPU to CPU
>giving a shit about AI

>Titan RTX
Why aren't these poorfags using the 5k Quadro RTX 8000

Graphics cards have always been faster than CPUs are certain tasks wtf are you smoking user

>with 1 billion times more round off errors
quality > quantity

Boy, aren't you retarded.

brainlet here
is this the fabled avx512 intel bet the house on?

VNNI is good for RNNs that can't be parallelized all that much. Obviously a highly parallel CNN does better on a highly parallel GPU architecture. On client platforms (icelake) it makes sense for everything but training as most people don't want to hook up an eGPU to their ultrabook to run some basic inference app.

You can have latency-sensitive but not very throughput-sensitive inference workloads, which do better on CPUs. Facebook has a lot of these for example

arxiv.org/pdf/1811.09886.pdf

3.5 = 4

What happened to the Xeon phi?

anandtech.com/show/14305/intel-xeon-phi-knights-mill-now-eol

itpeernetwork.intel.com/unleashing-high-performance-computing/
>One step we’re taking is to replace one of the future Intel® Xeon Phi™ processors (code name Knights Hill) with a new platform and new microarchitecture specifically designed for exascale.

>now in EOL

Attached: smug-3d-anime-girl.webm (380x480, 1.41M)

Attached: JUST.png (822x240, 17K)

It is also a lot more economical. However, I'm not so sure it is actually more efficient for large models.