Thursday, December 21st 2023

Intel Should be Leading the AI Hardware Market: Pat Gelsinger on NVIDIA Getting "Extraordinarily Lucky"

Intel CEO Pat Gelsinger considers NVIDIA "extraordinarily lucky" to be leading the AI hardware industry. In a recent public discussion with the students of MIT's engineering school to discuss the state of the semiconductor industry, Gelsinger said that Intel should be the one to be leading AI, but instead NVIDIA got lucky. We respectfully disagree. What Gelsinger glosses over with this train of thought is how NVIDIA got here. What NVIDIA has in 2023 is the distinction of being one of the hottest tech stocks behind Apple, the highest market share in a crucial hardware resource driving the AI revolution, and of course the little things, like market leadership over the gaming GPU market. What it doesn't have, is access to the x86 processor IP.

NVIDIA has, for long, aspired to be a CPU company, right from its rumored attempt to merge with AMD in the early/mid 2000s, to its stint with smartphone application processors with Tegra, an assortment of Arm-based products along the way, and most recently, its spectacularly unsuccessful attempt to acquire Arm from Softbank. Despite limited luck with the CPU industry, to level up to Intel, AMD, or even Qualcomm and MediaTek; NVIDIA never lost sight of its goal to be a compute hardware superpower, which is why, in our opinion, it owns the AI hardware market. NVIDIA isn't lucky, it spent 16 years getting here.
NVIDIA's journey to AI hardware leadership begins back in the late 2000s, when it saw the potential for GPU to be a general purpose processor, since programmable shaders essentially made the GPU a many-core processor with a small amount of fixed-function raster hardware on the side. The vast majority of an NVIDIA GPU's die-area is made up of streaming multiprocessors—the GPU's programmable SIMD muscle.

NVIDIA's primordial attempts to break into the HPC market with its GPUs bore fruit with its "Tesla" GPU, and the compute unified device architecture, or CUDA. NVIDIA's unique software stack that lets developers build and accelerate applications on its hardware dates all the way back to 2007. CUDA set in motion a long and exhaustive journey leading up to NVIDIA's first bets with accelerated AI on its GPUs a decade later, beginning with "Volta." NVIDIA realized that despite a vast amount of CUDA cores on its GPUs and HPC processors, it needed some fixed-function hardware to speed up deep learning neural network building, training, and inference, and developed the Tensor core.

In all this time, Intel continued to behave like a CPU company and not a compute company—the majority of its revenue came from client CPUs, followed by server CPUs, and it has consistently held accelerators at a lower priority. Even as Tesla and CUDA took off in 2007, Intel had its first blueprints for an SIMD accelerator, codenamed "Larrabee" as early as by 2008. The company hasn't accorded the focus Larrabee needed as a nascent hardware technology. But that's on Intel. AMD has been a CPU + GPU company since its acquisition of ATI in 2006, and has tried to played catch-up with NVIDIA by combining its Stream compute architecture with open compute software technologies. The reason AMD's Instinct CDNA processors aren't as successful as NVIDIA's A100 and H100 processors is the same reason Intel never stood a chance in this market with its "Ponte Vecchio"—it was slow to market, and didn't nurture an ecosystem around its silicon quite like NVIDIA did.

Hardware is a fraction of NVIDIA's growth story—the company has an enormous, top-down software stack, including its own programming language, APIs, prebuilt compute and AI models; and a thriving ecosystem of independent developers and ISVs that it has nurtured over these years. So by the time AI took off at scale as a revolution in computing, NVIDIA was ready with the fastest hardware, and the largest community of developers that could put it to use. We began this editorial by stating that it's a good thing NVIDIA didn't acquire an x86 license in the early 2000s. It could switch gears and look inward on the one thing it was already making that can crunch numbers at scale—GPUs with programmable shaders. What NVIDIA is extraordinarily lucky about is that it didn't get stuck with an x86 license.

You can watch Pat Gelsinger's interview over at MIT's YouTube channel, here.
Source: ExtremeTech
Add your own comment

55 Comments on Intel Should be Leading the AI Hardware Market: Pat Gelsinger on NVIDIA Getting "Extraordinarily Lucky"

#1
AleksandarK
News Editor
Software ate the world, and NVIDIA was there to accelerate it. Valid points in the editorial, and a great read!
Posted on Reply
#2
dj-electric
I really hope Pat does not actually believe the words he said about NVIDIA being lucky.
NVIDIA very much could be how the use of hardware accelerated machine learning got to where it is today. I remember attending a GTC in 2011 where it was already in the air.
Posted on Reply
#3
nguyen
Well Nvidia have one single leader whose sole focus is to innovate, whereas Intel have had CEOs who are more bean counters than anything else, that's why Intel cancelled projects at first sight of difficulty...
Posted on Reply
#4
ExcuseMeWtf
According to him, Intel should be leading every single market lol.
Posted on Reply
#5
Eternit
dj-electricI really hope Pat does not actually believe the words he said about NVIDIA being lucky.
NVIDIA very much could be how the use of hardware accelerated machine learning got to where it is today. I remember attending a GTC in 2011 where it was already in the air.
Basically since 22nm Intel had problems with its fabs. While 22nm was on time, it was initially no better than 32nm. Only two years latter with Haswel refresh it was proper. Even worse with 14nm as it took them 4 years to make it proper, and 10nm it was a disaster. Because of this they were unable to produce Larrabee and Atoms for phones. Later they were unable to produce next gen Phi, and now they must use TSMC fabs.
While AMD were struggling to survive and nVidia was investing in innovations, Intel was paying big dividends.
Posted on Reply
#7
john_
Well the article puts things in their right place and many of us where saying the same things for years. I really don't understand who Pat is trying to fool, especially at MIT. Maybe next time he should try Microcenter and talking to customers who need help to send an email. They might believe him.

Just adding a couple stuff here.
Nvidia tried to get an X86 license. Intel said no.
Huang said once(I think, read it once many many years ago), I believe around 2005 or even earlier "Nvidia is a software company that also happens to build the hardware where that software will run"

And one more thought.
Creative was the King in Audio 20+ years ago. With on board audio solutions becoming the normal, Creative just became an old name to remember.
I think Huang seen that and making the GPU a powerful co processor wasn't only a great vision, but also a necessary transformation of the simple graphics chip that could also be replaced in the future from cheap on board solutions that could be good enough for the majority of consumers.
Posted on Reply
#8
kondamin
Nvidia wasn't big enough to be able to ignore competition and had to step up a couple of times where intel didn't have to.
Posted on Reply
#9
dj-electric
EternitBasically since 22nm Intel had problems with its fabs. While 22nm was on time, it was initially no better than 32nm. Only two years latter with Haswel refresh it was proper. Even worse with 14nm as it took them 4 years to make it proper, and 10nm it was a disaster. Because of this they were unable to produce Larrabee and Atoms for phones. Later they were unable to produce next gen Phi, and now they must use TSMC fabs.
While AMD were struggling to survive and nVidia was investing in innovations, Intel was paying big dividends.
Fabrication isn't why hardware accelerated machine learning turned NVIDIA into a successful company.
Development, tools and community is why. NVIDIA actively running workshops, actively running courses and actively granting the tools to buy (that means even your local computer store's gaming GPU), learn and perform ML actions. No server grade or expensive hardware needed
Posted on Reply
#10
Assimilator
Intel's inability to be honest about their own failings is the company's biggest failing, and will destroy it if not rectified.
Posted on Reply
#11
AnotherReader
AssimilatorIntel's inability to be honest about their own failings is the company's biggest failing, and will destroy it if not rectified.
I hope they are more honest in internal discussions. This was, after all, a public venue.
Posted on Reply
#12
dyonoctis
People initially thought that getting an engineer as CEO would be a good thing for Intel, but as time passed, I was under the impression that they hired the biggest delusional fanboy. Pat is making a lot of unnecessary wild claims.
Posted on Reply
#13
Rares
I think PG is extraordinarily stupid! Everything he says is utter nonsense !
Posted on Reply
#14
mechtech
maybe intel shouldn't have thrown in the towel on larrabee then??
Posted on Reply
#15
thesmokingman
dyonoctisPeople initially thought that getting an engineer as CEO would be a good thing for Intel, but as time passed, I was under the impression that they hired the biggest delusional fanboy. Pat is making a lot of unnecessary wild claims.
Well he does come from the Netburst era when they were colluding behind closed doors. BS is not out of his repertoire.
Posted on Reply
#16
AnarchoPrimitiv
ALL success, and I mean ALL, has some degree of luck involved and lucky timing, to say otherwise is to imply that you can control every aspect of reality...EVERY success story involves luck.

*We should ALL be praying somebody knocks Nvidia off their high horse. Monopoly and AI alone are issues that represent many dangers, but together....that could be truly dangerous.
Posted on Reply
#17
Fouquin
mechtechmaybe intel shouldn't have thrown in the towel on larrabee then??
They didn't. Larrabee became Xeon Phi. By all accounts Xeon Phi was a success. It ran multiple generations, saw exponential improvements in performance and design, introduced the super-wide SIMD (AVX-512) to consumer markets, and pushed development forward on GPGPU at Intel where none was happening prior. Intel's Xe graphics chips owe their heritage in part to the Larrabee team with the dual-issue Vector/Matrix compute blocks.
thesmokingmanWell he does come from the Netburst era
He comes from the "Old" Intel. He exited after the Netburst years but Pat Gelsinger's signature is literally printed on the i386. Twice, actually. He also co-wrote the book on modern x86 as we know it. He's a well accomplished engineer, but also an incredibly brash and out of touch CEO.
btarunrIn all this time, Intel continued to behave like a CPU company and not a compute company—the majority of its revenue came from client CPUs, followed by server CPUs, and it has consistently held accelerators at a lower priority.
I would argue against this if only for the recent years. Intel purchased Habana labs in 2019, dedicating millions to AI accelerator development (Gaudi). Then in 2021 they bought up the entirety of CenTaur's x86 design team who were working on AI vision technology for VIA, as well as designing the first x86 CPU with an integrated ultra-wide VLIW neural processing unit.

Going back to the formation of the "Gen" graphics division (separate from Larrabee, those were two distinct teams) Intel invested heavily in the development of fixed-function accelerators for video with Quick Sync. They also began work on AMX way back in 2017 but made the poor decision to only incorporate that hardware in Sapphire Rapids which saw an historic 30+ months of delays before finally crawling over the finish line. So it's not that they sat back and did nothing, they simply weren't strategic enough in the use of their technologies.
Posted on Reply
#18
Mawkzin
dyonoctisPeople initially thought that getting an engineer as CEO would be a good thing for Intel, but as time passed, I was under the impression that they hired the biggest delusional fanboy. Pat is making a lot of unnecessary wild claims.
It's funny because Lisa is also a engineer and they're almost opposity as CEOs,.
Posted on Reply
#19
SOAREVERSOR
MawkzinIt's funny because Lisa is also a engineer and they're almost opposity as CEOs,.
Smaller companies are more scrappy, market leaders coast on name.
Posted on Reply
#20
stickleback123
AnotherReaderI hope they are more honest in internal discussions. This was, after all, a public venue.
I hope for their sake that they are, although from what I gather their internal culture may not allow that.

In any case it makes him look like a fool, if a competitor has done well through decades of careful planning and competent execution you acknowledge that, to do otherwise makes you look graceless, petty, and small.
Posted on Reply
#21
remixedcat
Pat Gelsigner digging for apples

reminds him of what he also lost.... apple... and now they are outperforming them with thier own silicon that does better with creative tasks like music production.
Posted on Reply
#22
john_
dj-electricFabrication isn't why hardware accelerated machine learning turned NVIDIA into a successful company.
But it was the reason for Intel's success in the past. The reason Intel could fake an X86 SOC being as efficient as an ARM SOC. When Intel lost that fabrication advantage, it stopped dreaming of winning market share on tablets against ARM. When Intel fell behind, AMD that was the worst company on the planet on efficiency, but was using TSMC's better manufacturing nodes, beat them.
dyonoctisPeople initially thought that getting an engineer as CEO would be a good thing for Intel, but as time passed, I was under the impression that they hired the biggest delusional fanboy. Pat is making a lot of unnecessary wild claims.
Pat is right about one thing. Trying to make Intel the top manufacturer in the world. Trying to beat TSMC and Samsung. He knows that having a manufacturing advantage is as big of an advantage as having the best architecture.
Posted on Reply
#23
Franzen4Real
Apparently, Salty Pat feels that his marketing team didn't make them look like complete fools while firing shots at AMD a couple of weeks ago, so he has taken it upon himself to finish off the job.
Clearly upset that he's riding on leather coat tails in to the AI market, while simultaneously being shown how to engineer a great CPU with less resources, by a team 1/3 their size. It is Intel that is 'lucky' to have this new market created by others, where they can now dump their -me too- "AI" products.
Posted on Reply
#24
remixedcat
john_But it was the reason for Intel's success in the past. The reason Intel could fake an X86 SOC being as efficient as an ARM SOC. When Intel lost that fabrication advantage, it stopped dreaming of winning market share on tablets against ARM. When Intel fell behind, AMD that was the worst company on the planet on efficiency, but was using TSMC's better manufacturing nodes, beat them.

Pat is right about one thing. Trying to make Intel the top manufacturer in the world. Trying to beat TSMC and Samsung. He knows that having a manufacturing advantage is as big of an advantage as having the best architecture.
bay trail and cherry trail tablets were such a pain... they were designed for windows but windows ran like a dog on tranq... I had to put linux on 3 of mine... and worst part is all 3 are technically 64 bit but they stuck a 32 bit bootloader/efi on them so it was hard af to put linux on them. They did that to lock windows in when when I got linux working it was massively better.

Intel failed so hard on tablets. So much waste
Posted on Reply
#25
Ferrum Master
Each article I see him in, I cannot decide either he is delusional, mad, a clown or just plain blind.
Posted on Reply
Add your own comment
Dec 21st, 2024 02:56 EST change timezone

New Forum Posts

Popular Reviews

Controversial News Posts