Dec 19th, 2024 04:50 EST change timezone

New Forum Posts

Popular Reviews

Controversial News Posts

Thursday, November 23rd 2023

NVIDIA Experiences Strong Cloud AI Demand but Faces Challenges in China, with High-End AI Server Shipments Expected to Be Below 4% in 2024

NVIDIA's most recent FY3Q24 financial reports reveal record-high revenue coming from its data center segment, driven by escalating demand for AI servers from major North American CSPs. However, TrendForce points out that recent US government sanctions targeting China have impacted NVIDIA's business in the region. Despite strong shipments of NVIDIA's high-end GPUs—and the rapid introduction of compliant products such as the H20, L20, and L2—Chinese cloud operators are still in the testing phase, making substantial revenue contributions to NVIDIA unlikely in Q4. Gradual shipments increases are expected from the first quarter of 2024.

The US ban continues to influence China's foundry market as Chinese CSPs' high-end AI server shipments potentially drop below 4% next year
TrendForce reports that North American CSPs like Microsoft, Google, and AWS will remain key drivers of high-end AI servers (including those with NVIDIA, AMD, or other high-end ASIC chips) from 2023 to 2024. Their estimated shipments are expected to be 24%, 18.6%, and 16.3%, respectively, for 2024. Chinese CSPs such as ByteDance, Baidu, Alibaba, and Tencent (BBAT) are projected to have a combined shipment share of approximately 6.3% in 2023. However, this could decrease to less than 4% in 2024, considering the current and potential future impacts of the ban.
China to expand investment in proprietary ASICs and develop general-purpose AI chips due to limited high-end AI chip demand
Facing the risk of expanded restrictions arising from the US ban, TrendForce believes Chinese companies will continue to buy existing AI chips in the short term. NVIDIA's GPU AI accelerator chips remain a top priority—including existing A800 or H800 inventories and new models like H20, L20, and L2—designed specifically for the Chinese market following the ban. In the long term, Chinese CSPs are expected to accelerate, with Alibaba's T-Head and Baidu being particularly active in this area, relying on foundries like TSMC and Samsung for production.

At the same time, major Chinese AI firms, such as Huawei and Biren, will continue to develop general-purpose AI chips to provide AI solutions for local businesses. Beyond developing AI chips, these companies aim to establish a domestic AI server ecosystem in China. TrendForce recognizes that a key factor in achieving success will come from the support of the Chinese government through localized projects, such as those involving Chinese telecom operators, which encourage the adoption of domestic AI chips.

Edge AI servers: A potential opportunity for Chinese firms amid high-end AI chip development constraints
A notable challenge in developing high-end chips in China is the limited access to advanced manufacturing technology. This is particularly true for Huawei, which remains on the US Entity List and relies on domestic foundries like SMIC for production. Despite SMIC's advancements, it faces similar issues created by the US ban—including difficulties in obtaining key advanced manufacturing equipment and potential yield issues. TrendForce believes that in trying to overcome these limitations, China may find opportunities in the mid to low-range edge AI server market. These servers, with lower AI computational demands, cater to applications like commercial ChatBOTs, video streaming, internet platforms, and automotive assistance systems. They might not be fully covered by US restrictions, presenting a possible growth direction for Chinese firms in the AI market.
Source: TrendForce
Add your own comment

20 Comments on NVIDIA Experiences Strong Cloud AI Demand but Faces Challenges in China, with High-End AI Server Shipments Expected to Be Below 4% in 2024

#1
A&P211
To the moon on their stock price.
Really their stock price is like rocket this year. I'm happy I got in early enough but I wont be adding to my portfolio and I'm not selling.

I wonder with a rumor going around that OpenAI made a breakthrough on artificial general intelligence might mean to Nvidia.
Posted on Reply
#2
thesmokingman
A&P211I wonder with a rumor going around that OpenAI made a breakthrough on artificial general intelligence might mean to Nvidia.
All the major players ie. Google, MSFT/OpenAI, Amazon, and Tesla are moving or have moved to custom silicon.
Posted on Reply
#3
Denver
thesmokingmanAll the major players ie. Google, MSFT/OpenAI, Amazon, and Tesla are moving or have moved to custom silicon.
Exactly, running AI on ASICs will significantly enhance efficiency, prompting companies to reduce service operating costs. What's noteworthy is that, unlike the CPU and GPU market, there is no nearly insurmountable IP barrier, allowing major players to freely develop their solutions.
Posted on Reply
#5
Assimilator
I love how certain people in all the NVIDIA/China threads are consistently making stupid claims that the Chinese ban is going to somehow destroy the company, and that NVIDIA is breaking the law to ship GPUs to circumvent this ban. Yet the actual numbers show that China makes up barely a 15th of all shipments of NVIDIA's ML-focused chips, and this will decrease by an incredible two percent in 2024.

Clearly NVIDIA will have to massively step up their imaginary lawbreaking in order to make up for this incredible shortfall! Also, and let me be completely clear here: you conspiracy theorists are complete and utter idiots, which is why you're broke and shitposting on tech forums, and NVIDIA shareholders aren't.
thesmokingmanAll the major players ie. Google, MSFT/OpenAI, Amazon, and Tesla are moving or have moved to custom silicon.
LMAO, and where do you think they going to fab that custom silicon? Where everyone else does, i.e. TSMC.

You know, the same TSMC that is massively backlogged with orders?
The company that NVIDIA has a contract with for a large percentage of wafer allocation, and all those other companies don't?
The company that isn't going to give time of day to anyone who isn't willing to commit to massive volumes, which the aforementioned players can't?

Or maybe they're going to use Intel... oh wait, they're outsourcing to TSMC too now.

Guess it's GloFo then, I'm sure that'll work just great!
Posted on Reply
#6
Denver
AssimilatorI love how certain people in all the NVIDIA/China threads are consistently making stupid claims that the Chinese ban is going to somehow destroy the company, and that NVIDIA is breaking the law to ship GPUs to circumvent this ban. Yet the actual numbers show that China makes up barely a 15th of all shipments of NVIDIA's ML-focused chips, and this will decrease by an incredible two percent in 2024.

Clearly NVIDIA will have to massively step up their imaginary lawbreaking in order to make up for this incredible shortfall! Also, and let me be completely clear here: you conspiracy theorists are complete and utter idiots, which is why you're broke and shitposting on tech forums, and NVIDIA shareholders aren't.


LMAO, and where do you think they going to fab that custom silicon? Where everyone else does, i.e. TSMC.

You know, the same TSMC that is massively backlogged with orders?
The company that NVIDIA has a contract with for a large percentage of wafer allocation, and all those other companies don't?
The company that isn't going to give time of day to anyone who isn't willing to commit to massive volumes, which the aforementioned players can't?

Or maybe they're going to use Intel... oh wait, they're outsourcing to TSMC too now.

Guess it's GloFo then, I'm sure that'll work just great
You get lost in your own arrogance by assuming a lot of things that are virtually impossible to know, whether it's about people's intellect or what Nvidia is doing behind the scenes. Sorry but that's a stupid attitude that only a blind fanatic would have.

ASICs are so many times more efficient that you can manufacture them in a more mature and cheaper process whose capacity is not compromised. TSMC's capacity is also increasing not decreasing.
Posted on Reply
#7
TheoneandonlyMrK
videocardz.com/newz/ai-craze-sweeps-china-banned-nvidia-rtx-4090-gpus-dismantled-en-masse-by-chinese-companies-for-ai-card-production

Or just use GPU? Bad times I'm guessing for buyers.

There's certainly enough evidence on show to display that the AI and ML market in china does matter to someone.

Let's hope the raids on Nvidia offices recently found a non loop hole jumping company behaving as it should then.

Because there's been a few coincidences lately.

like Nvidia bios are cracked.

Oh we are eol 4090/80 soon. Ready for super's.

Oh look millions of 4090 are now getting taken apart in china.
Posted on Reply
#8
Assimilator
DenverASICs are so many times more efficient that you can manufacture them in a more mature and cheaper process whose capacity is not compromised.
Fair, but fabbing is only half the equation.

You have to design your custom ASICs first.
You have to design the software that they're going to use.
You have to write that software.
You have to migrate over all your current software, or write a translation layer that does so.

All of the above take time.
All of the above involve significant risk.

Tesla is the only company that has tried this and that's because they're run by a man rich enough to not give a s**t. Other companies would much rather stick with what they know and what works, and just buy more hardware. It's not the cost that's the problem, it's the supply, which leads to...
DenverTSMC's capacity is also increasing not decreasing.
And they're going to offer that capacity to the companies that already have massive long-term contracts with them.

Companies like, I dunno, NVIDIA.
TheoneandonlyMrKBecause there's been a few coincidences lately.

like Nvidia bios are cracked.

Oh we are eol 4090/80 soon. Ready for super's.

Oh look millions of 4090 are now getting taken apart in china.
You need help.
Posted on Reply
#9
Denver
AssimilatorFair, but fabbing is only half the equation.

You have to design your custom ASICs first.
You have to design the software that they're going to use.
You have to write that software.
You have to migrate over all your current software, or write a translation layer that does so.

All of the above take time.
All of the above involve significant risk.

Tesla is the only company that has tried this and that's because they're run by a man rich enough to not give a s**t. Other companies would much rather stick with what they know and what works, and just buy more hardware. It's not the cost that's the problem, it's the supply, which leads to...


And they're going to offer that capacity to the companies that already have massive long-term contracts with them.

Companies like, I dunno, NVIDIA.


You need help.
All large corporations (Meta, Google, Microsoft, Amazon) developing AI services are mostly or partly software companies. Is it so difficult to imagine that they who not only have the expertise to create the software ecosystem for their hardware but also the necessary money have been working on it in silence for some time?

The state of the AI market can be illustrated by considering Tesla as an example. A while back, the company invested in a market that few believed in until there was an explosion in demand. Tesla, being ahead of the curve, was able to set its fees freely. However, with increasing competition, the battle over pricing levels is exerting pressure on profit margins, compelling companies to invest in maximizing efficiency. Simply swap Tesla for NVIDIA in this historical context, and you will observe the pattern repeating. Even Huang admitted the existence of this risk:

"Nvidia CEO Jensen Huang says his AI powerhouse is ‘always in peril’ despite a $1.1 trillion market cap: ‘We don’t have to pretend…we feel it’ "

Nvidia CEO says his AI powerhouse is ‘always in peril’ | Fortune
Posted on Reply
#10
R0H1T
AssimilatorTesla is the only company that has tried this and that's because they're run by a man rich enough to not give a s**t. Other companies would much rather stick with what they know and what works, and just buy more hardware. It's not the cost that's the problem, it's the supply, which leads to...
Nope Tesla's not the only one ~ Amazon, Google & now MS are moving to custom ARM chips. Granted this is about AI but their experience in developing(R&D) & then making these chips over several generations will help them everywhere even if Nvidia is the market leader right now. Also did you forget about Google's own TPU now at v4 o_O
Posted on Reply
#11
thesmokingman
R0H1TNope Tesla's not the only one ~ Amazon, Google & now MS are moving to custom ARM chips. Granted this is about AI but their experience in developing(R&D) & then making these chips over several generations will help them everywhere even if Nvidia is the market leader right now. Also did you forget about Google's own TPU now at v4 o_O
Yea, Googly has been on their iterations for a decade lmao. That poster is goin on my ignore list, such arrogance is hilarious yet insulting.
Posted on Reply
#12
Assimilator
thesmokingmanYea, Googly has been on their iterations for a decade lmao. That poster is goin on my ignore list, such arrogance is hilarious yet insulting.
Remind me how well Google's AI push is going.
Posted on Reply
#13
R-T-B
DenverASICs are so many times more efficient
At tasks that are fixed and easily defined. AI is very much outside that category. I do not see any way an ASIC designed for AI would be much different than what a GPU ASIC honestly already is: tons of simple stream processors.
Posted on Reply
#14
Minus Infinity
R-T-BAt tasks that are fixed and easily defined. AI is very much outside that category. I do not see any way an ASIC designed for AI would be much different than what a GPU ASIC honestly already is: tons of simple stream processors.
One com-any has already shown it's custom AI solution using far less energy to do the same calculations as H100 and do it 7x faster or something like that. The energy requirements for AI are growing exponentially, it'll make bitcoin energy use look pathetic. Companies are aware of this. They are not just trying to cut the costs of hardware but overall operating costs. These custom ASIC's are coming about precisely because of Huang's extreme greed and the egregious pricing of their H100/200 and GH200. With little compettion these scumbags charge what they like. Well no surprise several companies have had enough.
Posted on Reply
#15
Denver
R-T-BAt tasks that are fixed and easily defined. AI is very much outside that category. I do not see any way an ASIC designed for AI would be much different than what a GPU ASIC honestly already is: tons of simple stream processors.
Except AI isn't AI at all. There are several different applications within the spectrum of so-called AI, some companies including AMD/Xilinx have already demonstrated solutions that are more efficient than GPUs in some cases.

Nvidia sells GPUs at a very high price, which also encourages companies to look for alternatives.




Posted on Reply
#16
R-T-B
DenverExcept AI isn't AI at all.
I know exactly what "AI" (as it's being marketed) is when I spoke. It's weighted tables to make decisions. Same outcome.
Minus InfinityOne com-any has already shown it's custom AI solution using far less energy to do the same calculations as H100 and do it 7x faster or something like that.
I'd really like to see that independently validated. You'll gain some from losing the display circuitry sure, but not that much. I'm calling bullshit until someone (other than the company itself) proves that stat.
Denversome companies including AMD/Xilinx have already demonstrated solutions that are more efficient than GPUs in some cases.
More efficient I could buy, but not 7x. Like I said, losing the display circuitry and useless parts like ROPs alone is sure to gain you something, at least.
2x is feasible. Mind you nvidia could easily do the same.
Posted on Reply
#17
Assimilator
R-T-BAt tasks that are fixed and easily defined. AI is very much outside that category. I do not see any way an ASIC designed for AI would be much different than what a GPU ASIC honestly already is: tons of simple stream processors.
Oh thank god, finally someone who has a basic understanding of how this actually works.
DenverExcept AI isn't AI at all. There are several different applications within the spectrum of so-called AI, some companies including AMD/Xilinx have already demonstrated solutions that are more efficient than GPUs in some cases.

Nvidia sells GPUs at a very high price, which also encourages companies to look for alternatives.




Quoting performance numbers from three years ago isn't the slam-dunk you seem to think it is.
Posted on Reply
#18
Denver
AssimilatorOh thank god, finally someone who has a basic understanding of how this actually works.


Quoting performance numbers from three years ago isn't the slam-dunk you seem to think it is.
I just mentioned an example; you know, all these companies are working on their hardware internally. As I said, AI has a huge scope, and ASICs will be molded to fit each of these scenarios.

There's a big incentive to cut operating costs like this when Nvidia sells GPUs costing tens of thousands of dollars each.

"In 2013, Google realized that unless they could create a chip that could handle machine learning inference, they would have to double the number of data centers they possessed. Google claims that the resulting TPU has “15–30X higher performance and 30–80X higher performance-per-watt” than current CPUs and GPUs."

"Although both TPUs and GPUs can do tensor operations, TPUs are better at big tensor operations, which are more common in neural network training than 3D graphics rendering. The TPU core of Google is made up of two parts a Matrix Multiply Unit and a Vector Processing Unit. When it comes to the software layer, an optimizer is used to switch between bfloat16 and bfloat32 operations (where 16 and 32 are the number of bits) so that developers don’t have to rewrite their code. As a result, the TPU systolic array architecture has a large density and power advantage, as well as a non-negligible speed advantage over a GPU."

copperpod.medium.com/tensor-processing-unit-tpu-an-ai-powered-asic-for-cloud-computing-81ffb1256dd2
Posted on Reply
#19
TheoneandonlyMrK
AssimilatorOh thank god, finally someone who has a basic understanding of how this actually works.


Quoting performance numbers from three years ago isn't the slam-dunk you seem to think it is.
It was for Elon when he ditched Nvidia to make his own Dojo.
Posted on Reply
#20
R-T-B
TheoneandonlyMrKIt was for Elon when he ditched Nvidia to make his own Dojo.
Elon is not really the slamdunk genius many think either. I'm sure he did it mind, but that does not mean it was a good idea or "worth it."

I stand by what I said: independent verification or I'm not buying anything more than double the efficiency.
Posted on Reply
Add your own comment
Dec 19th, 2024 04:50 EST change timezone

New Forum Posts

Popular Reviews

Controversial News Posts