• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.

Social Media Imagines AMD "Navi 48" RDNA 4 to be a Dual-Chiplet GPU

btarunr

Editor & Senior Moderator
Staff member
Joined
Oct 9, 2007
Messages
47,230 (7.55/day)
Location
Hyderabad, India
System Name RBMK-1000
Processor AMD Ryzen 7 5700G
Motherboard ASUS ROG Strix B450-E Gaming
Cooling DeepCool Gammax L240 V2
Memory 2x 8GB G.Skill Sniper X
Video Card(s) Palit GeForce RTX 2080 SUPER GameRock
Storage Western Digital Black NVMe 512GB
Display(s) BenQ 1440p 60 Hz 27-inch
Case Corsair Carbide 100R
Audio Device(s) ASUS SupremeFX S1220A
Power Supply Cooler Master MWE Gold 650W
Mouse ASUS ROG Strix Impact
Keyboard Gamdias Hermes E2
Software Windows 11 Pro
A Chinese tech forum ChipHell user who goes by zcjzcj11111 sprung up a fascinating take on what the next-generation AMD "Navi 48" GPU could be, and put their imagination on a render. Apparently, the "Navi 48," which powers AMD's series-topping performance-segment graphics card, is a dual chiplet-based design, similar to the company's latest Instinct MI300 series AI GPUs. This won't be a disaggregated GPU such as the "Navi 31" and "Navi 32," but rather a scale-out multi-chip module of two GPU dies that can otherwise run on their own in single-die packages. You want to call this a multi-GPU-on-a-stick? Go ahead, but there are a couple of changes.

On AMD's Instinct AI GPUs, the chiplets have full cache coherence with each other, and can address memory controlled by each other. This cache coherence makes the chiplets work like one giant chip. In a multi-GPU-on-a-stick, there would be no cache coherence, the two dies would be mapped by the host machine as two separate devices, and then you'd be at the mercy of implicit or explicit multi-GPU technologies for performance to scale. This isn't what's happening on AI GPUs—despite multiple chiplets, the GPU is seen by the host as a single PCI device with all its cache and memory visible to software as a contiguously addressable block.



We imagine the "Navi 48" is modeled along the same lines as the company's AI GPUs. The graphics driver sees this package as a single GPU. For this to work, the two chiplets are probably connected by Infinity Fabric Fanout links—an interconnect with a much higher amount of bandwidth than a serial bus like PCIe. This is probably needed for the cache coherence to be effective. The "Navi 44" is probably just one of these chiplets sitting its own package.

In the render, the substrate and package is made to resemble that of the "Navi 32," which tends to agree with the theory that "Navi 48" will be a performance segment GPU, and a successor to the "Navi 32," "Navi 22," and "Navi 10," rather than being a successor to enthusiast-segment GPUs like the "Navi 21" and "Navi 31." This much was made clear by AMD in its recent interviews with the media.

Do we think the ChipHell rumor is plausible? Absolutely, considering nobody took the very first such renders about the AM5 package having an oddly-shaped IHS seriously. The "Navi 48" being a chiplet-based GPU is something within character for a company like AMD, which loves chiplets, MCMs, and disaggregated devices.

View at TechPowerUp Main Site | Source
 
Joined
May 19, 2011
Messages
106 (0.02/day)
With such an approach, their GPU compute dies would be much larger than they currently are. That would explain them exiting the high-end segment though… if that picture is anywhere near accurate, 2 dies would be a sensible maximum for a consumer product.
 
Joined
Jun 11, 2017
Messages
276 (0.10/day)
Location
Montreal Canada
So basicly Crossfire on a single card. Or now Single chip.

Remember the good ole days of SLI.
 
Joined
Apr 29, 2014
Messages
4,290 (1.11/day)
Location
Texas
System Name SnowFire / The Reinforcer
Processor i7 10700K 5.1ghz (24/7) / 2x Xeon E52650v2
Motherboard Asus Strix Z490 / Dell Dual Socket (R720)
Cooling RX 360mm + 140mm Custom Loop / Dell Stock
Memory Corsair RGB 16gb DDR4 3000 CL 16 / DDR3 128gb 16 x 8gb
Video Card(s) GTX Titan XP (2025mhz) / Asus GTX 950 (No Power Connector)
Storage Samsung 970 1tb NVME and 2tb HDD x4 RAID 5 / 300gb x8 RAID 5
Display(s) Acer XG270HU, Samsung G7 Odyssey (1440p 240hz)
Case Thermaltake Cube / Dell Poweredge R720 Rack Mount Case
Audio Device(s) Realtec ALC1150 (On board)
Power Supply Rosewill Lightning 1300Watt / Dell Stock 750 / Brick
Mouse Logitech G5
Keyboard Logitech G19S
Software Windows 11 Pro / Windows Server 2016
I know its a rumor/leak but its not surprising. AMD has been pushing/working on this idea for awhile. Would allow them to scale up performance with smaller chips allowing for (Hopefully) better yields. The issue is going to be making sure the chips can communicate fast enough so there is no latency.
 

FreedomEclipse

~Technological Technocrat~
Joined
Apr 20, 2007
Messages
24,045 (3.74/day)
Location
London,UK
System Name DarnGosh Edition
Processor AMD 7800X3D
Motherboard MSI X670E GAMING PLUS
Cooling Thermalright AM5 Contact Frame + Phantom Spirit 120SE
Memory G.Skill Trident Z5 NEO DDR5 6000 CL32-38-38-96
Video Card(s) Asus Dual Radeon™ RX 6700 XT OC Edition
Storage WD SN770 1TB (Boot)| 2x 2TB WD SN770 (Gaming)| 2x 2TB Crucial BX500| 2x 3TB Toshiba DT01ACA300
Display(s) LG GP850-B
Case Corsair 760T (White) {1xCorsair ML120 Pro|5xML140 Pro}
Audio Device(s) Yamaha RX-V573|Speakers: JBL Control One|Auna 300-CN|Wharfedale Diamond SW150
Power Supply Seasonic Focus GX-850 80+ GOLD
Mouse Logitech G502 X
Keyboard Duckyshine Dead LED(s) III
Software Windows 11 Home
Benchmark Scores ლ(ಠ益ಠ)ლ
Ive been saying this since the dual core days when I was running my AMD64 X2 3800+ (939, Manchester, clocked at 2.66Ghz...) If we could have dual core CPUs then why couldnt we have dual core GPUs? Then we went down the road of having two GPUs on one PCB... those all died. crossfire died, SLi died and now we are finally going somewhere.

Its confused me for the longest time why we couldnt have a dual or quad chiplet GPu design. I thought that all the knowledge and experience that AMD gained working on desktop and server chips would carry over to the GPu side of things but it never did till now.
 
Joined
Mar 7, 2011
Messages
4,543 (0.91/day)
So basicly Crossfire on a single card. Or now Single chip.

Remember the good ole days of SLI.
This is similar to how Apple has been building their large size M chips for desktops not surprising to see Crossfire/SLI onto GPU die instead of through PCIe Slots.
 

Space Lynx

Astronaut
Joined
Oct 17, 2014
Messages
17,199 (4.66/day)
Location
Kepler-186f
I already thought 7900 XTX was a dual chiplet gpu? what am I missing... why is this big news?
 
Joined
Jun 11, 2017
Messages
276 (0.10/day)
Location
Montreal Canada
Again driver see's one chip but it's actually dual core dual CPU.

Which is crossfire on a single chip with dual GPU.

And again it was done before. 25 years ago.

I sure Nvidia will do the same with something soon a Dual GPU on a Single card.

History repeating itself again. Two was always better than one.

Or if you were one of the lucky ones to get the monster before they went out of business

Even today with some tweak this card can still play games It was way way ahead of it's time.
 
Joined
Dec 12, 2016
Messages
1,828 (0.63/day)
I already thought 7900 XTX was a dual chiplet gpu? what am I missing... why is this big news?
The 7900XTX is one GPU die with six memory controller/cache dies. This reimaging is two GPUs 'glued' together with each GPU containing everything: cores, cache, memory controller, etc.

If this is the way AMD goes, then I'm guessing each die has 40 CUs. That might also make sense with Strix Halo having one of these dies.
 
Joined
Jun 11, 2017
Messages
276 (0.10/day)
Location
Montreal Canada
The 7900XTX is one GPU die with six memory controller/cache dies. This reimaging is two GPUs 'glued' together with each GPU containing everything: cores, cache, memory controller, etc.

If this is the way AMD goes, then I'm guessing each die has 40 CUs. That might also make sense with Strix Halo having one of these dies.
Ya well since the MB makers and Video card makers decided to drop crossfire and SLI because most of the hard core gamers became soft gamers no one wanted high end. I guess we were a dying breed of gamers that demanded the best from Video card companies. Now its all about the driver the DLSS and the software rendering.
 
Joined
Apr 29, 2019
Messages
20 (0.01/day)

gpu.jpg
 
Joined
Feb 3, 2017
Messages
3,747 (1.32/day)
Processor Ryzen 7800X3D
Motherboard ROG STRIX B650E-F GAMING WIFI
Memory 2x16GB G.Skill Flare X5 DDR5-6000 CL36 (F5-6000J3636F16GX2-FX5)
Video Card(s) INNO3D GeForce RTX™ 4070 Ti SUPER TWIN X2
Storage 2TB Samsung 980 PRO, 4TB WD Black SN850X
Display(s) 42" LG C2 OLED, 27" ASUS PG279Q
Case Thermaltake Core P5
Power Supply Fractal Design Ion+ Platinum 760W
Mouse Corsair Dark Core RGB Pro SE
Keyboard Corsair K100 RGB
VR HMD HTC Vive Cosmos
What exactly makes this rumor even remotely plausible?
 
Joined
Nov 27, 2022
Messages
57 (0.08/day)
Ive been saying this since the dual core days when I was running my AMD64 X2 3800+ (939, Manchester, clocked at 2.66Ghz...) If we could have dual core CPUs then why couldnt we have dual core GPUs? Then we went down the road of having two GPUs on one PCB... those all died. crossfire died, SLi died and now we are finally going somewhere.

Its confused me for the longest time why we couldnt have a dual or quad chiplet GPu design. I thought that all the knowledge and experience that AMD gained working on desktop and server chips would carry over to the GPu side of things but it never did till now.
Watch this, especially from 10:50
 
Joined
Dec 16, 2017
Messages
2,912 (1.15/day)
System Name System V
Processor AMD Ryzen 5 3600
Motherboard Asus Prime X570-P
Cooling Cooler Master Hyper 212 // a bunch of 120 mm Xigmatek 1500 RPM fans (2 ins, 3 outs)
Memory 2x8GB Ballistix Sport LT 3200 MHz (BLS8G4D32AESCK.M8FE) (CL16-18-18-36)
Video Card(s) Gigabyte AORUS Radeon RX 580 8 GB
Storage SHFS37A240G / DT01ACA200 / ST10000VN0008 / ST8000VN004 / SA400S37960G / SNV21000G / NM620 2TB
Display(s) LG 22MP55 IPS Display
Case NZXT Source 210
Audio Device(s) Logitech G430 Headset
Power Supply Corsair CX650M
Software Whatever build of Windows 11 is being served in Canary channel at the time.
Benchmark Scores Corona 1.3: 3120620 r/s Cinebench R20: 3355 FireStrike: 12490 TimeSpy: 4624
then why couldnt we have dual core GPUs?
We had something close to that with Crossfire and SLI.

I imagine that, due to GPUs having to present things on screen at a very fast rate and with extreme care of not resulting in, say, half the screen being a couple frames behind the other half of the screen, it's probably very complicated to do that when the two GPU "cores" are separated. Which is why SLI/Crossfire had their bridges at first and data synchronization over PCIE later, plus the performance gain wasn't that great.

Plus if this is exposed to the software, developers are gonna whine about having to code for basically explicit multi-GPU lite™.
 
Joined
Nov 26, 2021
Messages
1,645 (1.51/day)
Location
Mississauga, Canada
Processor Ryzen 7 5700X
Motherboard ASUS TUF Gaming X570-PRO (WiFi 6)
Cooling Noctua NH-C14S (two fans)
Memory 2x16GB DDR4 3200
Video Card(s) Reference Vega 64
Storage Intel 665p 1TB, WD Black SN850X 2TB, Crucial MX300 1TB SATA, Samsung 830 256 GB SATA
Display(s) Nixeus NX-EDG27, and Samsung S23A700
Case Fractal Design R5
Power Supply Seasonic PRIME TITANIUM 850W
Mouse Logitech
VR HMD Oculus Rift
Software Windows 11 Pro, and Ubuntu 20.04
We had something close to that with Crossfire and SLI.

I imagine that, due to GPUs having to present things on screen at a very fast rate and with extreme care of not resulting in, say, half the screen being a couple frames behind the other half of the screen, it's probably very complicated to do that when the two GPU "cores" are separated. Which is why SLI/Crossfire had their bridges at first and data synchronization over PCIE later, plus the performance gain wasn't that great.

Plus if this is exposed to the software, developers are gonna whine about having to code for basically explicit multi-GPU lite™.
If the die to die bandwidth is in the same ballpark as MI300X, then it won't be a problem. However, I'm very skeptical of this rumour. It makes sense for the putative high end RDNA 4, but doesn't seem likely for lower end products.

1730218075459.png
 
Joined
Feb 3, 2017
Messages
3,747 (1.32/day)
Processor Ryzen 7800X3D
Motherboard ROG STRIX B650E-F GAMING WIFI
Memory 2x16GB G.Skill Flare X5 DDR5-6000 CL36 (F5-6000J3636F16GX2-FX5)
Video Card(s) INNO3D GeForce RTX™ 4070 Ti SUPER TWIN X2
Storage 2TB Samsung 980 PRO, 4TB WD Black SN850X
Display(s) 42" LG C2 OLED, 27" ASUS PG279Q
Case Thermaltake Core P5
Power Supply Fractal Design Ion+ Platinum 760W
Mouse Corsair Dark Core RGB Pro SE
Keyboard Corsair K100 RGB
VR HMD HTC Vive Cosmos
Joined
Jun 22, 2014
Messages
446 (0.12/day)
System Name Desktop / "Console"
Processor Ryzen 5950X / Ryzen 5800X
Motherboard Asus X570 Hero / Asus X570-i
Cooling EK AIO Elite 280 / Cryorig C1
Memory 32GB Gskill Trident DDR4-3600 CL16 / 16GB Crucial Ballistix DDR4-3600 CL16
Video Card(s) RTX 4090 FE / RTX 2080ti FE
Storage 1TB Samsung 980 Pro, 1TB Sabrent Rocket 4 Plus NVME / 1TB Sabrent Rocket 4 NVME, 1TB Intel 660P
Display(s) Alienware AW3423DW / LG 65CX Oled
Case Lian Li O11 Mini / Sliger CL530 Conswole
Audio Device(s) Sony AVR, SVS speakers & subs / Marantz AVR, SVS speakers & subs
Power Supply ROG Loki 1000 / Silverstone SX800
VR HMD Quest 3
What exactly makes this rumor even remotely plausible?
This is what the nVidia B200 is doing, using two chiplets that behave as a single and are seen as a single die by the software. Or did you mean remotely plausible for AMD to pull off? Granted, the B200 is NOT a consumer level GPU...

"For their first multi-die chip, NVIDIA is intent on skipping the awkward “two accelerators on one chip” phase, and moving directly on to having the entire accelerator behave as a single chip. According to NVIDIA, the two dies operate as “one unified CUDA GPU”, offering full performance with no compromises."


I already thought 7900 XTX was a dual chiplet gpu? what am I missing... why is this big news?
It is different in that instead of making the entire GPU die on the 5nm node, they took the cache and memory controllers and fabbed them as chiplets on the older 6nm node because these parts do not benefits so much from a node shrink. All of the chiplets were then arranged to make a full die. This was an ingenious way to target the parts of the GPU getting the largest performance benefits of the 5nm node shrink, while saving cost by not using a cutting edge node on the parts that do not. Fantastic engineering in my opinion.

 
Last edited:
Joined
Jul 13, 2016
Messages
3,272 (1.07/day)
Processor Ryzen 7800X3D
Motherboard ASRock X670E Taichi
Cooling Noctua NH-D15 Chromax
Memory 32GB DDR5 6000 CL30
Video Card(s) MSI RTX 4090 Trio
Storage Too much
Display(s) Acer Predator XB3 27" 240 Hz
Case Thermaltake Core X9
Audio Device(s) Topping DX5, DCA Aeon II
Power Supply Seasonic Prime Titanium 850w
Mouse G305
Keyboard Wooting HE60
VR HMD Valve Index
Software Win 10
If the die to die bandwidth is in the same ballpark as MI300X, then it won't be a problem. However, I'm very skeptical of this rumour. It makes sense for the putative high end RDNA 4, but doesn't seem likely for lower end products.

View attachment 369338

My thoughts exactly. The MI300X is using a more expensive interconnect. The whole reason AMD went with an organic substrate on consumer cards is to save cost.

Doesn't make a lot of sense with the rumor that AMD is targeting mid-range and low end this gen unless they solved the cost issue.
 
Joined
Mar 10, 2024
Messages
11 (0.04/day)
Location
Hungary
System Name Main rig
Processor Intel Core i5-14600k
Motherboard TUF GAMING B760M-PLUS
Cooling Be Quiet! DARK ROCK PRO 5
Memory 32 GB DDR5 6000 MHz
Video Card(s) RTX 3060 Ti GDDR6X
Storage Kingston KC3000 1TB, Samsung 970 evo plus, Kingmax 480 GB SSD, Western Digital WD Red Plus 3.5 3TB
Display(s) 1080p
Case Fractal Design Focus 2
Power Supply Seasonic FOCUS GX Series 750W
What's next? "AI Imagines AMD "Navi 58" RDNA 5 to be a Quad-Chiplet GPU"?
 
Joined
Jun 11, 2017
Messages
276 (0.10/day)
Location
Montreal Canada
It' strange Nvidia did not go dual or quad GPU's on a single die. I mean it would have made sense, First power efficent for sure. Use 4 lower power GPU's into one die link them with Nvlink which they are using in their high end systems. You could have a single card GPU with like only 200 watts of power with more GPU power than a single big huge die sucking back 600 watts of power. I personally is still running an SLI rig. Still have twin 1080 ti GTX and games even out today run fine and dandy on them even though people say SLI is not supported. Ya it might not be supported in the Game as an Option to turn on. But in the driver is fully supported for all games. Most of the new games only see the one card even though the two is connected and Yes both are working in the game because both of them fans speed up and they are heating up. I also test this theory on some newer games with a watt o meter. I ran one game said no support so I run just one 1080 GTX TI measured wattage running game system was drawing about 460 watts in course of game then Switched back to SLI changed the driver config loaded up game and this time wattage was around 670 watts running the game full. So in turn it was using both card to render. I run 4k in all games so at 2160p one card is rendering the bottom and the other the top.

I remember back in the days of 1999, 2000 when people had two Diamond Monster Voodoo2's in SLI. I had a rig exactly the same. I met many people that bought the same setup and they complained they did not see much of a difference well guess what 90% of them did not know how to configure the drivers to make the game use the cards. I show them the settings what to change and where in the game and driver to change them and they were like OMG this is freaking awesome. Unreal orginial ran like butter on those Voodoo2's remember playing alot of ut tournament.

I'm an old school gamer. Heck yes my first home system was an atari 2600. So when voodoo's was around I was 25 year old. I'm an old timer now. So have lots of knowelege on systems since 1985 and on. It's like MS telling me my Core i7 3770k with 64 GB ram and 1 nvme 1TB and 500 SSD with a quadro card did not meet the min requirements to run windows 11 Pro. Um ok and your telling me a Core i5 that runs 2 GHZ slower than my i7 is going to be faster. I bypassed that BS and windows 11 on that computer at home runs faster then my work computers at office and they at i5 10500T's.

It's like comparing a harley motorcycle to and Ninja. Sure the Ninja has lots of CC's of power but that harley has raw power and it will win.
 
Last edited:
Top