• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.

Upcoming GDDR7

Joined
Sep 17, 2014
Messages
22,479 (6.03/day)
Location
The Washing Machine
System Name Tiny the White Yeti
Processor 7800X3D
Motherboard MSI MAG Mortar b650m wifi
Cooling CPU: Thermalright Peerless Assassin / Case: Phanteks T30-120 x3
Memory 32GB Corsair Vengeance 30CL6000
Video Card(s) ASRock RX7900XT Phantom Gaming
Storage Lexar NM790 4TB + Samsung 850 EVO 1TB + Samsung 980 1TB + Crucial BX100 250GB
Display(s) Gigabyte G34QWC (3440x1440)
Case Lian Li A3 mATX White
Audio Device(s) Harman Kardon AVR137 + 2.1
Power Supply EVGA Supernova G2 750W
Mouse Steelseries Aerox 5
Keyboard Lenovo Thinkpad Trackpoint II
VR HMD HD 420 - Green Edition ;)
Software W11 IoT Enterprise LTSC
Benchmark Scores Over 9000
As others have said VRAM allocated isn't the same thing as VRAM needed. That hit home to me about 10 years ago on a review of COD: Advanced Warfare here. The game was running fine on 3 GB cards at all resolutions and it was even stated in the review that the engine was just loading up the VRAM on cards with more VRAM just because it was there without actually needing it and one member here running a Maxwell Titan X said his card was showing 11 GB used.
Highly game specific statement, definitely not applicable as a blanket statement.

You can monitor the actual VRAM in use these days with RTSS. Do it. Its very interesting to see the differences. You will find some games absolutely do use almost everything they allocated, and if you allocate less, something's gonna give somewhere. Its not always stutter, these days, its also reduced texture quality, more pop in, etc. You can even sometimes detect post processing effects to generate less quickly each frame.
 
Joined
Feb 1, 2019
Messages
3,607 (1.69/day)
Location
UK, Midlands
System Name Main PC
Processor 13700k
Motherboard Asrock Z690 Steel Legend D4 - Bios 13.02
Cooling Noctua NH-D15S
Memory 32 Gig 3200CL14
Video Card(s) 4080 RTX SUPER FE 16G
Storage 1TB 980 PRO, 2TB SN850X, 2TB DC P4600, 1TB 860 EVO, 2x 3TB WD Red, 2x 4TB WD Red
Display(s) LG 27GL850
Case Fractal Define R4
Audio Device(s) Soundblaster AE-9
Power Supply Antec HCG 750 Gold
Software Windows 10 21H2 LTSC
To backup Vayra's point, those of us who have moaned about VRAM problems, we dont do it just because a tool reports high usage, its usually because we notice symptoms. I have experienced all of the following in games I played on my RTX 3080.

Game crashing due to running out of VRAM.
Texture pop in's.
Delayed texture loading due to constant flushes to make room.
Reduced LOD textures.
Stutters.
Glitches.
Textures that dont load in at all, completely missing.
Excessive texture unloading, so e.g. cant do a 360 in game whilst keeping everything in view loaded.

Many years back games would just routinely crash when no resources left, now days things are more dynamic and so people get the impression there is no problem. An example might be the game will let you pick ultra textures, and instead of crashing it will make you think you using ultra but instead it dynamically lowers the quality to keep the game running.
 
Last edited:
Joined
Feb 3, 2017
Messages
3,759 (1.32/day)
Processor Ryzen 7800X3D
Motherboard ROG STRIX B650E-F GAMING WIFI
Memory 2x16GB G.Skill Flare X5 DDR5-6000 CL36 (F5-6000J3636F16GX2-FX5)
Video Card(s) INNO3D GeForce RTX™ 4070 Ti SUPER TWIN X2
Storage 2TB Samsung 980 PRO, 4TB WD Black SN850X
Display(s) 42" LG C2 OLED, 27" ASUS PG279Q
Case Thermaltake Core P5
Power Supply Fractal Design Ion+ Platinum 760W
Mouse Corsair Dark Core RGB Pro SE
Keyboard Corsair K100 RGB
VR HMD HTC Vive Cosmos
Now the only question is the VRAM size. GDDR7 will still be produced using 2 GB memory chips, and later on with 3 GB in size. Maybe that's the reason Nvidia pushed back its next gen to 2025. Only way to increase the VRAM size without increasing bus width while still using 2 GB chips, would be to use socalled clamshell mode where memory chips are stacked at each side of the PCB, this however doesn't offer any memory bandwidth improvement, just doubles the memory size, and it's also cheaper for production. I believe that high end GPUs will use 3 GB chip size on a 384-bit (36 GB) wide bus, while mid-to-high tier will be on 2 GB size (24 GB) using 384-bit wide bus (something like a RTX 5080 perhaps).
I bet this is exactly why Nvidia postponed. There has been a lot of complaining about memory sizes and they need to do something. At least Micron seems to have 24Gbit GDDR7 in production so waiting for these would make sense. It would allow basically to increase memory capacity to the next size:
128bit bus, 4 chips - 8GB with 2GB chips, 12GB with 3GB chips
192bit bus, 6 chips - 12GB with 2GB chips, 18GB with 3GB chips
256bit bus, 8 chips - 16GB with 2GB chips, 24GB with 3GB chips
384bit bus, 12 chips - 24GB with 2GB chips, 36GB with 3GB chips

I don't think anything mid-high tier would get 386bit bus, not even 5080. At least not when 3GB GDDR7 is available. like you calculated, essentially 4090 bandwidth and 24GB will end up quote optimal there. And clamshell is not something a manufacturer wants to do. Costs more, is a bit more complex, needs cooling on the backside. GDDR6X was pretty hot and 3090s with clamshell VRAM showed that pretty convincingly.

Its funny how Nvidia prioritise bandwidth over capacity. Whilst also fighting tooth and nail to make the bus width as small as possible, causing their weird 6/10/12 gig setups.
Pushing bus width down was AMDs playbook, they did it a generation earlier. Same with large cache to mitigate this change. What bit Nvidia back was probably them expecting 2GB GDDR6X to be available sooner.

Looks like there are going to be even faster speeds for GDDR7.
GDDR7 spec sees up to 48Gbps and up to 8GB chips. Whether and when manufacturers will be able to build something like that we will see. Micron is currently at 2GB and 3GB sizes and modest speeds. Samsung announcing they have built faster stuff has been tricky before, it took years for them to get fast GDDR6 from announcement to production.
 
Last edited:
Joined
Mar 7, 2023
Messages
907 (1.43/day)
System Name BarnacleMan
Processor 14700KF
Motherboard Gigabyte B760 Aorus Elite Ax DDR5
Cooling ARCTIC Liquid Freezer II 240 + P12 Max Fans
Memory 32GB Kingston Fury Beast
Video Card(s) Asus Tuf 4090 24GB
Storage 4TB sn850x, 2TB sn850x, 2TB Netac Nv7000 + 2TB p5 plus, 4TB MX500 * 2 = 18TB. Plus dvd burner.
Display(s) Dell 23.5" 1440P IPS panel
Case Lian Li LANCOOL II MESH Performance Mid-Tower
Audio Device(s) Logitech Z623
Power Supply Gigabyte ud850gm pg5
We learned last year that GDDR7 will be running at around 32 Gbps, quite an improvement of the current GDDR6X. I did some research and came up with some interesting numbers regarding bandwidth, bus width and memory clocks for the upcoming memory. I may be wrong, so please do correct me if I'm mistaken.

Most importantly, GDDR7 will be using PAM3 and NRZ signaling, while current GDDR6X uses PAM4 signaling. NRZ (non return to zero) signaling can transfer 1 bit of data per cycle, PAM4 is transfering 2 bits of data per cycle, while PAM3 will be transfering 1,5 bits per cycle. NRZ still exists in GDDR7 spec because it lowers power consumption, and will be used instead of PAM3 when there is low bandwidth need. GDDR6X clocks slower compared to other memory types because of PAM4 signaling.

So as we're currently calculating the memory effective speed on GDDR6X:

Speed = memory clock * (bit rise/fall) * QDR (quad data rate, since GDDR5) * bits per cycle

So RTX 4080 memory speed would be: 1,4 * 2 * 4 * 2 (uses PAM4) = ~22,4 Gbps for example.

Announced 32 Gbps would equal to something like: 2667 MHz memory clock, or: 2,667 * 2 * 4 * 1,5 = ~32 Gbps

A lot of people are speculating that a larger than 384-bit wide bus will be used for the next gen high end GPUs. If we take a look at memory bandwidth on a 384-bit bus, coupled with 32 Gbps effective speed, that would equal to more than 1,5 TB/s of memory bandwidth! Memory bandwidth = memory effective speed * bus width / 8 (8 bits make a byte), so 32 * 384 / 8 = 1,536 TB/s. Seems like a perfect choice for the next high end GPUs.
Even if the 256-bit bus interface will be used coupled with ~32 Gbps, it will have a memory bandwidth of 1,02 TB/s, practically the same as current RTX 4090.

Now the only question is the VRAM size. GDDR7 will still be produced using 2 GB memory chips, and later on with 3 GB in size. Maybe that's the reason Nvidia pushed back its next gen to 2025. Only way to increase the VRAM size without increasing bus width while still using 2 GB chips, would be to use socalled clamshell mode where memory chips are stacked at each side of the PCB, this however doesn't offer any memory bandwidth improvement, just doubles the memory size, and it's also cheaper for production. I believe that high end GPUs will use 3 GB chip size on a 384-bit (36 GB) wide bus, while mid-to-high tier will be on 2 GB size (24 GB) using 384-bit wide bus (something like a RTX 5080 perhaps).

Any thoughts?
When you say high end you mean like the 5090? You think it will have 36gb? Wont that endanger its enterprise lineup? Then again if its expensive enough I suppose it wont matter that much.
 
Joined
May 10, 2023
Messages
273 (0.48/day)
Location
Brazil
Processor 5950x
Motherboard B550 ProArt
Cooling Fuma 2
Memory 4x32GB 3200MHz Corsair LPX
Video Card(s) 2x RTX 3090
Display(s) LG 42" C2 4k OLED
Power Supply XPG Core Reactor 850W
Software I use Arch btw
You can monitor the actual VRAM in use these days with RTSS.
You cannot. It only shows dedicated vs shared memory. To actually see the real, in-use VRAM usage you'd need to use a debugger like Renderdoc.
 
Joined
Jan 14, 2019
Messages
12,350 (5.75/day)
Location
Midlands, UK
System Name Nebulon B
Processor AMD Ryzen 7 7800X3D
Motherboard MSi PRO B650M-A WiFi
Cooling be quiet! Dark Rock 4
Memory 2x 24 GB Corsair Vengeance DDR5-4800
Video Card(s) AMD Radeon RX 6750 XT 12 GB
Storage 2 TB Corsair MP600 GS, 2 TB Corsair MP600 R2
Display(s) Dell S3422DWG, 7" Waveshare touchscreen
Case Kolink Citadel Mesh black
Audio Device(s) Logitech Z333 2.1 speakers, AKG Y50 headphones
Power Supply Seasonic Prime GX-750
Mouse Logitech MX Master 2S
Keyboard Logitech G413 SE
Software Bazzite (Fedora Linux) KDE
I'm on a "we'll see when it comes out" approach as (nearly) always.
 
Joined
Jan 2, 2024
Messages
570 (1.70/day)
Location
Seattle
System Name DevKit
Processor AMD Ryzen 5 3600 ↗4.0GHz
Motherboard Asus TUF Gaming X570-Plus WiFi
Cooling Koolance CPU-300-H06, Koolance GPU-180-L06, SC800 Pump
Memory 4x16GB Ballistix 3200MT/s ↗3800
Video Card(s) PowerColor RX 580 Red Devil 8GB ↗1380MHz ↘1105mV, PowerColor RX 7900 XT Hellhound 20GB
Storage 240GB Corsair MP510, 120GB KingDian S280
Display(s) Nixeus VUE-24 (1080p144)
Case Koolance PC2-601BLW + Koolance EHX1020CUV Radiator Kit
Audio Device(s) Oculus CV-1
Power Supply Antec Earthwatts EA-750 Semi-Modular
Mouse Easterntimes Tech X-08, Zelotes C-12
Keyboard Logitech 106-key, Romoral 15-Key Macro, Royal Kludge RK84
VR HMD Oculus CV-1
Software Windows 10 Pro Workstation, VMware Workstation 16 Pro, MS SQL Server 2016, Fan Control v120, Blender
Benchmark Scores Cinebench R15: 1590cb Cinebench R20: 3530cb (7.83x451cb) CPU-Z 17.01.64: 481.2/3896.8 VRMark: 8009
A lot of that constant mesh and texture load/unload cycling and LoD is desktop.
In VR it wants ALL of it loaded into memory, especially in some common engines that don't optimize the experience.
So for those of us doing both, we need way more vram than looks necessary or we get really bad glitches and slowdowns.
 
Joined
Sep 3, 2019
Messages
3,519 (1.84/day)
Location
Thessaloniki, Greece
System Name PC on since Aug 2019, 1st CPU R5 3600 + ASUS ROG RX580 8GB >> MSI Gaming X RX5700XT (Jan 2020)
Processor Ryzen 9 5900X (July 2022), 220W PPT limit, 80C temp limit, CO -6-14, +50MHz (up to 5.0GHz)
Motherboard Gigabyte X570 Aorus Pro (Rev1.0), BIOS F39b, AGESA V2 1.2.0.C
Cooling Arctic Liquid Freezer II 420mm Rev7 (Jan 2024) with off-center mount for Ryzen, TIM: Kryonaut
Memory 2x16GB G.Skill Trident Z Neo GTZN (July 2022) 3667MT/s 1.42V CL16-16-16-16-32-48 1T, tRFC:280, B-die
Video Card(s) Sapphire Nitro+ RX 7900XTX (Dec 2023) 314~467W (375W current) PowerLimit, 1060mV, Adrenalin v24.10.1
Storage Samsung NVMe: 980Pro 1TB(OS 2022), 970Pro 512GB(2019) / SATA-III: 850Pro 1TB(2015) 860Evo 1TB(2020)
Display(s) Dell Alienware AW3423DW 34" QD-OLED curved (1800R), 3440x1440 144Hz (max 175Hz) HDR400/1000, VRR on
Case None... naked on desk
Audio Device(s) Astro A50 headset
Power Supply Corsair HX750i, ATX v2.4, 80+ Platinum, 93% (250~700W), modular, single/dual rail (switch)
Mouse Logitech MX Master (Gen1)
Keyboard Logitech G15 (Gen2) w/ LCDSirReal applet
Software Windows 11 Home 64bit (v24H2, OSBuild 26100.2161), upgraded from Win10 to Win11 on Jan 2024
“Many games manage memory usage differently, automatically dialing down quality settings or removing textures altogether.”


“With the exception of entry-level options, you shouldn't be buying 8GB graphics cards anymore. 12GB is now the bare minimum, and 16GB is the ideal target.”
 

Raysterize

New Member
Joined
Jun 11, 2024
Messages
8 (0.05/day)
Personally, I think the 5090 will retain the 384-bit bus and employ the G7 memory. High throughput and adequate capacity - makes for a balanced configuration. It also keeps the cost relatively down compared to using a 512-bit bus and 16 memory devices. If the 6900 XT and the RTX 4080 can teach people a lesson is that sometimes more is not better, both Navi 21 and AD103 make for harmonious utilization of the hardware, you can extract pretty much everything out of these cards.

>24 GB is currently entirely unnecessary targeting 4K or 8K DLSS gaming, and as the 7900 XTX clearly shows, it does not benefit from having 24 GB as there is virtually no situation where it is a boon over the RTX 4080's 16 GB. Whenever the XTX pans out to be faster, it is usually in compute-heavy, D3D12 or Vulkan games (where its drivers don't get in the way), it is not because of its memory, or even bandwidth, but rather because that scenario favors its architecture.

Truth be told, if they wanted to hit it out of the park, they should release another graphics card like the RTX 3090 Ti: fully enabled core, fastest memory bin, no nonsense.
My 7900XTX was using 19GB VRAM on Warzone 4K on the Ultra Preset... Sold it and went back to Xbox Series X as got banned 3 times for 2 weeks at a time! (Like going from a Ferrari to a Robin Reliant!).
 
Joined
Nov 24, 2018
Messages
2,248 (1.02/day)
Location
south wales uk
System Name 1.FortySe7en VR rig 2. intel teliscope rig 3.MSI GP72MVR Leopard Pro .E-52699, Xeon play thing
Processor 1.3900x @stock 2. i7 7700k @5. 3. i7 7700hq
Motherboard 1.aorus x570 ultra 2. z270 Maximus IX Hero,4 MR9A PRO ATX X99
Cooling 1.Hard tube loop, cpu and gpu 2. Hard loop cpu and gpu 4 360 AIO
Memory 1.Gskill neo @3600 32gb 2.hyperxfury 32gb @3000 3. 16gb hyperx @2400 4 64GB 2133 in quad channel
Video Card(s) 1.GIGABYTE RTX 3080 WaterForce WB 2. Aorus RTX2080 3. 1060 3gb. 4 Arc 770LE 16 gb
Storage 1 M.2 500gb , 2 3tb HDs 2. 256gb ssd, 3tbHD 3. 256 m.2. 1tb ssd 4. 2gb ssd
Display(s) 1.LG 50" UHD , 2 MSI Optix MAG342C UWHD. 3.17" 120 hz display 4. Acer Preditor 144hz 32inch.z
Case 1. Thermaltake P5 2. Thermaltake P3 4. some cheapo case that should not be named.
Audio Device(s) 1 Onboard 2 Onboard 3 Onboard 4. onboard.
Power Supply 1.seasonic gx 850w 2. seasonic gx 750w. 4 RM850w
Mouse 1 ROG Gladius 2 Corsair m65 pro
Keyboard 1. ROG Strix Flare 2. Corsair F75 RBG 3. steelseries RBG
VR HMD rift and rift S and Quest 2.
Software 1. win11 pro 2. win11 pro 3, win11 home 4 win11 pro
Benchmark Scores 1.7821 cb20 ,cb15 3442 1c 204 cpu-z 1c 539 12c 8847 2. 1106 cb 3.cb 970
My God i've just taken in GDDR6, not again my brain canny not take it :) .
 
Joined
Feb 1, 2019
Messages
3,607 (1.69/day)
Location
UK, Midlands
System Name Main PC
Processor 13700k
Motherboard Asrock Z690 Steel Legend D4 - Bios 13.02
Cooling Noctua NH-D15S
Memory 32 Gig 3200CL14
Video Card(s) 4080 RTX SUPER FE 16G
Storage 1TB 980 PRO, 2TB SN850X, 2TB DC P4600, 1TB 860 EVO, 2x 3TB WD Red, 2x 4TB WD Red
Display(s) LG 27GL850
Case Fractal Define R4
Audio Device(s) Soundblaster AE-9
Power Supply Antec HCG 750 Gold
Software Windows 10 21H2 LTSC
I still have my 3080 (and my 9900k with 32 gigs of ram, so a reasonable base system too), so now have the luxury of comparing 10 gigs to 16 gigs in games, and have tested a couple and will test some more after a games purchasing spree last night, results are satisfying in terms of my 4080 purchase.
 
Last edited:
Joined
Feb 3, 2017
Messages
3,759 (1.32/day)
Processor Ryzen 7800X3D
Motherboard ROG STRIX B650E-F GAMING WIFI
Memory 2x16GB G.Skill Flare X5 DDR5-6000 CL36 (F5-6000J3636F16GX2-FX5)
Video Card(s) INNO3D GeForce RTX™ 4070 Ti SUPER TWIN X2
Storage 2TB Samsung 980 PRO, 4TB WD Black SN850X
Display(s) 42" LG C2 OLED, 27" ASUS PG279Q
Case Thermaltake Core P5
Power Supply Fractal Design Ion+ Platinum 760W
Mouse Corsair Dark Core RGB Pro SE
Keyboard Corsair K100 RGB
VR HMD HTC Vive Cosmos
I still have my 3080 (and my 9900k with 32 gigs of ram, so a reasonable base system too), so now have the luxury of comparing 10 gigs to 16 gigs in games, and have tested a couple and will test some more after a games purchasing spree last night, results are satisfying in terms of my 4080 purchase.
VRAM aside, 4080 is 45-50% faster :)
 
Joined
Feb 1, 2019
Messages
3,607 (1.69/day)
Location
UK, Midlands
System Name Main PC
Processor 13700k
Motherboard Asrock Z690 Steel Legend D4 - Bios 13.02
Cooling Noctua NH-D15S
Memory 32 Gig 3200CL14
Video Card(s) 4080 RTX SUPER FE 16G
Storage 1TB 980 PRO, 2TB SN850X, 2TB DC P4600, 1TB 860 EVO, 2x 3TB WD Red, 2x 4TB WD Red
Display(s) LG 27GL850
Case Fractal Define R4
Audio Device(s) Soundblaster AE-9
Power Supply Antec HCG 750 Gold
Software Windows 10 21H2 LTSC
VRAM aside, 4080 is 45-50% faster :)
Yeah, I am not comparing framerates, I am checking VRAM specific behaviour, texture quality issues, pop in, game stability, that sort of thing.
 
Joined
Mar 7, 2023
Messages
907 (1.43/day)
System Name BarnacleMan
Processor 14700KF
Motherboard Gigabyte B760 Aorus Elite Ax DDR5
Cooling ARCTIC Liquid Freezer II 240 + P12 Max Fans
Memory 32GB Kingston Fury Beast
Video Card(s) Asus Tuf 4090 24GB
Storage 4TB sn850x, 2TB sn850x, 2TB Netac Nv7000 + 2TB p5 plus, 4TB MX500 * 2 = 18TB. Plus dvd burner.
Display(s) Dell 23.5" 1440P IPS panel
Case Lian Li LANCOOL II MESH Performance Mid-Tower
Audio Device(s) Logitech Z623
Power Supply Gigabyte ud850gm pg5
I still have my 3080 (and my 9900k with 32 gigs of ram, so a reasonable base system too), so now have the luxury of comparing 10 gigs to 16 gigs in games, and have tested a couple and will test some more after a games purchasing spree last night, results are satisfying in terms of my 4080 purchase.
You know recently I was wondering if anybody had created a consumer facing utility to eat xxx amount of vram. It would be really helpful when trying to help find solutions for people on low vram cards. When you have a big pool of vram you never quite know if a game is caching things because it knows its has the room to do so, or if its just being used irresponsibly. Also yeah ' idk works fine on my 4090' is not a popular phrase.

Really though, I like to try and be helpful on steam and being able to limit to 8 or 12 would be super useful. Sometimes it feels like a fool's errand though. Especially when you're talking to people who are say used to being console gamers where things just work and don't really know about troubleshooting. You also got to be careful when using terminology that you forget an average gamer might not know. Sure its easy enough to say, something like 'reduce all your cores by 400mhz and turn xmp off and if the error no longer appears, then the cpu is most likely degraded.' I think I've said some variation of that maybe 5 times and let me tell you it is not something people like to hear even though it literally takes less than a couple minutes to set up. Instead I usually get some variant of 'but it worked before' Well yes, things generally work before they don't. I've even offered to walk people through it. But idk. They still respond with hostility.

Why do I bother? Idk. I guess I just want people to enjoy the games that I enjoy....

I feel bad for the developers who basically get put into tech support roles due to stuff like this.
 
Joined
May 10, 2023
Messages
273 (0.48/day)
Location
Brazil
Processor 5950x
Motherboard B550 ProArt
Cooling Fuma 2
Memory 4x32GB 3200MHz Corsair LPX
Video Card(s) 2x RTX 3090
Display(s) LG 42" C2 4k OLED
Power Supply XPG Core Reactor 850W
Software I use Arch btw
You know recently I was wondering if anybody had created a consumer facing utility to eat xxx amount of vram.
It's easily doable, but idk if there's such an app for windows.
See this video of mine I did years ago (I did it as an example of allocated vs actually used VRAM):
 
Joined
Mar 7, 2023
Messages
907 (1.43/day)
System Name BarnacleMan
Processor 14700KF
Motherboard Gigabyte B760 Aorus Elite Ax DDR5
Cooling ARCTIC Liquid Freezer II 240 + P12 Max Fans
Memory 32GB Kingston Fury Beast
Video Card(s) Asus Tuf 4090 24GB
Storage 4TB sn850x, 2TB sn850x, 2TB Netac Nv7000 + 2TB p5 plus, 4TB MX500 * 2 = 18TB. Plus dvd burner.
Display(s) Dell 23.5" 1440P IPS panel
Case Lian Li LANCOOL II MESH Performance Mid-Tower
Audio Device(s) Logitech Z623
Power Supply Gigabyte ud850gm pg5
It's easily doable, but idk if there's such an app for windows.
See this video of mine I did years ago (I did it as an example of allocated vs actually used VRAM):
You've posted this somewhere else haven't you? I've searched for such a tool and I am 99% sure I have seen this at one point.
 
Joined
Dec 25, 2020
Messages
6,806 (4.74/day)
Location
São Paulo, Brazil
System Name "Icy Resurrection"
Processor 13th Gen Intel Core i9-13900KS Special Edition
Motherboard ASUS ROG MAXIMUS Z790 APEX ENCORE
Cooling Noctua NH-D15S upgraded with 2x NF-F12 iPPC-3000 fans and Honeywell PTM7950 TIM
Memory 32 GB G.SKILL Trident Z5 RGB F5-6800J3445G16GX2-TZ5RK @ 7600 MT/s 36-44-44-52-96 1.4V
Video Card(s) ASUS ROG Strix GeForce RTX™ 4080 16GB GDDR6X White OC Edition
Storage 500 GB WD Black SN750 SE NVMe SSD + 4 TB WD Red Plus WD40EFPX HDD
Display(s) 55-inch LG G3 OLED
Case Pichau Mancer CV500 White Edition
Power Supply EVGA 1300 G2 1.3kW 80+ Gold
Mouse Microsoft Classic Intellimouse
Keyboard Generic PS/2
Software Windows 11 IoT Enterprise LTSC 24H2
Benchmark Scores I pulled a Qiqi~
You've posted this image somewhere else haven't you? I've searched for such a tool and I am 99% sure I saw that photo at one point.

Was here on TPU on a thread some time ago. Neat software, that one.
 
Joined
Mar 7, 2023
Messages
907 (1.43/day)
System Name BarnacleMan
Processor 14700KF
Motherboard Gigabyte B760 Aorus Elite Ax DDR5
Cooling ARCTIC Liquid Freezer II 240 + P12 Max Fans
Memory 32GB Kingston Fury Beast
Video Card(s) Asus Tuf 4090 24GB
Storage 4TB sn850x, 2TB sn850x, 2TB Netac Nv7000 + 2TB p5 plus, 4TB MX500 * 2 = 18TB. Plus dvd burner.
Display(s) Dell 23.5" 1440P IPS panel
Case Lian Li LANCOOL II MESH Performance Mid-Tower
Audio Device(s) Logitech Z623
Power Supply Gigabyte ud850gm pg5
I bet this is exactly why Nvidia postponed. There has been a lot of complaining about memory sizes and they need to do something. At least Micron seems to have 24Gbit GDDR7 in production so waiting for these would make sense. It would allow basically to increase memory capacity to the next size:
128bit bus, 4 chips - 8GB with 2GB chips, 12GB with 3GB chips
192bit bus, 6 chips - 12GB with 2GB chips, 18GB with 3GB chips
256bit bus, 8 chips - 16GB with 2GB chips, 24GB with 3GB chips
384bit bus, 12 chips - 24GB with 2GB chips, 36GB with 3GB chips

I don't think anything mid-high tier would get 386bit bus, not even 5080. At least not when 3GB GDDR7 is available. like you calculated, essentially 4090 bandwidth and 24GB will end up quote optimal there. And clamshell is not something a manufacturer wants to do. Costs more, is a bit more complex, needs cooling on the backside. GDDR6X was pretty hot and 3090s with clamshell VRAM showed that pretty convincingly.

Pushing bus width down was AMDs playbook, they did it a generation earlier. Same with large cache to mitigate this change. What bit Nvidia back was probably them expecting 2GB GDDR6X to be available sooner.

GDDR7 spec sees up to 48Gbps and up to 8GB chips. Whether and when manufacturers will be able to build something like that we will see. Micron is currently at 2GB and 3GB sizes and modest speeds. Samsung announcing they have built faster stuff has been tricky before, it took years for them to get fast GDDR6 from announcement to production.
Recently I've been worrying they might end up going 128 bit for the 5070 too. With 3gb chips thats 12GB with a respectable amount of bandwidth. I mean I hope its more but we know how nvidia rolls. Fastest, newest, ram modules but bare minimum bus, bare minimum capacity unless its a halo card or a very weak card that has no chance to compete with enterprise cards. Mid-range, or even high end 80 series cards still seem to get the bare minimum. I think the 3080 was the biggest insult. Have to go all the way to 3080 ti to get the same vram as the 3060??? And it was similar with the 40 series. Sure you can have a 16GB 4060 ti. But if you want vram and power, you gotta pay $1200 for a 4080. Mind you, its a little better now after the refresh, 4070 ti super can get you 16gb for $800. But thats still... a lot of money.
 
Last edited:
Joined
May 10, 2023
Messages
273 (0.48/day)
Location
Brazil
Processor 5950x
Motherboard B550 ProArt
Cooling Fuma 2
Memory 4x32GB 3200MHz Corsair LPX
Video Card(s) 2x RTX 3090
Display(s) LG 42" C2 4k OLED
Power Supply XPG Core Reactor 850W
Software I use Arch btw
You've posted this somewhere else haven't you? I've searched for such a tool and I am 99% sure I have seen this at one point.
Likely, I've posted this in many forums as an example of allocating ram without using it haha
 
Joined
Feb 3, 2017
Messages
3,759 (1.32/day)
Processor Ryzen 7800X3D
Motherboard ROG STRIX B650E-F GAMING WIFI
Memory 2x16GB G.Skill Flare X5 DDR5-6000 CL36 (F5-6000J3636F16GX2-FX5)
Video Card(s) INNO3D GeForce RTX™ 4070 Ti SUPER TWIN X2
Storage 2TB Samsung 980 PRO, 4TB WD Black SN850X
Display(s) 42" LG C2 OLED, 27" ASUS PG279Q
Case Thermaltake Core P5
Power Supply Fractal Design Ion+ Platinum 760W
Mouse Corsair Dark Core RGB Pro SE
Keyboard Corsair K100 RGB
VR HMD HTC Vive Cosmos
It's easily doable, but idk if there's such an app for windows.
See this video of mine I did years ago (I did it as an example of allocated vs actually used VRAM):
That's basically a virtual disk on VRAM, right?

Cool. I remember back during the GTX970 VRAM problems when trying to look into VRAM usage patterns and trying to force it into the 3.5-4GB range there were some solutions around to consume VRAM but they were not all too reliable. Simply allocating clearly did not work, filling the allocated range did help but there was some weirdness in that as well. Basically it seemed that user did not really get complete control over the actual VRAM management and drivers or firmware had their own say in what actually happened. Both on Nvidia and AMD cards. And yeah, worse on Windows of course (more handholding from what it looked like). Making a virtual drive out of it and copying actual files onto that does seem like a more reliable solution :)
 
Joined
May 10, 2023
Messages
273 (0.48/day)
Location
Brazil
Processor 5950x
Motherboard B550 ProArt
Cooling Fuma 2
Memory 4x32GB 3200MHz Corsair LPX
Video Card(s) 2x RTX 3090
Display(s) LG 42" C2 4k OLED
Power Supply XPG Core Reactor 850W
Software I use Arch btw
That's basically a virtual disk on VRAM, right?
Yup.
Basically it seemed that user did not really get complete control over the actual VRAM management and drivers or firmware had their own say in what actually happened. Both on Nvidia and AMD cards. And yeah, worse on Windows of course (more handholding from what it looked like). Making a virtual drive out of it and copying actual files onto that does seem like a more reliable solution :)
I think that even just copying files like I did would not really achieve what you want. Maybe manually doing some cudaMalloc for many regions, getting their addresses and freeing up the ones not in your desired range could work, but nothing stops the memory manager to just move those pages around.
 
Top