• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.

NVIDIA Could Ready HD 4670 Competitor

btarunr

Editor & Senior Moderator
Staff member
Joined
Oct 9, 2007
Messages
47,233 (7.55/day)
Location
Hyderabad, India
System Name RBMK-1000
Processor AMD Ryzen 7 5700G
Motherboard ASUS ROG Strix B450-E Gaming
Cooling DeepCool Gammax L240 V2
Memory 2x 8GB G.Skill Sniper X
Video Card(s) Palit GeForce RTX 2080 SUPER GameRock
Storage Western Digital Black NVMe 512GB
Display(s) BenQ 1440p 60 Hz 27-inch
Case Corsair Carbide 100R
Audio Device(s) ASUS SupremeFX S1220A
Power Supply Cooler Master MWE Gold 650W
Mouse ASUS ROG Strix Impact
Keyboard Gamdias Hermes E2
Software Windows 11 Pro
GPU Café published information on future competition lineups., which shows the entry of a "GeForce 9550 GT" stacked up against the Radeon HD 4670. Sources in the media have pointed to the the possibility that the the RV730 based HD 4670 from ATI outperforms NVIDIA cards in its current lineup, relative to the segments where GeForce 9500 GT sits. The HD 4650 could exchange a few blows with the GeForce 9500 GT with equal or better levels of performance while the HD 4670 surpasses it.

The entry of a GeForce 9550 GT shows the 9500 GT cannot compete with the HD 4650, a newer price demographic of ~ $129 is shown in that chart that not only indicates prices, but also shows the HD 4650's lead over 9500 GT is so significant that ATI could be comfortable with asking you $20 more than what 9500 GT asks, relative to the range. GPU Café reports that the 9550 GT would be a toned-down (and shrunk) G94, as in the 55 nm G94b, featuring 64 shader processors and a 192-bit memory bus (and presumably, memory configurations such as 384 MB or 768 MB of GDDR3 memory).



View at TechPowerUp Main Site
 
Last edited:

Apocolypse007

New Member
Joined
Apr 12, 2006
Messages
614 (0.09/day)
Location
New Sewickley, PA
System Name six pack
Processor Phenom II X6 1090T BE
Motherboard ASRock 890FX Deluxe3
Cooling Cool-it ECO ALC Water cooler
Memory 4GB G-Skill Flares CL7
Video Card(s) 1GB HIS HD4890 930core / 1025mem 4100mhz effective
Storage 320Gb Seagate Barracuda (w/perpendiular recording) and 750GB Western Digital Caviar Black
Display(s) Acer B273HU 2048 x 1152 native resolution
Case Xoxide X-Sentric Professional Series w/250mm Fan
Audio Device(s) SB Audigy 2 Value
Power Supply 750 Watt Apevia Warlock
Software Windows Vista Home Premium 64bit (w/dreamscene patch)/Windows 7 Pro 64bit
Benchmark Scores 3DMark06:10,516 3Dmark Vantage: P8537
$20 less for a slightly less powerful GPU. Honestly I think that if you will spend >100 on a GPU to begin with, the extra $20 for the 9600GT is nothing. I don't think there is much of a market for so many different versions of today's cards. There should only be a mainstream, performance, and enthusiast class product for each new generation of card. Any more products and its just going to confuse the consumer.
 

Black Light

New Member
Joined
Oct 2, 2007
Messages
156 (0.02/day)
System Name WeaponX2
Processor AMD64 7750 KUMA
Motherboard M2N-SLI DELUXE
Cooling Artic Cooling64
Memory 2X 1GB 4-4-4-10 Corasir
Video Card(s) Asus 4870 @ 870/1150
Storage 2x WD 640GB
Display(s) 19" Samsung Syncmaster
Case Black Widow
Audio Device(s) onboard
Power Supply 550w Rosewill
Software XP PRO SP2
$20 less for a slightly less powerful GPU. Honestly I think that if you will spend >100 on a GPU to begin with, the extra $20 for the 9600GT is nothing. I don't think there is much of a market for so many different versions of today's cards. There should only be a mainstream, performance, and enthusiast class product for each new generation of card. Any more products and its just going to confuse the consumer.

true but if it stays the was it is it is $40 from the 9500GT--->9600GT
 

newtekie1

Semi-Retired Folder
Joined
Nov 22, 2005
Messages
28,473 (4.10/day)
Location
Indiana, USA
Processor Intel Core i7 10850K@5.2GHz
Motherboard AsRock Z470 Taichi
Cooling Corsair H115i Pro w/ Noctua NF-A14 Fans
Memory 32GB DDR4-3600
Video Card(s) RTX 2070 Super
Storage 500GB SX8200 Pro + 8TB with 1TB SSD Cache
Display(s) Acer Nitro VG280K 4K 28"
Case Fractal Design Define S
Audio Device(s) Onboard is good enough for me
Power Supply eVGA SuperNOVA 1000w G3
Software Windows 10 Pro x64
I'm guessing the 9550GT is going to be the 55nm part with clocks slightly increased. The 9500GT should be able to handle the 4650 and the Pre-Overclocked 9500GT's should be able to handle the 4670.

I think nVidia is just going overboard with adding the 9550GT. They should have just left the 9500GT for $20 cheaper and let their partners Pre-Overclock the cards to make up the difference in performance and price.
 
Joined
Dec 28, 2006
Messages
4,378 (0.67/day)
Location
Hurst, Texas
System Name The86
Processor Ryzen 5 3600
Motherboard ASROCKS B450 Steel Legend
Cooling AMD Stealth
Memory 2x8gb DDR4 3200 Corsair
Video Card(s) EVGA RTX 3060 Ti
Storage WD Black 512gb, WD Blue 1TB
Display(s) AOC 24in
Case Raidmax Alpha Prime
Power Supply 700W Thermaltake Smart
Mouse Logitech Mx510
Keyboard Razer BlackWidow 2012
Software Windows 10 Professional
the 9600GSO is an even 100 these days, your stupid to get one of these with the 9600GSO out there
 

newtekie1

Semi-Retired Folder
Joined
Nov 22, 2005
Messages
28,473 (4.10/day)
Location
Indiana, USA
Processor Intel Core i7 10850K@5.2GHz
Motherboard AsRock Z470 Taichi
Cooling Corsair H115i Pro w/ Noctua NF-A14 Fans
Memory 32GB DDR4-3600
Video Card(s) RTX 2070 Super
Storage 500GB SX8200 Pro + 8TB with 1TB SSD Cache
Display(s) Acer Nitro VG280K 4K 28"
Case Fractal Design Define S
Audio Device(s) Onboard is good enough for me
Power Supply eVGA SuperNOVA 1000w G3
Software Windows 10 Pro x64
the 9600GSO is an even 100 these days, your stupid to get one of these with the 9600GSO out there

You have to kind of ignore the prices in the picture, the 9500GT doesn't retail for $109, so the 9550GT won't be $129. I wonder if these prices are even USD? A 9500GT is more around $60 for the DDR2 version and $80($75 with Mail-In) for the DDR3 version.

I am in agreement with you though, the 9600GSO can be had for $90 even with free shipping right now from newegg. So, IMO, these lower class cards aren't worth saving the $10-20. The 9600GSO is even cheaper if you consider rebates, they can be had for $80.
 
Joined
Dec 28, 2006
Messages
4,378 (0.67/day)
Location
Hurst, Texas
System Name The86
Processor Ryzen 5 3600
Motherboard ASROCKS B450 Steel Legend
Cooling AMD Stealth
Memory 2x8gb DDR4 3200 Corsair
Video Card(s) EVGA RTX 3060 Ti
Storage WD Black 512gb, WD Blue 1TB
Display(s) AOC 24in
Case Raidmax Alpha Prime
Power Supply 700W Thermaltake Smart
Mouse Logitech Mx510
Keyboard Razer BlackWidow 2012
Software Windows 10 Professional
agreed, I'd love to know when we started getting 2 low end series cards. The 9500GT should be set lower MSRP, and the 9400GT abdononed
 

newtekie1

Semi-Retired Folder
Joined
Nov 22, 2005
Messages
28,473 (4.10/day)
Location
Indiana, USA
Processor Intel Core i7 10850K@5.2GHz
Motherboard AsRock Z470 Taichi
Cooling Corsair H115i Pro w/ Noctua NF-A14 Fans
Memory 32GB DDR4-3600
Video Card(s) RTX 2070 Super
Storage 500GB SX8200 Pro + 8TB with 1TB SSD Cache
Display(s) Acer Nitro VG280K 4K 28"
Case Fractal Design Define S
Audio Device(s) Onboard is good enough for me
Power Supply eVGA SuperNOVA 1000w G3
Software Windows 10 Pro x64
There has always been 4 basic levels. The extreme low end, low end, mid-range, high end.

With the 8 Series:
8400, 8500, 8600, 8800
With the 7 series:
7100/7200, 7300, 7600, 7800/7900
With the 6 series:
6200-TC, 6200, 6600, 6800
With the 5 series:
5200, 5500, 5600/5700, 5800/5900

Though, in todays market, I don't see a place for the extreme low end anymore.
 
Joined
Dec 28, 2006
Messages
4,378 (0.67/day)
Location
Hurst, Texas
System Name The86
Processor Ryzen 5 3600
Motherboard ASROCKS B450 Steel Legend
Cooling AMD Stealth
Memory 2x8gb DDR4 3200 Corsair
Video Card(s) EVGA RTX 3060 Ti
Storage WD Black 512gb, WD Blue 1TB
Display(s) AOC 24in
Case Raidmax Alpha Prime
Power Supply 700W Thermaltake Smart
Mouse Logitech Mx510
Keyboard Razer BlackWidow 2012
Software Windows 10 Professional
There has always been 4 basic levels. The extreme low end, low end, mid-range, high end.

With the 8 Series:
8400, 8500, 8600, 8800
With the 7 series:
7100/7200, 7300, 7600, 7800/7900
With the 6 series:
6200-TC, 6200, 6600, 6800
With the 5 series:
5200, 5500, 5600/5700, 5800/5900

Though, in todays market, I don't see a place for the extreme low end anymore.


yes and now actully.

For Starters the 5500 came out after the 5200 to replace the 5200 Ultra that was more expensive to produce.

The 6200TC is the same low end generation as the normal 6200, the normal PCIe 6200 was just so they had something there, the 6200TC replaced it

The 7100/7200 line granted where lower end, though the 7100GS was faster than the 7200GS, and the 7200GS was just to get rid of NV44 cores. Though it started here.

Personally I want it simple again.

Geforce MX for low end

Geforce TI for high end.

GeforceMX 220, GeforceMX 240, GeforceMX 260

GeforceTI 220, GeforceTI 240, GeforceTI 260,

that would simpliy life enough for me
 
Joined
Mar 1, 2008
Messages
284 (0.05/day)
Location
Antwerp, Belgium
I'm guessing the 9550GT is going to be the 55nm part with clocks slightly increased. The 9500GT should be able to handle the 4650 and the Pre-Overclocked 9500GT's should be able to handle the 4670.

I think nVidia is just going overboard with adding the 9550GT. They should have just left the 9500GT for $20 cheaper and let their partners Pre-Overclock the cards to make up the difference in performance and price.

I really don't think so.
The 9500 is basically a higher clocked 8600.
The HD4600 is basically a HD3870 with a 128bit bus (+faster AA unit).
Considering that these cards are mainly used by users who own 19" monitors (~1280x1024), the low memory bandwidth won't be a major criteria.
For refrence, a HD3850 is around 2x faster than a 8600GTS @ 1280x1024.
The HD4670 will have 480GFlops (peak) and 9500GT has around 132GFlops (peak - depending on the model). You can't close that gap with an overclock.
 

newtekie1

Semi-Retired Folder
Joined
Nov 22, 2005
Messages
28,473 (4.10/day)
Location
Indiana, USA
Processor Intel Core i7 10850K@5.2GHz
Motherboard AsRock Z470 Taichi
Cooling Corsair H115i Pro w/ Noctua NF-A14 Fans
Memory 32GB DDR4-3600
Video Card(s) RTX 2070 Super
Storage 500GB SX8200 Pro + 8TB with 1TB SSD Cache
Display(s) Acer Nitro VG280K 4K 28"
Case Fractal Design Define S
Audio Device(s) Onboard is good enough for me
Power Supply eVGA SuperNOVA 1000w G3
Software Windows 10 Pro x64
I really don't think so.
The 9500 is basically a higher clocked 8600.
The HD4600 is basically a HD3870 with a 128bit bus (+faster AA unit).
Considering that these cards are mainly used by users who own 19" monitors (~1280x1024), the low memory bandwidth won't be a major criteria.
For refrence, a HD3850 is around 2x faster than a 8600GTS @ 1280x1024.
The HD4670 will have 480GFlops (peak) and 9500GT has around 132GFlops (peak - depending on the model). You can't close that gap with an overclock.

You are wrong, the 128-bit bus makes a huge performance hit.

The HD4670 is just an overclocked HD4650. All the information we have seen says the 9500GT matches the HD4650, so an overclocked 9500GT should be able to match an HD4670.

And the FLOPS rating of either card doesn't matter one bit, and has no real affect on graphical performance. If it did, we wouldn't see the 9600GT rated at 208 GFLOPS outperforming the HD3870 rated at 496 GFLOPS.
 
Joined
Feb 18, 2006
Messages
5,147 (0.75/day)
Location
AZ
System Name Thought I'd be done with this by now
Processor i7 11700k 8/16
Motherboard MSI Z590 Pro Wifi
Cooling Be Quiet Dark Rock Pro 4, 9x aigo AR12
Memory 32GB GSkill TridentZ Neo DDR4-4000 CL18-22-22-42
Video Card(s) MSI Ventus 2x Geforce RTX 3070
Storage 1TB MX300 M.2 OS + Games, + cloud mostly
Display(s) Samsung 40" 4k (TV)
Case Lian Li PC-011 Dynamic EVO Black
Audio Device(s) onboard HD -> Yamaha 5.1
Power Supply EVGA 850 GQ
Mouse Logitech wireless
Keyboard same
VR HMD nah
Software Windows 10
Benchmark Scores no one cares anymore lols
You are wrong, the 128-bit bus makes a huge performance hit.

The HD4670 is just an overclocked HD4650. All the information we have seen says the 9500GT matches the HD4650, so an overclocked 9500GT should be able to match an HD4670.

And the FLOPS rating of either card doesn't matter one bit, and has no real affect on graphical performance. If it did, we wouldn't see the 9600GT rated at 208 GFLOPS outperforming the HD3870 rated at 496 GFLOPS.

yeah and I thought the 4670 has 12rop's which would further lower performance. or am I wrong on that?
 

WarEagleAU

Bird of Prey
Joined
Jul 9, 2006
Messages
10,812 (1.61/day)
Location
Gurley, AL
System Name Pandemic 2020
Processor AMD Ryzen 5 "Gen 2" 2600X
Motherboard AsRock X470 Killer Promontory
Cooling CoolerMaster 240 RGB Master Cooler (Newegg Eggxpert)
Memory 32 GB Geil EVO Portenza DDR4 3200 MHz
Video Card(s) ASUS Radeon RX 580 DirectX 12 DUAL-RX580-O8G 8GB 256-Bit GDDR5 HDCP Ready CrossFireX Support Video C
Storage WD 250 M.2, Corsair P500 M.2, OCZ Trion 500, WD Black 1TB, Assorted others.
Display(s) ASUS MG24UQ Gaming Monitor - 23.6" 4K UHD (3840x2160) , IPS, Adaptive Sync, DisplayWidget
Case Fractal Define R6 C
Audio Device(s) Realtek 5.1 Onboard
Power Supply Corsair RMX 850 Platinum PSU (Newegg Eggxpert)
Mouse Razer Death Adder
Keyboard Corsair K95 Mechanical & Corsair K65 Wired, Wireless, Bluetooth)
Software Windows 10 Pro x64
I dont think you are wrong. I also dont think the 9500GT is even par with the HD 4650. But good luck to me trying to convince you of that.
 
Joined
Dec 28, 2006
Messages
4,378 (0.67/day)
Location
Hurst, Texas
System Name The86
Processor Ryzen 5 3600
Motherboard ASROCKS B450 Steel Legend
Cooling AMD Stealth
Memory 2x8gb DDR4 3200 Corsair
Video Card(s) EVGA RTX 3060 Ti
Storage WD Black 512gb, WD Blue 1TB
Display(s) AOC 24in
Case Raidmax Alpha Prime
Power Supply 700W Thermaltake Smart
Mouse Logitech Mx510
Keyboard Razer BlackWidow 2012
Software Windows 10 Professional
320x16x8 is the RV730 core config, thats 8 rops, and 16 TMU's with 64x5 ALU's which only 1 set will really be used in games of course.

So 64x750 = 48,000

now the 9500GT 32x16x8 core config, now the shader config is 32 and only 32 but all those will be used, unlike the extra ALU's on the 4650.

so 32x1400 = 44,800

numbers are fairly close on shader op's per second for most games actully.

so you tell me can the 9500GT keep up?

what ATI needs is a 9600GT killer the RV670 is supposed to stop production soon leaving nothing to compete
 
Joined
Mar 1, 2008
Messages
284 (0.05/day)
Location
Antwerp, Belgium
320x16x8 is the RV730 core config, thats 8 rops, and 16 TMU's with 64x5 ALU's which only 1 set will really be used in games of course.

So 64x750 = 48,000

now the 9500GT 32x16x8 core config, now the shader config is 32 and only 32 but all those will be used, unlike the extra ALU's on the 4650.

so 32x1400 = 44,800

numbers are fairly close on shader op's per second for most games actully.

so you tell me can the 9500GT keep up?

what ATI needs is a 9600GT killer the RV670 is supposed to stop production soon leaving nothing to compete

What you write here is pure nonsence. You need to study more into this matter before making such a statement.
ATI and nVidia use a runtime compiler and this compiler tries to make best use of the shaders available. I don't think there is any situation where the compiler is that inefficient.

@newtekie1 about the 9600GT vs HD3870:
While GFlops are NOT the only factor that make a chip perform in a certain way, they are for sure very important. It just crazy to say they don't matter one bit.
Only using the 9600GT vs HD3870 as refrence and concluding that is wrong.
The problem lies in the fact that the HD3870 has very high shader power while the other units are not that powerful. That's why you get a skewed view when using 'older' games.
To give an example, check out these numbers from Crysis - 'very high setting' (extremely shader heavy):
8600GTS - 4.3
HD3650 - 6.4
9600GT - 14.9
HD3870 - 16.1
9800GTX - 21.9

You can immediately see that the HD3870 is faster that the 9600GT but even more important is the fact that the 9800GTX is 47% faster. GFlops don't matter? They matter now and even more in the future.
 

newtekie1

Semi-Retired Folder
Joined
Nov 22, 2005
Messages
28,473 (4.10/day)
Location
Indiana, USA
Processor Intel Core i7 10850K@5.2GHz
Motherboard AsRock Z470 Taichi
Cooling Corsair H115i Pro w/ Noctua NF-A14 Fans
Memory 32GB DDR4-3600
Video Card(s) RTX 2070 Super
Storage 500GB SX8200 Pro + 8TB with 1TB SSD Cache
Display(s) Acer Nitro VG280K 4K 28"
Case Fractal Design Define S
Audio Device(s) Onboard is good enough for me
Power Supply eVGA SuperNOVA 1000w G3
Software Windows 10 Pro x64
What you write here is pure nonsence. You need to study more into this matter before making such a statement.
ATI and nVidia use a runtime compiler and this compiler tries to make best use of the shaders available. I don't think there is any situation where the compiler is that inefficient.

@newtekie1 about the 9600GT vs HD3870:
While GFlops are NOT the only factor that make a chip perform in a certain way, they are for sure very important. It just crazy to say they don't matter one bit.
Only using the 9600GT vs HD3870 as refrence and concluding that is wrong.
The problem lies in the fact that the HD3870 has very high shader power while the other units are not that powerful. That's why you get a skewed view when using 'older' games.
To give an example, check out these numbers from Crysis - 'very high setting' (extremely shader heavy):
8600GTS - 4.3
HD3650 - 6.4
9600GT - 14.9
HD3870 - 16.1
9800GTX - 21.9

You can immediately see that the HD3870 is faster that the 9600GT but even more important is the fact that the 9800GTX is 47% faster. GFlops don't matter? They matter now and even more in the future.

Using your numbers, and assuming GFLOPS matter, why doesn't the HD3870 outperform the 9800GTX? Though I don't know where you are even getting your numbers.

http://www.techpowerup.com/reviews/Galaxy/GeForce_9500_GT_Overclocked/9.html

The HD3650 doesn't outperform the 8600GTS in Crysis despite the nearly 100 GFLOP advantage the HD3650 has. Face it, GFLOPS can't be used to determin gaming performance.

I dont think you are wrong. I also dont think the 9500GT is even par with the HD 4650. But good luck to me trying to convince you of that.

We will have to wait until the HD4650 is released and see. However, judging by the performance of the HD3650, which is about 60% of the 9500GT, and the fact that the HD4650 appears to be the HD3650 with everything on the core double, I think the two will be very close in the end.
 
Last edited:
Joined
Mar 1, 2008
Messages
284 (0.05/day)
Location
Antwerp, Belgium
http://www.computerbase.de/artikel/...st_ati_radeon_hd_4870_x2/20/#abschnitt_crysis

Well i also should have mentioned that since ATI and nVidia use completely different architectures, it's hard to compare their GFlops. But within one brand it's easy to see that GFlops do matter and that's why i was pointing to the 9600GT and 9800GTX comparison.

Your techpowerup review of Crysis has one flaw:
We tested the DX9 version with graphics set to "High", which is the highest non-DX10 setting in the game.

ComputerBase uses DX10 and 'Very High'. This setting is much more shader demanding!
BTW in this same review the 9500GT scores 7.0fps and that's only 9% more than a HD3650.

If you go here:
http://www.computerbase.de/artikel/...on_hd_4870_x2/23/#abschnitt_performancerating => these are the results of all games combined.
Here you see that the 9500GT scores only 19% more on average than a HD3650.

Enough talking and let's just wait a month.
 

DarkMatter

New Member
Joined
Oct 5, 2007
Messages
1,714 (0.27/day)
Processor Intel C2Q Q6600 @ Stock (for now)
Motherboard Asus P5Q-E
Cooling Proc: Scythe Mine, Graphics: Zalman VF900 Cu
Memory 4 GB (2x2GB) DDR2 Corsair Dominator 1066Mhz 5-5-5-15
Video Card(s) GigaByte 8800GT Stock Clocks: 700Mhz Core, 1700 Shader, 1940 Memory
Storage 74 GB WD Raptor 10000rpm, 2x250 GB Seagate Raid 0
Display(s) HP p1130, 21" Trinitron
Case Antec p180
Audio Device(s) Creative X-Fi PLatinum
Power Supply 700W FSP Group 85% Efficiency
Software Windows XP
MrMilli your point has a big flaw: who cares how this cards will perform in future games? As you have said in the future the HD card could perform better because the demand for shader power will be higher. Again who cares? It's not even able to use the higher settings of the most demanding games today, let alone in the future ones...

If the card has some shader power left now (assuming that is true, which I don't think) then the card is bottlenecked by the other parts. That will not change in the future and that only means that while the 9500GT will go down to 5fps from the 10fps that renders today, the HD card will mantain a framerate close to that 10. WOOhhooo! Big deal. Same happens with the X1000 family, now they are like 50%++ faster than GF7 counterparts but always on higher settings and thus unplayable frames.

I have said this like hundreds of times: ever since the X1000 series Ati seems more concerned about how the cards could perform in the future than making the better card they can for the present.
 
Joined
Dec 28, 2006
Messages
4,378 (0.67/day)
Location
Hurst, Texas
System Name The86
Processor Ryzen 5 3600
Motherboard ASROCKS B450 Steel Legend
Cooling AMD Stealth
Memory 2x8gb DDR4 3200 Corsair
Video Card(s) EVGA RTX 3060 Ti
Storage WD Black 512gb, WD Blue 1TB
Display(s) AOC 24in
Case Raidmax Alpha Prime
Power Supply 700W Thermaltake Smart
Mouse Logitech Mx510
Keyboard Razer BlackWidow 2012
Software Windows 10 Professional
What you write here is pure nonsence. You need to study more into this matter before making such a statement.
ATI and nVidia use a runtime compiler and this compiler tries to make best use of the shaders available. I don't think there is any situation where the compiler is that inefficient.

@newtekie1 about the 9600GT vs HD3870:
While GFlops are NOT the only factor that make a chip perform in a certain way, they are for sure very important. It just crazy to say they don't matter one bit.
Only using the 9600GT vs HD3870 as refrence and concluding that is wrong.
The problem lies in the fact that the HD3870 has very high shader power while the other units are not that powerful. That's why you get a skewed view when using 'older' games.
To give an example, check out these numbers from Crysis - 'very high setting' (extremely shader heavy):
8600GTS - 4.3
HD3650 - 6.4
9600GT - 14.9
HD3870 - 16.1
9800GTX - 21.9

You can immediately see that the HD3870 is faster that the 9600GT but even more important is the fact that the 9800GTX is 47% faster. GFlops don't matter? They matter now and even more in the future.

then explain to me one thing please, why does the R700 and RV770 preform more like a 64sp card than a 320sp card. The reason is only one of those ALU's is a complex shader, 64 of those are simple the rest arn't even related to shader work actully. And very few games use simple shaders because its harder to program for 2 types of shaders than just one.
 
Joined
Mar 1, 2008
Messages
284 (0.05/day)
Location
Antwerp, Belgium
then explain to me one thing please, why does the R700 and RV770 preform more like a 64sp card than a 320sp card. The reason is only one of those ALU's is a complex shader, 64 of those are simple the rest arn't even related to shader work actully. And very few games use simple shaders because its harder to program for 2 types of shaders than just one.

The R600 uses 64 superscalar unified shader clusters, each consisting of 5 stream processing units for a total of 320 stream processing units. Each of the first 4 stream processing units is able to retire a finished single precision floating point MAD (or ADD or MUL) instruction per clock, dot product (dp, and special cased by combining ALUs), and integer ADD. The fifth unit is more complex and can additionally handle special transcendental functions such as sine and cosine. Each of the 64 shader clusters can execute 6 instructions per clock cycle (peak), consisting of 5 shading instructions plus 1 branch.

The claimed theoretical processing power for the 8 Series cards given in FLOPS may not be correct at all times. For example the GeForce 8800 GTX has 518.43 GigaFLOPs theoretical performance given the fact that there are 128 stream processors at 1.35 GHz with each SP being able to run 1 Multiply-Add and 1 Multiply instruction per clock [(MADD (2 FLOPs) + MUL (1 FLOP))×1350MHz×128 SPs = 518.4 GigaFLOPs]. This figure may not be correct because the Multiply operation is not always available giving a possibly more accurate performance figure of (2×1350×128) = 345.6 GigaFLOPs.

So ... just to recap for you:
ATI: 5 units can do MADD (or ADD or MUL)
The 5th (and complex) unit is a special unit. It can also do transcedentals like SIN, COS, LOG, EXP. That's it.
1 MADD (=Multiply-Add) = 2 Flops
1 ADD or MUL = 1 Flops
And these are all usable. The developer doesn't need to program this. The compiler takes care of this. A real life scenario with some bad code could be something like 2 MADD + 1 MUL. If we average this over the 64 units then that would give 240GFlops.

nVidia: basically each scalar unit can do 2 Flops per clock. That would result in a real life performance of around 90GFlops.

So on shader performance ATI will win hands down.

Considering how close the HD4870 performs to the GTX 280 and how much more texel fillrate and bandwidth the GTX has, then it seems to me that shader performance is darn important these days.
 
Last edited:

Frick

Fishfaced Nincompoop
Joined
Feb 27, 2006
Messages
19,565 (2.86/day)
Location
Piteå
System Name White DJ in Detroit
Processor Ryzen 5 5600
Motherboard Asrock B450M-HDV
Cooling Be Quiet! Pure Rock 2
Memory 2 x 16GB Kingston Fury 3400mhz
Video Card(s) XFX 6950XT Speedster MERC 319
Storage Kingston A400 240GB | WD Black SN750 2TB |WD Blue 1TB x 2 | Toshiba P300 2TB | Seagate Expansion 8TB
Display(s) Samsung U32J590U 4K + BenQ GL2450HT 1080p
Case Fractal Design Define R4
Audio Device(s) Plantronics 5220, Nektar SE61 keyboard
Power Supply Corsair RM850x v3
Mouse Logitech G602
Keyboard Cherry MX Board 1.0 TKL Brown
Software Windows 10 Pro
Benchmark Scores Rimworld 4K ready!
Though, in todays market, I don't see a place for the extreme low end anymore.

Until they make good integrated graphics (780G/HD3200 is kinde nice though...) there will always be a need for extreme low-end imo. Like the HD3450. It's not really a bad card for a HTPC and it's cheaper than a gaming mouse.
 

DarkMatter

New Member
Joined
Oct 5, 2007
Messages
1,714 (0.27/day)
Processor Intel C2Q Q6600 @ Stock (for now)
Motherboard Asus P5Q-E
Cooling Proc: Scythe Mine, Graphics: Zalman VF900 Cu
Memory 4 GB (2x2GB) DDR2 Corsair Dominator 1066Mhz 5-5-5-15
Video Card(s) GigaByte 8800GT Stock Clocks: 700Mhz Core, 1700 Shader, 1940 Memory
Storage 74 GB WD Raptor 10000rpm, 2x250 GB Seagate Raid 0
Display(s) HP p1130, 21" Trinitron
Case Antec p180
Audio Device(s) Creative X-Fi PLatinum
Power Supply 700W FSP Group 85% Efficiency
Software Windows XP
So ... just to recap for you:
ATI: 5 units can do MADD (or ADD or MUL)
The 5th (and complex) unit is a special unit. It can also do transcedentals like SIN, COS, LOG, EXP. That's it.
1 MADD (=Multiply-Add) = 2 Flops
1 ADD or MUL = 1 Flops
And these are all usable. The developer doesn't need to program this. The compiler takes care of this. A real life scenario with some bad code could be something like 2 MADD + 1 MUL. If we average this over the 64 units then that would give 240GFlops.

nVidia: basically each scalar unit can do 2 Flops per clock. That would result in a real life performance of around 90GFlops.

So on shader performance ATI will win hands down.

Considering how close the HD4870 performs to the GTX 280 and how much more texel fillrate and bandwidth the GTX has, then it seems to me that shader performance is darn important these days.

Some inacuracies and missinformation there:

- Theoretically both Ati and Nvidia shaders can do MADD+MUL. What you quoted above was about the G80, it has been long fixed in later releases. Assuming Ati can do both at a ime, while Nvidia can't, is stupid consideing Ati doesn't outperform Nvidia by so much even on shader specific benchmarks...
-You so conveniently forgot Nvidia shaders run at doeble the speed when calculating the "real life" performance...
- R600 and R700 are SIMD for each cluster and VLIW for each shader. This means that the instruction for all 5 units in the shader have to be written at the same time in the compilation (Very Long Instruction Word) and that all 80 shaders (R600=80x4, R700=80x10) in each cluster must calculate the same instruction. By constrast Nvidia's are scalar and also organiced on SIMD arrays, but only 16 or 24 long. (G80/9x and GT200 respectively)

This has two effects:

1. VLIW means that even if shaders (5 ALUs) are superscalar for the programmer or the drivers in this case, each shader IS a vector unit.
2. SIMD over such large arrays means that if a state change occurs, you have to calculate it in a different cluster, potentially losing a complete cluster or even the entire chip in tha clock.

That's why Ati is comparable to Nvidia when it comes to "real life" shader power.
 

btarunr

Editor & Senior Moderator
Staff member
Joined
Oct 9, 2007
Messages
47,233 (7.55/day)
Location
Hyderabad, India
System Name RBMK-1000
Processor AMD Ryzen 7 5700G
Motherboard ASUS ROG Strix B450-E Gaming
Cooling DeepCool Gammax L240 V2
Memory 2x 8GB G.Skill Sniper X
Video Card(s) Palit GeForce RTX 2080 SUPER GameRock
Storage Western Digital Black NVMe 512GB
Display(s) BenQ 1440p 60 Hz 27-inch
Case Corsair Carbide 100R
Audio Device(s) ASUS SupremeFX S1220A
Power Supply Cooler Master MWE Gold 650W
Mouse ASUS ROG Strix Impact
Keyboard Gamdias Hermes E2
Software Windows 11 Pro
Sorry for the inaccuracy. Fixed.
 
Joined
Mar 1, 2008
Messages
284 (0.05/day)
Location
Antwerp, Belgium
@Darkmatter

- Only GT200 can dual-issue MADD and MUL ops all the time. G8x/G9x generation chips can't do it all the time. There are a select few scenarios where you can dual-issue MAD and MUL ops.

- I didn't: 1375Mhz * 2 Flops * 32 shaders = 88 GFlops

- You are wrong about it being SIMD. ATI's shader involves a MIMD 5-way vectoriel unit, MIMD signifying (contrary to SIMD) that several different instructions can be processed in parallel. The compiler is going to try to assemble simple operations in order to fill the MIMD 5D unit. But these 5 instructions cannot be dependant on each other. So even one shader can process different instructions at a time, let alone one cluster!
I simulated that only 3 instructions/shader can be done on average in my real life calculation because of less than optimal code and inefficiencies.
So basically your conclusion is wrong!

Using my real life caculation (it's just a simulation):
9800GTX 432GFlops
HD3870 248
HD4670 240
9600GT 208
9500GT 88
HD3650 87

If you check out my Crysis scores i posted previously, things start to make sence.
Now i know the HD4670 won't beat the 9600GT in Crysis because of many factors but what ATI has done is basically slapped the HD3870 shader engine into it. Add the RV700 generation architectural improvements.
nVidia on the contrary has made a die shrink of G84 and clocked it higher.

(pls read my previous posts before you reply)
 
Top