• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.
  • The forums have been upgraded with support for dark mode. By default it will follow the setting on your system/browser. You may override it by scrolling to the end of the page and clicking the gears icon.

AMD "Ellesmere" ASIC Pictured Up Close in RX 480 PCB Picture Leak

I said nothing at all about performance in my comments so it seems you are missing the point.

1. Go back and look at the post i was replying to, that post was talking about how amd has similar or even better oc headroom that nvidia, which is true because amd is clocking this thing very conservatively at stock, so i simply said this round nvidia got mad clocks and i wonder if it has anything to do with 14nm vs 16nm. Does that make more sence now? How about you guys contribute to a logical argument rather than being all offended. And secondly, any sane person will not compare gpu to cpu, in cpu single threaded performance is important so clock as well as ipc go hand in hand, then after than comes number of cores, gpus on the other hand arent that way since its all about fitting as many cores and clocking them as high as you possibly can before reaching thermal headroom. So while clock is not everything, it still works differently than with cpus

You missed the entire point of my post. Clockspeed doesn't matter. GPU's do not have cores you are correct, but do you know what an SP is, do you know how they work? Did you know they stuff as many of those into a GPU as is possible and those basically make the GPU. Something about number of transistors and those actually do the work.

So instead of arguing with the entire forum, all of whom has said you are wrong how about do some research.

Oh and just in case you missed it clockspeed doesn't matter. The GTX980 hit 2.3ghz in 2015, it also wasn't the first GPU to do so. http://www.kitguru.net/components/g...eforce-gtx-980-hits-2-30ghz-sets-new-records/
 
"Naked" 480, not sure why it is a thing, but here we go:

wkumr.jpg



@cdawall
Why do you think it would be "Fury replacement"? (at stock)
I could see 300$ (1500+) version going there, but don't 1266Mhz leaks show it somewhere at 980 (stock), and even slightly below it?
 
"Naked" 480, not sure why it is a thing, but here we go:

wkumr.jpg



@cdawall
Why do you think it would be "Fury replacement"? (at stock)
I could see 300$ (1500+) version going there, but don't 1266Mhz leaks show it somewhere at 980 (stock), and even slightly below it?
36 CU's says it all.

This card is an R9 390 replacement at a better price.

Vega will bring the Fury replacement cards. Only these will be called the RX 490/x this time.

I don't expect much better performance than the R9 390 if both cards are clocked the same in DX11 titles. This card will sit firmly in the R9 390x performance range but it's clockspeeds will help it pull away in some games.
 
@cdawall
Why do you think it would be "Fury replacement"? (at stock)
I could see 300$ (1500+) version going there, but don't 1266Mhz leaks show it somewhere at 980 (stock), and even slightly below it?

When did I say it replaced a fury? It should compete with a nano on the top end, but they will both be power limited minus the high end ones I feel.
 
from everything i've seen its a 390(x) replacement at half the TDP.

odd/annoying that the cooler is longer than the PCB.
 
from everything i've seen its a 390(x) replacement at half the TDP.

odd/annoying that the cooler is longer than the PCB.

Fury is the exact same way though.
 
You missed the entire point of my post. Clockspeed doesn't matter. GPU's do not have cores you are correct, but do you know what an SP is, do you know how they work? Did you know they stuff as many of those into a GPU as is possible and those basically make the GPU. Something about number of transistors and those actually do the work.

So instead of arguing with the entire forum, all of whom has said you are wrong how about do some research.

Oh and just in case you missed it clockspeed doesn't matter. The GTX980 hit 2.3ghz in 2015, it also wasn't the first GPU to do so. http://www.kitguru.net/components/g...eforce-gtx-980-hits-2-30ghz-sets-new-records/
Lol ok I feel like you arent even reading my comments, or maybe you are slightly skipping through the lines. This time i will try to break things down as much as possible. There are 3 major factors in gpu production.
1- thermal headroom (dictates form factor)
2-die size (effects cost of production)
3- performance (number of cores/sps and clockspeed dictate this)

So in a gpu, engineers want to design power efficient sps to cram as many of them before reaching thermal headroom, but at the same time you want to clock them as high as you can as well in order to build reasonably priced/smaller chips that perform well. So there is always a trade off.

The question i presented here, is wether you believe clockspeed is a design decision by Amd, or is due to the variance between 14nm glofo and 16nm tsmc, id appreciate it if you stop refuting what you are assuming im trying to say(bash amd, which I'm not) and instead contribute to the conversation by answering my question.

Is amd prioratizing core count and bigger die with lower power over smaller chips with higher clocks and higher power?
And i also pointed out that I wonder if 14nm glofo having slightly higher density and possibly lower cost effected that decision from AMDs side.

Again sorry for the confusion, but you guys seriously didnt get what i was trying to say initially, but i hope you do now :)
 
Lol ok I feel like you arent even reading my comments, or maybe you are slightly skipping through the lines. This time i will try to break things down as much as possible. There are 3 major factors in gpu production.
1- thermal headroom (dictates form factor)
2-die size (effects cost of production)
3- performance (number of cores/sps and clockspeed dictate this)

So in a gpu, engineers want to design power efficient sps to cram as many of them before reaching thermal headroom, but at the same time you want to clock them as high as you can as well in order to build reasonably priced/smaller chips that perform well. So there is always a trade off.

The question i presented here, is wether you believe clockspeed is a design decision by Amd, or is due to the variance between 14nm glofo and 16nm tsmc, id appreciate it if you stop refuting what you are assuming im trying to say(bash amd, which I'm not) and instead contribute to the conversation by answering my question.

Is amd prioratizing core count and bigger die with lower power over smaller chips with higher clocks and higher power?
And i also pointed out that I wonder if 14nm glofo having slightly higher density and possibly lower cost effected that decision from AMDs side.

Again sorry for the confusion, but you guys seriously didnt get what i was trying to say initially, but i hope you do now :)

What you have said is wrong. Claiming efficiency purely based off of clockspeed is ignorant. This tangent about sp's has never been in any of your comments and your understanding of it makes me think you watched one too many YouTube videos. Go back to Tom's hardware.
 
sergio: you're comparing apples to oranges and arguing over which is better because it has more blue.

clockspeed alone means nothing, you cannot use it as a reference between different models of the same GPU, let alone different GPU designs.
 
What you have said is wrong. Claiming efficiency purely based off of clockspeed is ignorant. This tangent about sp's has never been in any of your comments and your understanding of it makes me think you watched one too many YouTube videos. Go back to Tom's hardware.
Because previously i used the generic term cores rather than sp's. i only pointed out clock speed and that nvidia went mad on the clocks, i assumed people here would get what i meant, instead it seems people including yourself are butthurt, and even after saying sorry for the misunderstanding and explaining in details what i was going for you still wanna go on and on about technicalities and what i said and what i didnt say rather than moving on and having a mature conversation.
 
Last edited:
And i also pointed out that I wonder if 14nm glofo having slightly higher density and possibly lower cost effected that decision from AMDs side.
Note that AMD crammed in more transistors into the same die area than Nvidia even when on the same node.
More things at lower clock or less at higher clocks is a design decision, with neither way being clearly "superior".

If RX 480 OCing rumors are true, AMD has clearly "first chips released" round, with 15-20%-ish boost from OC, vs 10-14%-ish from nVidia.
 
Note that AMD crammed in more transistors into the same die area than Nvidia even when on the same node.
More things at lower clock or less at higher clocks is a design decision, with neither way being clearly "superior".

If RX 480 OCing rumors are true, AMD has clearly "first chips released" round, with 15-20%-ish boost from OC, vs 10-14%-ish from nVidia.
Yeah that makes sense, and having more oc headroom is good and was clear from your original post, but idk if its just me but i still feel like amd is clocking these chips conservatively even with the density vs clocks thing you mentioned, probably to make validation possible on a higher number of chips, but on later revisions snd probably rebrands im sure the clock will be much higher and somewhat closer to the clocks nvidia is reaching. We saw this with the hd7970 when it was released at 925mhz, then a ghz edition at 1050mhz was released on the next revision. One thing i know for sure though, i cant wait for the official reviews! If my educated guesses and assumptions are in their place, i suspect polaris 10 at its mid/end of its lifecycle and rebrands will be clocked and perform similar or close to gtx1070, but initially with current clocks and somewhat early silicon it wont come as close
 
Because previously i used the generic term cores rather than sp's. i only pointed out clock speed and that nvidia went mad on the clocks, i assumed people here would get what i meant, instead it seems people including yourself are butthurt, and even after saying sorry for the misunderstanding and explaining in details what i was going for you still wanna go on and on about technicalities and what i said and what i didnt say rather than moving on and having a mature conversation.

No one is butt hurt everyone is scratching their collective heads trying to figure out why you think the ocean is full of soda.
 
Why are you so illogical? He said top spot. With no reference to anything else. Top spot. Not middle top. Bottom top. Top.

Sooooooooooooooooooo yeah.

Maybe Radeon Pro would be Top Spot but with occasional profile absences, the average would possibly drag it below GTX 1080.

top spot at populatiry, like GTX 970 is right now (which isn't the king of performance either... just most popular; best seller). http://store.steampowered.com/hwsurvey
 
I think there is some confusion about how GPUs work, so maybe a little bit of explanation is in order because everyone is acting like the number of shaders are the things that can do independent computations and this is not a true statement. Both nVidia and AMD have terms for what constitutes a full "core". For nVidia that's a "Streaming Multiprocessor" or SM and for AMD it's a "Compute Unit" or CU. Each CU/SM has all of the inner workings to do independent calculations however, it's usually these units that dispatch more fine-grained computations. So, for regular CPUs, it's just having cores inside of a package. For GPUs, it's like having very simple cores which are put inside more CPU-like complex cores, which are then put into the package.

Just like regular CPUs, there is a limit to how much parallel work can be accomplished on a GPU. There is an argument that could be made that despite similar shader counts, the fact that nVidia used fewer SMs clocked higher yielded a better benefit than more CUs clocked lower (also with fewer shaders per CU.)

So with that said, I think there is a lot more that goes into this than clock speed, or even hardware for that matter. How the drivers and applications dispatch any instructions to the GPU has just as much to do with it as the way the hardware is setup... but simply put, the only time you can boil everything down to clock speed is on the same architecture.

tl;dr: GPUs do have "cores", they're just a little different than traditional CPU cores and they're not the shaders though, they're the CUs and SMs.
 
"Naked" 480, not sure why it is a thing, but here we go:

wkumr.jpg
Just wondering, the picture shows the fins on the heatsink in a vertical alignment, but shouldn't they mount 90 degrees to that for air flow to have a clean path out the rear vents?
 
to me, based on the base of that heatsink it needs to be rotated 90 degrees to the right - then the fins are horizontal, and it looks like it all lines up.
 
to me, based on the base of that heatsink it needs to be rotated 90 degrees to the right - then the fins are horizontal, and it looks like it all lines up.
That's what I thought, but the mounting holes look like they wont line up that way.
Maybe that is an illusion.
 
I think there is some confusion about how GPUs work, so maybe a little bit of explanation is in order because everyone is acting like the number of shaders are the things that can do independent computations and this is not a true statement. Both nVidia and AMD have terms for what constitutes a full "core". For nVidia that's a "Streaming Multiprocessor" or SM and for AMD it's a "Compute Unit" or CU. Each CU/SM has all of the inner workings to do independent calculations however, it's usually these units that dispatch more fine-grained computations. So, for regular CPUs, it's just having cores inside of a package. For GPUs, it's like having very simple cores which are put inside more CPU-like complex cores, which are then put into the package.

Just like regular CPUs, there is a limit to how much parallel work can be accomplished on a GPU. There is an argument that could be made that despite similar shader counts, the fact that nVidia used fewer SMs clocked higher yielded a better benefit than more CUs clocked lower (also with fewer shaders per CU.)

So with that said, I think there is a lot more that goes into this than clock speed, or even hardware for that matter. How the drivers and applications dispatch any instructions to the GPU has just as much to do with it as the way the hardware is setup... but simply put, the only time you can boil everything down to clock speed is on the same architecture.

tl;dr: GPUs do have "cores", they're just a little different than traditional CPU cores and they're not the shaders though, they're the CUs and SMs.
Well said, i personally didnt intend to get into sucj depth though and just used the term "cores" whereas in a cpu it was a handful while in a gpu its multi parallelism, and what i was saying the relationships between clockspeed to performance isnt the same as in cpus because in a cpu single core performance is key. In a gpu as i mentioned the big constrain is thermal headroom and die size, and you want to keep a good balance hence why i had all these question marks and everyone lost it. I said nvidia went mad on clocks and they surely did, yes you cant compare clockspeed directly between different architectures but i mean this is an over 30% clock difference between amd and nvidia, and this is kinda on similar nodes, so idk if im missing something but i dont remember the clockspeed gap ever being this wide to be honest, correct me if im wrong tho. I did point out 14nm vs 16nm so technicaly this is the first time amd and nvidia use different processes, still i though they should be similar for the most part
 
Well said, i personally didnt intend to get into sucj depth though and just used the term "cores" whereas in a cpu it was a handful while in a gpu its multi parallelism, and what i was saying the relationships between clockspeed to performance isnt the same as in cpus because in a cpu single core performance is key. In a gpu as i mentioned the big constrain is thermal headroom and die size, and you want to keep a good balance hence why i had all these question marks and everyone lost it. I said nvidia went mad on clocks and they surely did, yes you cant compare clockspeed directly between different architectures but i mean this is an over 30% clock difference between amd and nvidia, and this is kinda on similar nodes, so idk if im missing something but i dont remember the clockspeed gap ever being this wide to be honest, correct me if im wrong tho. I did point out 14nm vs 16nm so technicaly this is the first time amd and nvidia use different processes, still i though they should be similar for the most part

They aren't similar in anyway. Just because two companies use 65nm or 14/16nm doesn't mean anything is the same internally....
 
They aren't similar in anyway. Just because two companies use 65nm or 14/16nm doesn't mean anything is the same internally....
They're a lot more similar than you think. Sure, there may be some implementation differences but when push comes to shove, they're doing it the same way. When push comes to shove, it really comes down to nVidia opting for lower latency and low parallel throughput compared to AMD which opts for higher overall latency but, more parallel throughput. The amount of performance increasing clocks for any given design will almost always result in performance improvements however, increasing the width of the engine (such as increasing the number of shaders/SMs/CUs,) has limited benefit depending on the workload and how that workload was dispatched.

Either way, all of this is besides the point. I don't think anyone really understands what @sergionography is trying to say and all I'm saying is this:
sergio: you're comparing apples to oranges and arguing over which is better because it has more blue.

clockspeed alone means nothing, you cannot use it as a reference between different models of the same GPU, let alone different GPU designs.
 
They're a lot more similar than you think. Sure, there may be some implementation differences but when push comes to shove, they're doing it the same way. When push comes to shove, it really comes down to nVidia opting for lower latency and low parallel throughput compared to AMD which opts for higher overall latency but, more parallel throughput. The amount of performance increasing clocks for any given design will almost always result in performance improvements however, increasing the width of the engine (such as increasing the number of shaders/SMs/CUs,) has limited benefit depending on the workload and how that workload was dispatched.

Either way, all of this is besides the point. I don't think anyone really understands what @sergionography is trying to say and all I'm saying is this:

Same argument could be made with AMD vs intel CPU's...AMD uses a different everything for everything.

Oh well at the end of the day I just got a 980Ti for my PC at work so I am going to enjoy that

per.png


Lol because I so need a 980Ti to run web browsers.
 
The amount of performance increasing clocks for any given design will almost always result in performance improvements however, increasing the width of the engine (such as increasing the number of shaders/SMs/CUs,) has limited benefit depending on the workload and how that workload was dispatched.
You don't need that "almost" at all, unless you mean increasing the clocks beyond stability ... increasing parallelism in hardware must be harnessed by equally parallelized software, luckily all graphics processing is very SIMD parallelized by definition (same shader code being run for many pixels and vertices) ... so higher clocks and more throughput are both valid strategies and both are used - so far AMD and Nvidia are finding a sweet spot power efficiency wise for what they think gaming workload will consist of in next year or so (balancing fill rate vs. compute/shading and raw throughput vs. caching in their architectures)
 
I bet somebody will overclock/overvolt RX480 to near 50%, that will par with 1070 easily. Looking at the performance graph and clock speed, I'm sure it can be done.
 
Back
Top