Monday, January 4th 2016
Microsoft Could Refresh Xbox One Design with Polaris Based SoC
Microsoft could make the Xbox One "slimmer" and more energy-efficient by leveraging AMD's upcoming 4th generation Graphics CoreNext architecture, codenamed "Polaris." Found in the footnotes of AMD's "Polaris" press-deck, mentioned as an ominous-sounding "Xbox One Polaris," the mention hints at a revision of Xbox One that features an SoC running a "Polaris" based SoC. Microsoft could leverage the energy-efficiency improvements, and possibly upcoming processes, such as 14 nm FinFET, to bring down power-draw, thermal requirements, and possibly cost.
Source:
Overclock.net
17 Comments on Microsoft Could Refresh Xbox One Design with Polaris Based SoC
Just another BS rumour from one of those misterXmedia lunatics...
Isn't this pretty much assumed? The 360 had so many processor changes that it wasn't funny (though having to cripple the later systems in order to match performance on the initial systems did warrant a chuckle).
As much as I hate to say this, hasn't the PS4 pretty much already won this generation of consoles? MS rode the Kinect requirement into the ground, their games library is somewhat lackluster (when you remove all the multi-platform games), and MS's cash cow Halo is starting to get criticism even from the die hard players.
Interesting that, despite both being built on AMD's SOCs, this hasn't been teased out of Sony yet.
Will be interesting to see how things play out. If Nintendo high balls, which it doesn't have to do much to do, then we might see next gen sooner. Think about it, a Zen + Polaris based console going up against these current paperweights...lol.
I'm working under the assumption that the successor to the WiiU is what you're talking about. I'm also working under the assumption that "under specified" nature of the PS3 and Xbox One is a reference to the fact that both Sony and MS tried to have their consoles near break even price, rather than have to have a console on the market for 5+ years before they turned a profit by decreasing production costs.
Under the above two assumptions, how does Nintendo fit in? Their 3rd party support started to dry up in the Gamecube generation, and has been functionally non-existent in the current generation. If they decided to compete with MS and Sony they'd have to release a console that not only made money, but was also as easy to port to as the functionally similar PS4 and Xbox One. Similarly, Nintendo would have to churn out first party software at an insane rate to make its market penetration substantial enough to warrant the development costs for the publishers. They'd have to do all of this, while still being price competitive (which means insane specifications aren't likely). How exactly does this make Nintendo a force that MS needs to compete with?
Now I can see Nintendo actually capable of a genuine "year of Nintendo." They've got enough financial resources, enough old IP, and more than enough know-how to bring themselves back onto the market in a big way, but why? Nintendo basically has the lock on handheld gaming. They can sell a respectable (if not substantial) amount of hardware based only on their first party production. Nintendo has demonstrated that they understand the market is overdue for a massive shift, and they're sitting back and waiting for MS or Sony to overplay their hands and let a publisher drag them down. It's a quintessentially Japanese tradition to believe that doing it best is enough, as their management has time and again demonstrated. Nintendo doesn't have to be the biggest, but they have to make the best games. That mentality doesn't lend itself to getting involved with petty squabbles amongst "upstart" developers. I honestly believe Nintendo is insulating themselves for another games crash, which is why they are silent. They're more than content to let Sony and MS waste their efforts competing, while toxic business practices brings the game market down around them.
This is a very, VERY suspicious rumor and looks quite baseless, and yet more and more "tech" sites are repeating it, even though they should know better. Would you care to elaborate? What do you mean "cripple"? All those CPU and GPU changes were basically die-shrinks. AFAIK, they all were the exact same designs with the exact same specs, simply built on different nodes and different chip combinations.
Perhaps you meant this: arstechnica.com/gaming/2010/08/microsoft-beats-intel-amd-to-market-with-cpugpu-combo-chip/
The CPU and the GPU were the same. The only issue was the FSB. Yeah, I see why it might look like crippling, but it's quite a small one and had to be done to maintain absolute compatibility.
en.wikipedia.org/wiki/Xbox_360_technical_specifications
The 360 started out with a separate 90 nm CPU and GPU. Over the course of several revisions it managed to get down to a single chip 45nm design, with the CPU and GPU being of different sizes for a substantial number of revisions. Your linked article does an excellent job of showing how they radically redesigned the silicon from a dual chip system into an SOC. They had to cripple FSB speeds to maintain parity among the models on offer blog.gsmarena.com/heres-what%E2%80%99s-inside-the-latest-xbox-360-slim-gaming-console/, because their architecture changed sufficiently enough that they'd experience better performance without the crippling.
I'm not sure exactly what's powering your logic here, so I need some assistance from you. They changed the structure of the chips; they altered how it accessed memory, and they managed to go to a process which effectively allowed them to do all of this by decreasing the area required to create their chips by 75%. That, in my book, is completely redesigning their architecture to allow an SOC to perform at exactly the same level as a dual chip system. Just because they didn't create additional performance doesn't mean they didn't redesign their silicon.
If that isn't correct, then I don't exactly know what you'd call a redesign. Simply adopting a new architecture on the GPU would require as much effort to maintain parity (as well as determining the right level of crippling) as MS put into making sure their SOC solution matched the performance of their dual chip systems (those that managed to avoid the fiery RROD deaths that killed so many of the first couple of generations of hardware). I can't find the link to the article now, but at the time I remember a hardware take down even making note of the vastly increased trace lengths on some of the new 360's interconnects designed to induce enough latency into the system to match older model's performance.
The only part of that SoC that was entirely different was that new FSB replacement block which was necessary since the interconnections inside an SoC have much lower latency than two separate chips.
All this brings us to the main discussion that introducing a major architectural change, like the one the article is suggesting, is very doubtful. How can they make a completely new GPU architecture to behave exactly the same as an old one?
My qualifier is anything requiring substantial redesign of silicon structure of layout. This ranges from something as simple as changing how memory is accessed (HBM versus GDDR5, or in the case of the Xbox 360 the shared memory access), to as substantial as redesigning the pipeline for the processes. Under that definition, the integration of the GPU and CPU, along with a substantial change in the memory access, is redesigning the architecture of the 360.
I can acquiesce that if your definition is different, and much more narrow, that there has never been an architectural change to a console. I reject that limited of a definition, but can respect the logic in coming to it.
A simple example of nvidia architectures: ..., Tesla, Fermi, Kepler, Maxwell 1st gen, Maxwell 2nd gen, ... (IINM, there were also some architectural changes within the Fermi and Kepler families too, when going from 1xx line to 11x for the former and going from 1xx to 2xx for the latter, so I suppose they could be listed separately too. The Tesla architecture also changed once or twice in its lifetime.) Did you miss the part where I separated the chip/SoC design from the CPU and the GPU? I even agreed with you that the chip itself is different from before.
I am specifically talking about the architectures of the CPU and GPU blocks. As far as we know, those two blocks in the newer 45nm SoCs are exactly the same as the blocks in the older 2-chip setups. Only the nodes are different.
Yes, the "architecture" of the 360 itself changed, but the architecture of the main computing components and blocks didn't.
Red_Machine explained it quite nicely in his comment.
They function on a tick-tock cycle. They theoretically release a "new" microarchitecture, then the following cycle they die shrink it. This die shrink hasn't often coincided with any real changes to the functional blocks on their processor, but instead of being general let's focus on Sandy Bridge and Ivy Bridge. Intel lists these as separate microarchitectures, despite them functionally being the same structure on the CPU side. The only real change between the two is the die shrink, and all of the peripherals that are included. To prove my point, take a gander at the two pictures:
www.guru3d.com/articles-pages/core-i7-3960x-processor-amp-msi-x79a-gd65-review,2.html
www.guru3d.com/articles-pages/core-i7-4820k-processor-review,2.html
So I ask again, how do you define an architecture change? In the case of Intel, something as simple as changing interconnection possibilities qualifies. Hell, I'd even be willing to admit I was wrong if Intel only called them an architecture change and you wanted to argue microarchitecture. Problem is, in Intel's book IB and SB are separate microarchitectures. Is that not exactly what the integration of the Xbox 360 did. More than simply slapping the two chips onto a single die, they offered a new way to access memory. They didn't just die shrink, they created an entirely new chunk of silicon. This new silicon is designed only to parity the performance of the old, but we've accepted that from Intel since SB.
I'm going to have to go with how Intel labels this stuff. Arguing against them is pointless, even if you don't accept their liberal definitions. They've written the book on consumer CPUs, by nearly obliterating their competition.
www.anandtech.com/show/4830/intels-ivy-bridge-architecture-exposed/2
Read those. It does not have the "same" structure. Ivy bridge was not a simple die shrink with just some changes to the interconnections. Even though it is based on sandy bridge and is VERY similar to it, nearly identical, it's not the same. They wouldn't have bothered to label it as a different architecture if it was the exact same. They do exaggerate though, for changes that are rather small. They could've maybe called it sandy bridge 2nd gen.
It is not exactly what the integration of the Xbox 360 did, because there were changes to ivy bridge's pipeline itself. Quite minor ones, but still. As I pointed out before, the CPU and GPU blocks of the chips used in 360s remained the same, as far as we know.
On the other hand, changing the memory or the interconnects between the different blocks in a chip/SoC would not change the blocks themselves. It's quite obvious. Unless you can find a paper where it says the CPU and GPU blocks of the newer 45nm 360s are different, even in the smallest, then their architecture/microarchitecture didn't change.
It would defy console logic for MS to willingly change the GPU and CPU on 360. The FSB situation was an unfortunate side effect that was unavoidable, unless they decided to stay with a two-chip design, but they didn't because they wanted to reduce the manufacturing cost first and foremost.
I'm not saying you're wrong, but you're mixing chip/SoC design with CPU and GPU design.
Yes, the new SoC itself perhaps can be called a new chip "architecture", but that's not the word that is used for such purpose. It's simply a new package, containing the same computational blocks as before.