Tuesday, June 30th 2020

Microsoft Rumored To Release Budget Xbox Series S Console

In a recent report by the Verge, Microsoft is reportedly planning a budget console to go along with the Xbox Series X known as the Xbox Series S. The Xbox Series S will be a successor to the popular Xbox One S, codenamed Lockhart it's set to feature the key next-generation improvements found in the Xbox Series X and provide them at a lower price point.

Sources say that the Xbox Series S will target 1080p 60FPS and 1440p 6̶0̶F̶P̶S̶ 30FPS gaming performance, this will be achieved with roughly 1/3rd of the GPU power at 4 TeraFLOPs, and 10 GB of GDDR6 RAM. The Xbox Series S console is rumored to share the same 7 nm AMD Zen 2 SoC a̶t̶ ̶a̶ ̶l̶o̶w̶e̶r̶ ̶c̶l̶o̶c̶k̶ and the same ultra-fast PCIe 4.0 SSD. The console may not feature a disc drive like the Xbox One S All Digital Edition. Microsoft is yet to publicly say anything about the console, so take these rumors with a grain of salt.

Update Jun 30th: Sources tell Eurogamer that Microsoft is planning to unveil the console at an event in August.
Sources: The Verge, @tomwarren, Eurogamer
Add your own comment

57 Comments on Microsoft Rumored To Release Budget Xbox Series S Console

#26
kayjay010101
Vya DomusEven if so it will have less memory.
As @Valantar mentions, the memory is a shared pool between GPU and CPU. Going from 16GB to like 8GB wouldn't be as bad since the GPU is 1/3rd the performance anyway, and therefore will use much less memory, opening up more for the CPU. Also going from 4K to 1080p will drastically reduce texture sizes and therefore memory usage. I imagine both console versions still have the same "budget" of memory that the CPU can use.

I'm sure you, and not Microsoft, have managed to spot a critical flaw that means everything will be bad. It's not like Microsoft does any kind of R&D, testing or thinking about how their products work. Not one person at that trillion dollar company has even thought about that until you did.
Posted on Reply
#27
Vya Domus
ValantarBut these consoles allocate memory dynamically between the CPU and GPU, and the high end concoles will use a lot as VRAM due to their high resolutions. Running at 1080p or 1440p with lower texture quality will easily overcome this, and won't affect development for the higher end consoles whatsoever.
6 GB is a lot even when you take into consideration lower textures and resolutions. As a developer you will be forced to make sure the game fits in those 10GB, there is no going around it. You can fill the rest with trivial stuff like textures and whatnot on the higher end console but it's a waste of potential.
kayjay010101I'm sure you, and not Microsoft, have managed to spot a critical flaw that means everything will be bad. It's not like Microsoft does any kind of R&D, testing or thinking about how their products work. Not one person at that trillion dollar company has even thought about that until you did.
Don't get cheeky, it's not a critical flaw, it's an extra constraint developers will have to deal with. And, yes these trillion dollar companies have designed countless pieces of hardware with terrible constraints, look at Cell and it's separate GPU/system RAM, look at the ESRAM on the Xbox One, etc.
kayjay010101Also going from 4K to 1080p will drastically reduce texture sizes and therefore memory usage.
The render target resolution is independent from texture resolution.
Posted on Reply
#28
THANATOS
ValantarThere's no way the price difference will be only $100 with a GPU ⅓ the performance and less memory. If this is true it is clearly a value play, likely aiming for $300 or possibly even less.
$100 was a bit conservative estimate from me so I will change It to $150, but I don't think It will be $200.
If there was some option to play PC games on Xbox It would be a killer feature, but then most of PC gaming market(hardware) would crumble, so It won't happen.
Posted on Reply
#29
Valantar
Vya Domus6 GB is a lot even when you take into consideration lower textures and resolutions. As a developer you will be forced to make sure the game fits in those 10GB, there is no going around it. You can fill the rest with trivial stuff like textures and whatnot on the higher end console but it's a waste of potential.



Don't get cheeky, it's not a critical flaw, it's an extra constraint developers will have to deal with. And, yes these trillion dollar companies have designed countless pieces of hardware with terrible constraints, look at Cell and it's separate GPU/system RAM, look at the ESRAM on the Xbox One, etc.



The render target resolution is independent from texture resolution.
Put it this way: it's entirely possible the XSX is aiming for (an equivalent of) 4K "High" or even borderline "Ultra" settings levels in most games. The amount of extra work needed to implement a "Medium" level that focuses on running smoothly at 1080p60 or 1440p30 while staying within a 7.5GB memory footprint (2.5GB is reserved for the OS) rather than the 13.5GB of the XSX? Not that much, frankly. Especially if the game is cross-platform, as most of the work would then be done for the PC port no matter what. It's also worth pointing out that consoles with a unified memory architecture need (often significantly) less total memory than an equivalent PC due to the lack of duplicated assets between RAM and VRAM.
THANATOS$100 was a bit conservative estimate from me so I will change It to $150, but I don't think It will be $200.
If there was some option to play PC games on Xbox It would be a killer feature, but then most of PC gaming market(hardware) would crumble, so It won't happen.
They would struggle to sell this at those kinds of prices, given that it is after all an explicit low-end console. "Get ⅓ the performance for ⅔ the price" is a dirt-poor sales pitch. No optical drive, less RAM, smaller and cheaper motherboard, less expensive VRM components, smaller PSU, either a smaller and cheaper die or harvested chips that would otherwise go in the trash... $150 savings for this sounds like a relatively bad deal.
Posted on Reply
#30
kayjay010101
Vya DomusDon't get cheeky, it's not a critical flaw, it's an extra constraint developers will have to deal with. And, yes these trillion dollar companies have designed countless pieces of hardware with terrible constraints, look at Cell and it's separate GPU/system RAM, look at the ESRAM on the Xbox One, etc.
The render target resolution is independent from texture resolution.
Right, I see my hyperbole didn't quite stick its landing. Oh well.

I'm aware manufacturers can make massive mistakes, my point is this is quite a trivial oversight if it ends up being the case (which I'm quite confident in saying it won't) and not something that is quite on the same level as some of the other mistakes.

The render target res is in fact independent, but the overall memory usage is still lower by lowering res. Most PC games today will struggle with 4GB of VRAM at 4K, but same settings at 1080p and you'd see ~1GB VRAM usage. If the Series S ships with 12GB like the One X and we assume VRAM usage to be 2-3GB at most at 1080p (and the SSD will help get that number much lower as lower-priority assets can be stored on the SSD) and 4-5GB at 1440p, then that leaves 7GB at least for the rest of the system.
Posted on Reply
#31
Vya Domus
kayjay010101The render target res is in fact independent, but the overall memory usage is still lower by lowering res. Most PC games today will struggle with 4GB of VRAM at 4K, but same settings at 1080p and you'd see ~1GB VRAM usage. If the Series S ships with 12GB like the One X and we assume VRAM usage to be 2-3GB at most at 1080p (and the SSD will help get that number much lower as lower-priority assets can be stored on the SSD) and 4-5GB at 1440p, then that leaves 7GB at least for the rest of the system.
I know of no game that uses 1GB at 1080p and over 4 GB at 4K, the difference in memory usage is much less stark than people believe. Everyone makes the trivial error that 4 times the pixels means 4 times the overall VRAM usage. It doesn't work like that, a lot of data stored in VRAM doesn't need to scale up and remains the same in size. All of these are modern games :






See ? Are you all still not convinced games don't actually use that much more memory at 4K versus 1080p yet ? I know what's the next probably : "yeah but all those games just fill up all the available memory even if it wont use it all". No, it's not that, it's just not that big of a difference.

The being said I stand by my conclusion that 6GB is a huge chunk of memory which will act as as big constraint, forcing developer to simplify game logic and assets. And I wonder how much cheaper would this console really be, the SoC is the same, MS will pay the same for every wafer they get from AMD irrespective of where it will be used.
Posted on Reply
#32
THANATOS
ValantarThey would struggle to sell this at those kinds of prices, given that it is after all an explicit low-end console. "Get ⅓ the performance for ⅔ the price" is a dirt-poor sales pitch. No optical drive, less RAM, smaller and cheaper motherboard, less expensive VRM components, smaller PSU, either a smaller and cheaper die or harvested chips that would otherwise go in the trash... $150 savings for this sounds like a relatively bad deal.
I forgot It probably won't have an optical drive, so maybe $200 less in total.
I don't know about your country, but In my country Xbox One X with one included game costs €399. So If the big one costs €499 and the smaller one only €299, that wouldn't be such a bad deal in my opinion, but I would rather pay more for the stronger one.
Vya DomusThe being said I stand by my conclusion that 6GB is a huge chunk of memory which will act as as big constraint, forcing developer to simplify game logic and assets. And I wonder how much cheaper would this console really be, the SoC is the same, MS will pay the same for every wafer they get from AMD irrespective of where it will be used.
I don't think the SoC is the same.
For 4Tflops you need a GPU with only 20CU at 1560Mhz or 16CU at 1950Mhz and we know the SoC has 56CU in total. They won't disable that many CU, just to put It in a much cheaper console.
Posted on Reply
#33
Vya Domus
THANATOSI don't think the SoC is the same.
For 4Tflops you need a GPU with only 20CU at 1560Mhz or 16CU at 1950Mhz and we know the SoC has 56CU in total. They won't disable that many CU, just to put It in a much cheaper console.
I can't see them commissioning another custom die from AMD, these custom designs are expensive as hell.
Posted on Reply
#34
THANATOS
And disabling 32-36CU + other parts of GPU isn't? That's a lot of disabled space of the whole SoC and said SoC will be produced in millions.
Posted on Reply
#35
hat
Enthusiast
Vya DomusThis is a horrible idea and it will cripple game development across all platforms. Developers will be forced to adhere to the CPU/RAM limits of the weaker consoles, meaning the games will be as complex as the weaker consoles allows. The GPU situation can be dealt with easily but having a lower clocked CPU with less cores potentially and less RAM will ruin progress with this generation.
I'm inclined to agree. I thought consoles were all about offering a simple, hassle-free device for your average Joe to play games on. Having multiple models at different performance levels muddles up the waters a bit. Sure, there were like 9 versions of the PS2, but they were all relatively minor changes with little impact to the end user. There are some competitive people I know with original PS4s that are put off by playing competitively because their PS4 isn't as good as some of the later ones, so they're already at a flat disadvantage. It looks like it's only going to get worse.
Posted on Reply
#36
Vya Domus
THANATOSAnd disabling 32-36CU + other parts of GPU isn't? We are talking about millions of said SoCs used for the weaker brother.
The wafers they get from AMD will have defective CUs, AMD doesn't care about that, MS will pay for those anyway. So might as well use them in the cheaper console and by the way I don't actually think it will be a 4TF GPU, that seems way too weak. 6TF seems more reasonable.
Posted on Reply
#37
THANATOS
Vya DomusThe wafers they get from AMD will have defective CUs, AMD doesn't care about, MS will pay for those anyway. So might as well use them in the cheaper console and by the way I don't actually think it will be a 4TF GPU, that seems way too weak. 6TF seems more reasonable.
Wafers from AMD? Are you sure? In my opinion MS just pays a fixed licence fee per produced SoC to AMD, but the production is commissioned by MS to Samsung or TSMC.
Posted on Reply
#38
Vya Domus
THANATOSWafers from AMD? Are you sure? In my opinion MS just pays a fixed licence fee for produced SoCs to AMD, but the production is commissioned by MS to Samsung or TSMC.
It never works that way, you pay for the entire wafer. These SoCs will be produced in the millions, they agree to certain yields/bins and whatnot. AMD designs the chip you want and you do whatever you like with the wafer the chips are printed on.
Posted on Reply
#39
THANATOS
My point was that AMD just designed It, so MS can hardly pay AMD for wafers. MS pays TMSC or Samsung for wafers, because they make It and also pay AMD for each SoC the agreed licence fee.
If you have two different SoCs, then you will pay two different licence fees to AMD and even If It was the same the production cost for a smaller SoC is still lower and with millions of produced SoCs It's a lot of saved money.
Posted on Reply
#40
Vya Domus
MS -> AMD -> TSMC/Samsung or whoever else. That's how it works, you can't really design a chip independent from the manufacturing process, so AMD has to be the one which deals with TSMC not MS. It would make absolutely no sense to get a design from AMD and then try and get TSMC to make your chips when AMD already has the know-how on 7nm. That's the whole idea of the semi-custom concept AMD made, you can let AMD handle all of the logistics, you just say what you want them to make.

MS gets the chips from AMD not TSMC, I don't know why you insist on this. TSMC just makes the wafers with whatever is printed on them, you could draw stick figures with transistors onto them for all they care.
THANATOSthen you will pay two different licence fees to AMD and even If It was the same the production cost for a smaller SoC is still lower and with millions of produced SoCs It's a lot of saved money.
I don't think you quite get this, no matter the chip you pay the same amount of money for one wafer, the cost per die goes up and down because of yields. That's why it's probably too expensive to design another chip when you can just get the defective ones that you would have gotten anyway. This of course works for relatively small chips, if you'd have had something like a 700mm^2 monstrosity, yes, it would probably be cheaper to just make another chip.
Posted on Reply
#41
THANATOS
Naturally If AMD designed It, then It needs to work with TSMC, but I thought the deal was between MS-AMD(design) and MS-TMSC(production) and not that AMD was commissioning production to TMSC instead of MS.
Fo MS It should be better to have 2 SoCs rather than just 1.
Vya DomusI don't think you quite get this, no matter the chip you pay the same amount of money for one wafer, the cost per die goes up and down because of yields. That's why it's probably too expensive to design another chip when you can just get the defective ones that you would have gotten anyway. This of course works for relatively small chips, if you'd have had something like a 700mm^2 monstrosity, yes, it would probably be cheaper to just make another chip.
If Project Scarlett SoC is 360mm^2, but Lockhart SoC is <=300mm^2, then you can make more Lockhart SoCs from one wafer even with the same yield(failure) rate. With smaller SoC you don't need to order as much wafers as you would need with a bigger one so MS won't need to pay AMD for so many wafers.
With millions of SoCs needed for the cheaper console you won't have enough faulty SoCs to cover the demand and would need to partially disable a fully functional Scarlett SoC. It's not out of question that It would be cheaper to use Scarlett SoC instead of a smaller SoC, but only If the weaker console sells less units than the stronger one.
Posted on Reply
#42
demian_vi
Vya DomusI can't see them commissioning another custom die from AMD, these custom designs are expensive as hell.
the reality is we don't know if they have 2 SoC designs or will just use the defective ones for XSS. it's not like they lack the money to get 2 custom SoC designs. either way they can easily absorb the extra cost
Posted on Reply
#43
THANATOS
Out of curiosity I made a comparison.
I found a wafer yield calculator. Here
I found that defect density is 0.09 per cm^2 for 7nm.
Let's say 300mm 7nm wafer costs $10,000.

Scarlett SoC - 360mm^2 -> 112 Good ones and 41 faulty ones.
Cost per good SoC is 10,000/112 = $89,3

"Lockhart" SoC - 285mm^2 -> 152 Good ones and 44 faulty ones.
Cost per good SoC is 10,000/152 = $65.8

If you wanted to make 10,000,000 Scarlett SoCs for the Lockhart console you would need 89,286 wafers which would cost $892,860,000.
If you wanted to make 10,000,000 Lockhart SoCs for the Lockhart console you would need 65,790 wafers which would cost $657,900,000.
You would save $234,960,000 on wafer cost. This is when I excluded the faulty Scarlett SoCs, from which some can be certainly reused.

If let's say 30 can be reused then 142 will be used in total, It would be 10,000/142 = $70,4 per SoC.
If you wanted to make 10,000,000 Scarlett SoCs for the Lockhart console you would need 70,423 wafers which would cost $704,230,000 and that's only $46,330.000 more.
Of course this difference can decrease or increase depending on how many Lockhart consoles will be made.
Posted on Reply
#44
Valantar
Remember that the Xbox Series X SoC has more CUs on it than MS is enabling (56 in silicon, 52 enabled), which given the % area of the die dedicated to the GPU will mean that a huge proportion of faulty dice will be perfectly usable for the full XSX by disabling whatever CU the error is in. The chances of a die having more than four defective CUs yet no faults that make it useless otherwise (memory controllers, CPU cores, etc.) sound mind-bogglingly small.
hatI'm inclined to agree. I thought consoles were all about offering a simple, hassle-free device for your average Joe to play games on. Having multiple models at different performance levels muddles up the waters a bit. Sure, there were like 9 versions of the PS2, but they were all relatively minor changes with little impact to the end user. There are some competitive people I know with original PS4s that are put off by playing competitively because their PS4 isn't as good as some of the later ones, so they're already at a flat disadvantage. It looks like it's only going to get worse.
IMO, "pay X for full performance and all the features, or pay 3/5X for a lower performance version that still does all the same things and plays all the same games" is still a relatively simple proposition for end users, even if it adds a single layer of complexity (that is frankly already there with the X1X/X1S and PS4s/PS4p). Keeping the CPU the same across the board also means that (outside of draw calls increasing CPU load for the higher end SKU) the only relevant difference will be in graphical fidelity. Same responsiveness, same framerates (given that developers adjust for the less powerful GPU), same games, same disks, same UI, though likely smaller installations/downloads on the lower end SKU, and likely no RTRT.
Vya DomusThe wafers they get from AMD will have defective CUs, AMD doesn't care about that, MS will pay for those anyway. So might as well use them in the cheaper console and by the way I don't actually think it will be a 4TF GPU, that seems way too weak. 6TF seems more reasonable.
See above. The chance of a die from one of these 360mm wafers being unusable for the XSX is very, very small. Which leads me to think any potential Lockhart would be a different die - after all, to reach 4 TFlops at the same clock as the XSX (1825MHz) they would just need 18 CUs. If a smaller die allowed them to clock higher they could get away with 16 or even 15. Make that into a 20 CU die for some redundancy, reduce the number of memory controllers and interfaces to match the 10GB of RAM, and you've got a relatively small die. Not Renoir size (150mm2), but likely ~200mm2. That would allow for dramatic increases in chip yields - a 13x16mm die (208mm2) run through the calculator @THANATOS used above results in 244 good dice and 46 partials, nearly 2x the die count of the full Scarlett SoC. Also, as discussed above 4TF of RDNA compute won't lag much behind 6TF of GCN compute in gaming performance - and unlike the X1X this would be targeting lower resolutions than 4k, meaning 4TF ought to be perfectly fine as long as other parts of the SoC don't choke. Speaking of which:
Vya DomusI know of no game that uses 1GB at 1080p and over 4 GB at 4K, the difference in memory usage is much less stark than people believe. Everyone makes the trivial error that 4 times the pixels means 4 times the overall VRAM usage. It doesn't work like that, a lot of data stored in VRAM doesn't need to scale up and remains the same in size. All of these are modern games :






See ? Are you all still not convinced games don't actually use that much more memory at 4K versus 1080p yet ? I know what's the next probably : "yeah but all those games just fill up all the available memory even if it wont use it all". No, it's not that, it's just not that big of a difference.

The being said I stand by my conclusion that 6GB is a huge chunk of memory which will act as as big constraint, forcing developer to simplify game logic and assets. And I wonder how much cheaper would this console really be, the SoC is the same, MS will pay the same for every wafer they get from AMD irrespective of where it will be used.
a) Those tests are all run on Nvidia GPUs with Nvidia drivers, which are known to (on average, though it depends on the title) have more VRAM labeled "in use" while gaming than equivalent AMD GPUs. Whether they actually use or need more VRAM or if this is just a difference in how the drivers work is another issue entirely, but the point nonetheless stands.
b) A lot of that data in VRAM is loaded "just in case", with a relatively small portion ever actually being used. Given that the new consoles are built around NVMe SSDs and superior-to-Windows storage architectures, it's not that much of a stretch to assume that they can do this less, relying on the SSD to stream in necessary data on the fly (to a certain extent, of course).
c) Those tests are run on Ultra settings. No console runs PC Ultra settings, as Ultra inevitably activates all the stupidly expensive stuff that you barely even notice. Most games have significant drops from Ultra to High.
d) It's entirely within reason to expect a lower-end console to use a lower tier of graphics settings too - likely something akin to PC "medium" settings, further lowering VRAM usage.
Posted on Reply
#45
Vya Domus
Valantarse tests are all run on Nvidia GPUs with Nvidia drivers, which are known to (on average, though it depends on the title) have more VRAM labeled "in use" while gaming than equivalent AMD GPUs. Whether they actually use or need more VRAM or if this is just a difference in how the drivers work is another issue entirely, but the point nonetheless stands.
b) A lot of that data in VRAM is loaded "just in case", with a relatively small portion ever actually being used. Given that the new consoles are built around NVMe SSDs and superior-to-Windows storage architectures, it's not that much of a stretch to assume that they can do this less, relying on the SSD to stream in necessary data on the fly (to a certain extent, of course).
c) Those tests are run on Ultra settings. No console runs PC Ultra settings, as Ultra inevitably activates all the stupidly expensive stuff that you barely even notice. Most games have significant drops from Ultra to High.
d) It's entirely within reason to expect a lower-end console to use a lower tier of graphics settings too - likely something akin to PC "medium" settings, further lowering VRAM usage.
Why not just show me instances of games using a lot more memory at higher resolution ? None of your explanations have a real foundation.

a) It makes no sense at all why an AMD card would use more VRAM. Does a character model gain extra vertices on an AMD GPU ? Please explain, what exactly is the driver doing.

b) Of course not all stored data is used in rendering every single frame, that holds true whether you run a game at the lowest possible settings or if you run it at 4K with everything turned all the way up. VRAM usage is VRAM usage, it makes no sense to argue that some of the data is not used. No program is ever using every single byte simultaneously at once, come on.

c) Why does it matter if it's ultra or not ? You do realize that if the settings were to be turned down, the difference between lower and higher resolutions would be even smaller, right ?

d) Of course it is, still can't justify how all that cut down memory somehow only has to do with graphical settings. I'll remind you that you can currently fit entire frame buffers of modern games in something like 6GB. How much more memory do you think higher resolution shadows maps use for example ?
Posted on Reply
#46
Valantar
Vya DomusWhy not just show me instances of games using a lot more memory at higher resolution ? None of your explanations have a real foundation.

a) It makes no sense at all why an AMD card would use more VRAM. Does a character model gain extra vertices on an AMD GPU ? Please explain, what exactly is the driver doing.

b) Of course not all stored data is used in rendering every single frame, that holds true whether you run a game at the lowest possible settings or if you run it at 4K with everything turned all the way up. VRAM usage is VRAM usage, it makes no sense to argue that some of the data is not used. No program is ever using every single byte simultaneously at once, come on.

c) Why does it matter if it's ultra or not ? You do realize that if the settings were to be turned down, the difference between lower and higher resolutions would be even smaller, right ?

d) Of course it is, still can't justify how all that cut down memory somehow only has to do with graphical settings. I'll remind you that you can currently fit entire frame buffers of modern games in something like 6GB. How much more memory do you think higher resolution shadows maps use for example ?
So what you are worried about is non-graphics portions of games needing that memory? Such as what, precisely? Most PC games run perfectly fine on 8GB of RAM, which under Windows 10 at best means 6GB free, though likely more like 4. And again, those games aren't built around a storage architecture delivering near-instant loads (no, it doesn't matter if you have an NVMe SSD in your PC, the game is programmed to pre-load anything it might need at the rate necessary if running from a 2.5" HDD. All your SSD is doing is speeding up the pre-load), meaning a lot of "maybe we'll need this too" data can be kept on the SSD, with the game knowing exactly how quickly it can load it, reducing actual RAM usage.

As for your responses:
a) as I said, it's quite well documented that VRAM usage differs in the same game at the same settings between Nvidia and AMD. But you misread me; AMD typically has less memory flagged as in use than Nvidia. I would guess this is down to various driver related processes, such as the previously discussed pre-loading of data, compression/decompression, etc. I definitely didn't say this means either OEM's cards use more or less VRAM than the other while gaming, in fact I explicitly underscored that this is likely not what is happening.

b) so... if not all data in RAM is needed, and the system and game both know the only storage present is fast NVMe, perhaps some of that unnecessary data can be kept out of RAM and on the SSD?

c) What? No. Resolution is one of dozens if not hundreds of variables in graphics settings. Reducing rendering quality does not equate to lowering the resolution. 1080p with amazing shadows and lighting is still less sharp than 4k with poor shadows and lighting. As to whether sharpness or rendering quality is the biggest determinant of visual quality? Both, in various situations and ways. But then again here we are talking about a lower end SKU that might likely run at BOTH a lower resolution AND lower settings. Makes sense, no? And, again, this would lower VRAM needs for the lower end SKU.

As for d), you really ought to have pointed out earlier that you were talking about RAM usage outside of graphics, as that is what the entire debate has been around up until this. Being vague and overly general doesn't get your point across. Beyond this, I've covered this above.
Posted on Reply
#47
danbert2000
I think that this is a lot smarter than Sony's attempt at a lower price point by removing the disc drive. Microsoft will be able to squeeze Sony by having the "cheapest next gen console" even though it will be limited to 1080p. Also, by baking in a lower performance point into the equation, we could see something like a portable Xbox down the line. That's wishful thinking, of course, but at 5 nm or 3 nm maybe the Xbox APU for the Series S could be shoved into a handheld.

Thinking about cost savings, the lower price point should allow for a much cheaper PSU, smaller die size for the processor or maybe harvested Series X chips with too many defects, lower RAM costs, and potentially a missing disc drive too. I hope that adds up to a $300 launch as anything more would likely not be worth releasing separately.
Posted on Reply
#48
demian_vi
Switch Lite showed that there's a market for cheaper versions of consoles. not everyone has a 4k tv and someone who can't afford a 4k tv, would likely prefer a cheaper console.
Microsoft has definitely done some market research before deciding this
danbert2000Thinking about cost savings, the lower price point should allow for a much cheaper PSU, smaller die size for the processor or maybe harvested Series X chips with too many defects, lower RAM costs, and potentially a missing disc drive too. I hope that adds up to a $300 launch as anything more would likely not be worth releasing separately.
Cpu will be the same.
Posted on Reply
#49
Valantar
demian_viSwitch Lite showed that there's a market for cheaper versions of consoles. not everyone has a 4k tv and someone who can't afford a 4k tv, would likely prefer a cheaper console.
Microsoft has definitely done some market research before deciding this


Cpu will be the same.
CPU being the same does not mean the SoC is the same, and it's highly unlikely that there will be enough defective XSX dice to support this being based on cut down chips. If this is real it's likely a second die design with the same CPU but less GPU CUs and RAM channels.
Posted on Reply
#50
Naito
CybrshrkYou do know 4k requires roughly 4x the GPU power vs 1080p and 1440(at 30 fps) is within the 3x less power (4/12) that differentiates them.

By keeping the exact same everything else (with reduced memory again because that'd directly tied to resolution) they can offer. Slower cost lower resolution version without compromising the experience of series x owners.
I understand that. However, 4K TVs have become very prevalent (I've seen as low as $300 where I am) and for a console which will probably be in the market for 7 to 8 years, it should still be able to drive 4K at least as well as the Scorpio (One X) currently does. Scaling back performance to only support 1080/1440p seems short sighted. Though, as others have pointed out, 4TF of RDNA2 may perform as well or better than the 6TF on hand in the Scorpio.
Posted on Reply
Add your own comment
Jul 31st, 2024 17:13 EDT change timezone

New Forum Posts

Popular Reviews

Controversial News Posts