Thursday, May 7th 2020

Intel Ready with 144-layer 3D NAND On its Own, Talks 4-layer 3DXP, "Alder Stream" and "Keystone Harbor"

Intel's memory and storage products division now has a completely independent NAND flash technology development team post its split with Micron Technology, with which it was part of the IMFlash Technologies joint-venture. Intel is close to gaining a technological lead over Micron with a new 144-layer 3D NAND flash chip which will ship roughly around the time Micron begins pushing out its 128-layer 3D NAND chips. SK Hynix will begin shipping its 128-layer 3D NAND flash chips later this year. KIOXIA will put out 112-layer chips before the turn of the year. YMTC is developing its portfolio at a breakneck pace.

The 144-layer 3D NAND flash chip by Intel can handle up to four bits per cell (QLC), and can be configured to function as TLC or SLC, at lower densities. Intel will launch its first SSD based on this 144-layer QLC NAND flash chip, codenamed "Keystone Harbor," later this year. Development is underway at Intel for PLC (5 bits per cell) technology, which should drive up densities by 25 percent. Intel is also close to launching its second generation 3D X-point memory technology.
The 2nd gen 3D X-point memory chip by Intel will feature four layers on the physical media, as opposed to two on the first-gen. The first Optane product based on the new chip is codenamed "Alder Stream." The drive will launch later this year in a single-port variant, with dual-port variants planned for 2021. The drive uses an advanced new controller that features PCI-Express gen 4.0 host interface. It's likely that "Alder Stream" will be part of Intel's big next-gen Xeon "Sapphire Rapids" processor and "Eagle Stream" platform launch, which features PCIe gen 4.0.
Source: Blocks and Files
Add your own comment

45 Comments on Intel Ready with 144-layer 3D NAND On its Own, Talks 4-layer 3DXP, "Alder Stream" and "Keystone Harbor"

#27
efikkan
AnarchoPrimitivJust use the 380gb 110mm long m.2....most boards support at least one 110mm m.2
And if not, there are adapters if you have free PCIe slots.

I probably will use such adapters in my future builds, just to make it more convenient to pull them out. The M2 form factor might make sense in laptops, but in desktop they are just a pain.

When I build computers, they are usually in service for 8-10 years, but their role changes over time. And this is one of the biggest strengths of desktop computers; many things can be adapted or changed, each usable part can be reused somewhere else. So graphics cards are usually swapped a couple of times, SSDs and HDDs swapped a lot, etc. Then it's a pain if I have to unplug half of the stuff to swap an SSD, especially if I'm troubleshooting something. I'm also tired of small cases, so unlike londiste, I will do my future builds in spacious towers, do only basic cable management, etc. Computers to me are not meant to be pretty, but serve as tools. ;)
Posted on Reply
#28
londiste
efikkanWhen I build computers, they are usually in service for 8-10 years, but their role changes over time. And this is one of the biggest strengths of desktop computers; many things can be adapted or changed, each usable part can be reused somewhere else. So graphics cards are usually swapped a couple of times, SSDs and HDDs swapped a lot, etc. Then it's a pain if I have to unplug half of the stuff to swap an SSD, especially if I'm troubleshooting something. I'm also tired of small cases, so unlike londiste, I will do my future builds in spacious towers, do only basic cable management, etc. Computers to me are not meant to be pretty, but serve as tools. ;)
It gets offtopic but I am seeing less and less need for big computer these days. For a fully functional computer, put CPU with cooler on the motherboard, add RAM and M.2 drive or two and the only cables you need are 24-pin and 4/8 pin for power. If you need GPU, it goes to PCIe slot with possibly a couple power cables of its own and done. Very nice and clean. This is also why I tend to like the idea of 10-pin ATX12VO - that would get a few more cables out of the way.

Depends on what you use the computer for but I am noticing that I swap parts and do random testing and troubleshooting less and less every year. There is mostly no need for that. Swapping GPU is the same regardless of case size (excluding some extremes like my A4SFX maybe), CPU is more likely to be replaced with motherboard and when it comes to drives how often would you want to swap those and why? For testing HDDs/SSDs, a USB dock has been my go-to method for well over a decade now and troubleshooting M.2/mSATA etc drives is PITA one way or another.

Edit:
For mass storage, NASes are pretty good these days. I would not run games or storage-dependent work off of them but a cheap enough Synology box for example will nicely saturate a 1Gbit Ethernet bandwidth which isn't much slower than what drives are realistically capable of.

I would bet constant testing and swapping and adapting is pretty niche thing in itself :)
Posted on Reply
#29
efikkan
londisteFor testing HDDs/SSDs, a USB dock has been my go-to method for well over a decade now and troubleshooting M.2/mSATA etc drives is PITA one way or another.
Not what I meant. I have a dock too.
londisteI would bet constant testing and swapping and adapting is pretty niche thing in itself :)
It is, but it's not what I meant. Those who do usually have open rigs for that.
Click the spoiler for details.

Like many for many power users, my needs evolve over time. As of now I have two operational desktops, one laptop and two servers, 14 HDDs and 7 SSDs in use between them. Every once in a while a drive needs replacing, either because it's bad or because it's too small, if it's still good then it moves to another machine. The same with GPUs; if I buy a new one, the old one is moved to another machine. Then it's really appreciated when installation takes 2 minutes instead of 20+, like one of my machines which sits in a Fractal Node 304; I've had to open it five times in the last few months; replacing a HDD, adding another SSD, two times because some cable fell out because it's too cramped, and once because the PSU blew up (but difficulties changing a PSU is understandable and rare).

My "work" machine have one SSD for the OS, two SSDs in RAID 1 for work, and two HDDs for long-term storage, all marked with labels to make it easy to identify and replace. If a drive goes bad I can pull it and replace it, if the machine goes bad, I can move the drives to another machine and keep working. If I were to put a M2 drive in there, I would probably put it in a PCIe adapter to make it easier to pull out.

Back on topic,
QLC is bad enough, but with PLC I would be really worried about file integrity. I would certainly put OS and other stuff on separate drives, but this still sounds scary. I do wonder if controllers will move things around more to reduce data rot. Still, I think NAND flash is in dire need of a replacement.
Posted on Reply
#30
londiste
efikkanQLC is bad enough, but with PLC I would be really worried about file integrity.
Return of the WORM? :)
Posted on Reply
#31
lexluthermiester
efikkanQLC is bad enough, but with PLC I would be really worried about file integrity.
I'm not touching either until the process can be proved reliable for a minimum of 5 years with an expected useful lifespan of 10years+. And before anyone whines that such is not a reasonable expectation, save it. I still have OCZ SSD's from 2009 that are still going strong. If that level of durability is not a factor in the engineering of an SSD, I'm not buying it. MLC and TLC have been refined and engineered to the point where that level of durability(or close to it) can be expected.
Posted on Reply
#32
voltage
I am very curious to see what second generation 3D X-point memory technology will offer.
Posted on Reply
#33
efikkan
If the second generation 3D XPoint cuts the price per GB in half, then many prosumers will start to get interested, including myself.
Posted on Reply
#34
londiste
efikkanIf the second generation 3D XPoint cuts the price per GB in half, then many prosumers will start to get interested, including myself.
Current XPoint is manufactured on 20nm class node. This is very likely to be shrunk to some 10nm class node that DRAM and NAND Flash is manufactured on these days.

What I am a bit vary of is that 2nd gen XPoint might actually get a nerf of some kind. XPoint itself is overspecced for anything it is used for today in consumer or enthusiast space. The exception is probably the Optane DIMM technology. NVMe controllers cannot seem to use XPoint speed and latency to fullest. This leaves some space for simplification or relaxed specs.
lexluthermiesterI'm not touching either until the process can be proved reliable for a minimum of 5 years with an expected useful lifespan of 10years+. And before anyone whines that such is not a reasonable expectation, save it. I still have OCZ SSD's from 2009 that are still going strong. If that level of durability is not a factor in the engineering of an SSD, I'm not buying it. MLC and TLC have been refined and engineered to the point where that level of durability(or close to it) can be expected.
Depends on what you use these for. 5 years is a long time when we look at how SSDs have evolved. In 2009 the reasonably priced SSDs were 120GB, maybe 240GB with especially write speeds that are low by today's standards. Manufacturers have learned from their mistakes for the most part when it comes to SSD endurance. Today, I am quite convinced that endurance rating claim in manufacturer spec is legitimate, at least for major manufacturers.

Price remains a factor and QLC has its place. I have a 2TB 660p for my secondary drive (mainly for games) which I got for ~160€ last year. Intel spec says it is good for 400 TBW and in about 9 months I have put 10TB on it. If what Intel claims is even remotely accurate, I am fine for a long while. The downsides of QLC I can live with in this particular configuration - unless I fill the drive to 95+%, it's faster than a SATA SSD. Filling the last 5% is indeed very painful but awareness of that helps :)
Posted on Reply
#35
Breit
londisteCurrent XPoint is manufactured on 20nm class node. This is very likely to be shrunk to some 10nm class node that DRAM and NAND Flash is manufactured on these days.

What I am a bit vary of is that 2nd gen XPoint might actually get a nerf of some kind. XPoint itself is overspecced for anything it is used for today in consumer or enthusiast space. The exception is probably the Optane DIMM technology. NVMe controllers cannot seem to use XPoint speed and latency to fullest. This leaves some space for simplification or relaxed specs.

Depends on what you use these for. 5 years is a long time when we look at how SSDs have evolved. In 2009 the reasonably priced SSDs were 120GB, maybe 240GB with especially write speeds that are low by today's standards. Manufacturers have learned from their mistakes for the most part when it comes to SSD endurance. Today, I am quite convinced that endurance rating claim in manufacturer spec is legitimate, at least for major manufacturers.

Price remains a factor and QLC has its place. I have a 2TB 660p for my secondary drive (mainly for games) which I got for ~160€ last year. Intel spec says it is good for 400 TBW and in about 9 months I have put 10TB on it. If what Intel claims is even remotely accurate, I am fine for a long while. The downsides of QLC I can live with in this particular configuration - unless I fill the drive to 95+%, it's faster than a SATA SSD. Filling the last 5% is indeed very painful but awareness of that helps :)
You can always overprovision that SSD yourself if you create your partitions to 90% of the total available space and forget about that.

In fact, where does all this "QLC is unreliable as shit" attitude come from? Are there any real tests or experiences published that I'm not aware of?
If QLC is used for longterm storage, you barely write to it anyways and if used in a redundant array with a proper file system in a NAS, you can scrub your stored data to detect and repair flipped bits as often as you like? Nobody said that QLC is good for every application...
Posted on Reply
#36
londiste
The usually quoted numbers are roughly 100 000 write cycles or more for SLC (1 bit per cell), 10 000 - 30 000 for MLC (2 bits per cell), 3 000 - 5 000 for TLC (3 bits per cell) and 1 000 for QLC. Now, all this is not necessarily true. The exact number for specific manufacturers' Flash dies are not publicly known. Also, 3D or V-NAND is not just for marketing but was an actual innovation in building the cells to height to increase endurance. MLC and TLC numbers earlier are for currently used technologies, good old planar Flash would have far lower endurance.

QLC definitely has much less write cycles compared to TLC. In a write-heavy environment this is important. For consumer use cases, might not be too crucial. The other downside of QLC is its speed, particularly write speed. Due to the accuracy of voltage levels needed for that many levels, writes are painfully slow. The approach controllers and drives take to mitigate this is the same as with TLC Flash - SLC cache. Some part of the drive is treated as SLC and that allows very fast write speeds. After writes to the drive finish, drive itself will move the data around to QLC and frees up the SLC cache for more writes.
Posted on Reply
#37
efikkan
londistePrice remains a factor and QLC has its place. I have a 2TB 660p for my secondary drive (mainly for games) which I got for ~160€ last year. Intel spec says it is good for 400 TBW and in about 9 months I have put 10TB on it. If what Intel claims is even remotely accurate, I am fine for a long while. The downsides of QLC I can live with in this particular configuration - unless I fill the drive to 95+%, it's faster than a SATA SSD. Filling the last 5% is indeed very painful but awareness of that helps :)
Those endurance ratings are probably as inflated as MTBF numbers for HDDs. It also depends what the maker defines as "working". To me, a drive is no longer "usable" when it gets SMART errors, especially if it's reformatted and still keeps getting them.

While I can't claim to have the empirical data to support a scientific conclusion, the trend I'm seeing is really concerning. In the past couple of years I've seen a decent number of SSDs which become very unstable after 2-3 years of use, and most of them nowhere near their endurance rating. I've seen a lot of HDDs go bad in the last 25 years, but nothing close to this failure rate.

There are at least two major problems with SSDs; stability of the cells and data rot. Especially data rot gets a lot worse when they cram more bits into each cell. I wish SSD specs would list what the controllers do to deal with data rot, if they do anything at all.
Posted on Reply
#38
Tomorrow
londisteCurrent XPoint is manufactured on 20nm class node. This is very likely to be shrunk to some 10nm class node that DRAM and NAND Flash is manufactured on these days.
14nm++++++++++++++++++ :D
londistePrice remains a factor and QLC has its place. I have a 2TB 660p for my secondary drive (mainly for games) which I got for ~160€ last year.
That's a good price for 660p. Unfortunately the prices have risen considerably making QLC all but useless. 660p costs ~250€ for 2TB and 665p costs nearly ~450€. Jesus F christ. For 450 i can buy a 2TB TLC based PCIe 4.0 drive. Crucial P1 seems to be all but gone looking at prices.

The cheapest ones are the slower SATA based Samsung 860 QVO and ADATA SU630 around 200€ for 2TB. Still too much.
SATA based 2TB QLC should be 150€ max and PCIe versions should be around 200€ like they were last year.
Posted on Reply
#39
londiste
efikkanThose endurance ratings are probably as inflated as MTBF numbers for HDDs. It also depends what the maker defines as "working". To me, a drive is no longer "usable" when it gets SMART errors, especially if it's reformatted and still keeps getting them.
TechReport's SSD Endurance test (techreport.com/review/27909/the-ssd-endurance-experiment-theyre-all-dead/) is probably the biggest one of the sort and these seem to show manufacturers' numbers in a good enough light. Also keep in mind that these are 2nd gen SSDs and everyone has much much more experience with everything involved.
efikkanWhile I can't claim to have the empirical data to support a scientific conclusion, the trend I'm seeing is really concerning. In the past couple of years I've seen a decent number of SSDs which become very unstable after 2-3 years of use, and most of them nowhere near their endurance rating. I've seen a lot of HDDs go bad in the last 25 years, but nothing close to this failure rate.
I have seen only a couple SSDs fail but with the amount of SSDs I have had or deployed, the defect rate is actually lower than with HDDs. Granted, I tend to not use the cheapest SSDs and stick to the models I know - Intel in early days, Samsungs, Crucials with some Kingstons and Adatas sprinkled in. Reallocated sector count seems to be the one to watch for SSDs and get the drive replaced as soon as that starts to happen. Endurance rating as well of course but I do not have write-heavy use cases that would use up all of that in drive's lifetime. I have 10 years old SSDs that are still going strong without issues, mostly relegated to a desktop machine of family or friends due to something like 1st gen 60GB or 2nd gen 120GB SSD being pretty slow by today's standards.
efikkanThere are at least two major problems with SSDs; stability of the cells and data rot. Especially data rot gets a lot worse when they cram more bits into each cell. I wish SSD specs would list what the controllers do to deal with data rot, if they do anything at all.
Data rot is definitely a concern. From what I have seen tested this is a problem primarily with unpowered drives (where data rots relatively quickly) and powered archive or NAS type use where there is a suspicion data could rot unnoticed but there seems to be little research into that. If you have good sources, links would be appreciated.
Posted on Reply
#40
InVasMani
londisteIt gets offtopic but I am seeing less and less need for big computer these days. For a fully functional computer, put CPU with cooler on the motherboard, add RAM and M.2 drive or two and the only cables you need are 24-pin and 4/8 pin for power. If you need GPU, it goes to PCIe slot with possibly a couple power cables of its own and done. Very nice and clean.
You're overlooking all the case fans and wiring to them as well that can be about as messy or more than all the SATA cabling with HDD's/CD-ROMs of the past in it's own right. I say though that NVMe has helped a lot to eliminate a lot of messy case cable management which is great. I'd like to see HDD's that fit in PCIe slots and bus powered that utilize NVMe protocol and have more cache than current HDD's. Look at the Gigabyte I-RAM and it's easy to see how bad a bottleneck can be to the overall design fast memory with a slow interface is still slow no matter how often it bumps it's head against the wall blocking it.
Posted on Reply
#41
R-T-B
lexluthermiesterThat would only be useful if the durability is on par with native TLC/MLC/SLC tech.
I've read up on this a bit. The tl;dr basically is it does improve longevity but not nearly as much as being native TLC or MLC/SLC. It's not even close. It helps, but does not fix QLCs issues. The only way to fix QLC is to ramp up process node to a larger size, which really, defeats the point altogether of saving space with it.
efikkanThose endurance ratings are probably as inflated as MTBF numbers for HDDs.
They aren't. They are pretty defined numbers. "No longer working" is defined as when the cell no longer retains data properly, so a proper failure for the data block. This isn't like BS MTBF on SSDs, this is the true rating of how many times you can program a cell on the chip on average, and it's usually got testing to back it.
Posted on Reply
#42
lexluthermiester
R-T-BI've read up on this a bit. The tl;dr basically is it does improve longevity but not nearly as much as being native TLC or MLC/SLC. It's not even close. It helps, but does not fix QLCs issues. The only way to fix QLC is to ramp up process node to a larger size, which really, defeats the point altogether of saving space with it.
And that's more or less what I've been reading elsewhere. Just not worth it.
Posted on Reply
#43
Scrizz
Tomorrow14nm++++++++++++++++++ :D

That's a good price for 660p. Unfortunately the prices have risen considerably making QLC all but useless. 660p costs ~250€ for 2TB and 665p costs nearly ~450€. Jesus F christ. For 450 i can buy a 2TB TLC based PCIe 4.0 drive. Crucial P1 seems to be all but gone looking at prices.

The cheapest ones are the slower SATA based Samsung 860 QVO and ADATA SU630 around 200€ for 2TB. Still too much.
SATA based 2TB QLC should be 150€ max and PCIe versions should be around 200€ like they were last year.
Where do you see those prices?
The retailers are ripping you off.
I see $239 and $369 respectively.That 665p price is a bit high doesn't make sense when you look at the 1TB being $139
www.newegg.com/intel-660p-series-2tb/p/N82E16820167461
www.newegg.com/intel-665p-series-2tb/p/N82E16820167470
Posted on Reply
#44
Tomorrow
ScrizzWhere do you see those prices?
The retailers are ripping you off.
I see $239 and $369 respectively.That 665p price is a bit high doesn't make sense when you look at the 1TB being $139
www.newegg.com/intel-660p-series-2tb/p/N82E16820167461
www.newegg.com/intel-665p-series-2tb/p/N82E16820167470
geizhals.eu/intel-ssd-660p-2tb-ssdpeknw020t801-ssdpeknw020t8xt-ssdpeknw020t8x1-a1859205.html

US pricing is meaningless to me as it does not include VAT and shipping costs.
But still $239 is too much for 660p.
Posted on Reply
#45
user112
optane as a caching technology is worthless but as a swap file/partition it works nicely. to the point I think once next gen games launch that can take advantage of fast storage 3DXpoint will be a life saver in that use case. the key thing is that it be cheaper than high capacity ram and I have no doubt it will be cheaper than 64GB of DDR5 for the 1st few years.
Posted on Reply
Add your own comment
Oct 25th, 2024 19:58 EDT change timezone

New Forum Posts

Popular Reviews

Controversial News Posts