• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.

Neuchips to Showcase Industry-Leading Gen AI Inferencing Accelerators at CES 2024

TheLostSwede

News Editor
Joined
Nov 11, 2004
Messages
18,010 (2.44/day)
Location
Sweden
System Name Overlord Mk MLI
Processor AMD Ryzen 7 7800X3D
Motherboard Gigabyte X670E Aorus Master
Cooling Noctua NH-D15 SE with offsets
Memory 32GB Team T-Create Expert DDR5 6000 MHz @ CL30-34-34-68
Video Card(s) Gainward GeForce RTX 4080 Phantom GS
Storage 1TB Solidigm P44 Pro, 2 TB Corsair MP600 Pro, 2TB Kingston KC3000
Display(s) Acer XV272K LVbmiipruzx 4K@160Hz
Case Fractal Design Torrent Compact
Audio Device(s) Corsair Virtuoso SE
Power Supply be quiet! Pure Power 12 M 850 W
Mouse Logitech G502 Lightspeed
Keyboard Corsair K70 Max
Software Windows 10 Pro
Benchmark Scores https://valid.x86.fr/yfsd9w
Neuchips, a leading AI Application-Specific Integrated Circuits (ASIC) solutions provider, will demo its revolutionary Raptor Gen AI accelerator chip (previously named N3000) and Evo PCIe accelerator card LLM solutions at CES 2024. Raptor, the new chip solution, enables enterprises to deploy large language models (LLMs) inference at a fraction of the cost of existing solutions.

"We are thrilled to unveil our Raptor chip and Evo card to the industry at CES 2024," said Ken Lau, CEO of Neuchips. "Neuchips' solutions represent a massive leap in price to performance for natural language processing. With Neuchips, any organisation can now access the power of LLMs for a wide range of AI applications."




Democratising Access to LLMs
Together, Raptor and Evo provide an optimised stack that makes market-leading LLMs readily accessible for enterprises. Neuchips' AI solutions significantly reduce hardware costs compared to existing solutions. The high energy efficiency also minimizes electricity usage, further lowering the total cost of ownership.

At CES 2024, Neuchips will demo Raptor and Evo, accelerating the Whisper and Llama AI chatbots on a Personal AI Assistant application. This solution highlights the power of LLM inferencing for real business needs.

Enterprises interested in test-driving Neuchips' breakthrough performance can visit booth 62700 to enrol in a free trial program. Additional technical sessions will showcase how Raptor and Evo can slash deployment costs for speech-to-text applications.

Raptor Gen AI Accelerator Powers Breakthrough LLM Performance
The Raptor chip delivers up to 200 tera operations per second (TOPS) per chip. Its outstanding performance for AI inferencing operations such as Matrix Multiply, Vector, and embedding table lookup suits Gen-AI and transformer-based AI models. This groundbreaking throughput is achieved via Neuchips' patented compression and efficiency optimisations tailored to neural networks.

Evo Gen 5 PCIe Card Sets New Standard for Acceleration and Low Power Consumption
Complementing Raptor is Neuchips' ultra-low powered Evo acceleration card. Evo combines PCIe Gen 5 with eight lanes and LPDDR5 32 GB to achieve 64 GB/s host I/O bandwidth and 1.6-Tbps per second of memory bandwidth at just 55 watts per card.

As demonstrated with DLRM, Evo also features 100% scalability, allowing customers to linearly increase performance by adding more chips. This modular design ensures investment protection for future AI workloads.

An upcoming half-height half-length (HHHL) form factor product, Viper, set to be launched by the second half of 2024, will provide even greater deployment flexibility. The new series brings data centre-class AI acceleration in a compact design.

View at TechPowerUp Main Site | Source
 
Joined
Aug 23, 2013
Messages
593 (0.14/day)
So begins the return of Nvidia back to gamers. Imagine Jensen, always thinking he's found his niche that isn't graphics and always being dragged kicking and screaming back to gaming every time. Poor AMD only getting into AI acceleration as it goes back home to custom solutions. Lisa's a day late and a dollar short.

Let them make graphics cards and be content with only most of the money instead of all the money.
 
Joined
Sep 2, 2014
Messages
672 (0.18/day)
Location
Scotland
Processor 5800x
Motherboard b550-e
Cooling full - custom liquid loop
Memory cl16 - 32gb
Video Card(s) 6800xt
Storage nvme 1TB + ssd 750gb
Display(s) xg32vc
Case hyte y60
Power Supply 1000W - gold
Software 10
So begins the return of Nvidia back to gamers. Imagine Jensen, always thinking he's found his niche that isn't graphics and always being dragged kicking and screaming back to gaming every time. Poor AMD only getting into AI acceleration as it goes back home to custom solutions. Lisa's a day late and a dollar short.

Let them make graphics cards and be content with only most of the money instead of all the money.
Are you drunk or smoking some s**t? :) looks like you have a poor connection in your CAVE with the world :)
 

silentbogo

Moderator
Staff member
Joined
Nov 20, 2013
Messages
5,594 (1.37/day)
Location
Kyiv, Ukraine
System Name WS#1337
Processor Ryzen 7 5700X3D
Motherboard ASUS X570-PLUS TUF Gaming
Cooling Xigmatek Scylla 240mm AIO
Memory 64GB DDR4-3600(4x16)
Video Card(s) MSI RTX 3070 Gaming X Trio
Storage ADATA Legend 2TB
Display(s) Samsung Viewfinity Ultra S6 (34" UW)
Case ghetto CM Cosmos RC-1000
Audio Device(s) ALC1220
Power Supply SeaSonic SSR-550FX (80+ GOLD)
Mouse Logitech G603
Keyboard Modecom Volcano Blade (Kailh choc LP)
VR HMD Google dreamview headset(aka fancy cardboard)
Software Windows 11, Ubuntu 24.04 LTS
So begins the return of Nvidia back to gamers. Imagine Jensen, always thinking he's found his niche that isn't graphics and always being dragged kicking and screaming back to gaming every time. Poor AMD only getting into AI acceleration as it goes back home to custom solutions. Lisa's a day late and a dollar short.

Let them make graphics cards and be content with only most of the money instead of all the money.
Don't get too excited. Better read up on AI in spare time.
This is an "inference" hardware, not the "training" hardware. I doubt it'll even make sense to develop an ASIC specifically for training, cause it's supposed to be flexible by design.
NVidia doesn't even make that much of their money off edge AI devices and inference hardware(excluding GPUs). Jettson boards are niche dev. kits, which NVidia can't even produce in numbers. And all of their post-Mellanox stuff is even more of a niche-of-a-niche. Drive PX and CX aren't that hyped up anymore... and as far as I know Tesla dropped it awhile ago, while Mercedes and Toyota either gave up on it, or waiting for Tesla to pave the road and hit all bumps along the way for self-driving and assistive driving regulation(or as it usually goes with Musk - f it up completely).
I think THE biggest reason their ARM acquisition got blocked, is to prevent any possibility of NV creating a monopoly in inference hardware.

I was wondering when the ASIC makers were going to get on board for this cash cow and shake things up a bit.
There are already quite a few AI ASICs on the market, under a bunch of different "catchy" names, like IPU(Inference Processing Unit), VPU(Visual Processing Unit) etc. etc. etc.
Heck, your shiny new flagship phones all have those in them. Even without ASICs and GPUs, you can do it on other commodity hardware or Raspberry Pi, or even on a microcontroller, depending on tasks and performance requirements.
 

Solaris17

Super Dainty Moderator
Staff member
Joined
Aug 16, 2005
Messages
27,186 (3.83/day)
Location
Alabama
System Name RogueOne
Processor Xeon W9-3495x
Motherboard ASUS w790E Sage SE
Cooling SilverStone XE360-4677
Memory 128gb Gskill Zeta R5 DDR5 RDIMMs
Video Card(s) MSI SUPRIM Liquid X 4090
Storage 1x 2TB WD SN850X | 2x 8TB GAMMIX S70
Display(s) 49" Philips Evnia OLED (49M2C8900)
Case Thermaltake Core P3 Pro Snow
Audio Device(s) Moondrop S8's on schitt Gunnr
Power Supply Seasonic Prime TX-1600
Mouse Razer Viper mini signature edition (mercury white)
Keyboard Monsgeek M3 Lavender, Moondrop Luna lights
VR HMD Quest 3
Software Windows 11 Pro Workstation
Benchmark Scores I dont have time for that.
Heck, your shiny new flagship phones all have those in them. Even without ASICs and GPUs, you can do it on other commodity hardware or Raspberry Pi, or even on a microcontroller, depending on tasks and performance requirements.

I literally have like a 4tops NPU I bought in like 2019
 
Joined
Jun 29, 2023
Messages
107 (0.19/day)
At the lowest level, AI operations are simple math operations on huge multidimentional matrices and vectors. It's not the hardest stuff it seems, and that's why everyone and his uncle seem to come with specialized circuitry to do just that.

But the real important stuff is the higher level software that makes use of this, and with CUDA, Nvidia has a large head start.
 

Solaris17

Super Dainty Moderator
Staff member
Joined
Aug 16, 2005
Messages
27,186 (3.83/day)
Location
Alabama
System Name RogueOne
Processor Xeon W9-3495x
Motherboard ASUS w790E Sage SE
Cooling SilverStone XE360-4677
Memory 128gb Gskill Zeta R5 DDR5 RDIMMs
Video Card(s) MSI SUPRIM Liquid X 4090
Storage 1x 2TB WD SN850X | 2x 8TB GAMMIX S70
Display(s) 49" Philips Evnia OLED (49M2C8900)
Case Thermaltake Core P3 Pro Snow
Audio Device(s) Moondrop S8's on schitt Gunnr
Power Supply Seasonic Prime TX-1600
Mouse Razer Viper mini signature edition (mercury white)
Keyboard Monsgeek M3 Lavender, Moondrop Luna lights
VR HMD Quest 3
Software Windows 11 Pro Workstation
Benchmark Scores I dont have time for that.
Joined
Jul 5, 2013
Messages
28,613 (6.79/day)

It sits in a closet now, but I did a ton of object identification training on this thing.
Ah, ok. One of those is "slow" by today's standards, but still useful. It also makes for a great Android device. Granted, it's only Android 9, but it's still updated. That little board would make for a solid little AI platform with the right software on Ubuntu or Armbian.
 

Solaris17

Super Dainty Moderator
Staff member
Joined
Aug 16, 2005
Messages
27,186 (3.83/day)
Location
Alabama
System Name RogueOne
Processor Xeon W9-3495x
Motherboard ASUS w790E Sage SE
Cooling SilverStone XE360-4677
Memory 128gb Gskill Zeta R5 DDR5 RDIMMs
Video Card(s) MSI SUPRIM Liquid X 4090
Storage 1x 2TB WD SN850X | 2x 8TB GAMMIX S70
Display(s) 49" Philips Evnia OLED (49M2C8900)
Case Thermaltake Core P3 Pro Snow
Audio Device(s) Moondrop S8's on schitt Gunnr
Power Supply Seasonic Prime TX-1600
Mouse Razer Viper mini signature edition (mercury white)
Keyboard Monsgeek M3 Lavender, Moondrop Luna lights
VR HMD Quest 3
Software Windows 11 Pro Workstation
Benchmark Scores I dont have time for that.
Ah, ok. One of those is "slow" by today's standards, but still useful. It also makes for a great Android device. Granted, it's only Android 9, but it's still updated. That little board would make for a solid little AI platform with the right software on Ubuntu or Armbian.

I ran ubuntu, but I was only in it for the NPU. It didnt do anything else.
 
Joined
Dec 29, 2010
Messages
3,820 (0.74/day)
Processor AMD 5900x
Motherboard Asus x570 Strix-E
Cooling Hardware Labs
Memory G.Skill 4000c17 2x16gb
Video Card(s) RTX 3090
Storage Sabrent
Display(s) Samsung G9
Case Phanteks 719
Audio Device(s) Fiio K5 Pro
Power Supply EVGA 1000 P2
Mouse Logitech G600
Keyboard Corsair K95
So begins the return of Nvidia back to gamers. Imagine Jensen, always thinking he's found his niche that isn't graphics and always being dragged kicking and screaming back to gaming every time. Poor AMD only getting into AI acceleration as it goes back home to custom solutions. Lisa's a day late and a dollar short.

Let them make graphics cards and be content with only most of the money instead of all the money.
Day late, lmao. AI is gonna be around for the rest of humanity, that is until it kills us lol. AMD missed out on the fleecing of the market. ANd the real pain is Nvidia, knowing that they had a captive market but instead of growing it organically they went the route of fleecing the fuck out of the market. It is so bad that the top 3 major AI players are all well on their way to custom chips, with Google on their 4th gen and Tesla 2nd gen. No company in their right mind is ok with feeding a suppliers gross margin of 75%. Thus Nvidia killed their own dominance, that's the thing you seem to be missing, smh.
 

silentbogo

Moderator
Staff member
Joined
Nov 20, 2013
Messages
5,594 (1.37/day)
Location
Kyiv, Ukraine
System Name WS#1337
Processor Ryzen 7 5700X3D
Motherboard ASUS X570-PLUS TUF Gaming
Cooling Xigmatek Scylla 240mm AIO
Memory 64GB DDR4-3600(4x16)
Video Card(s) MSI RTX 3070 Gaming X Trio
Storage ADATA Legend 2TB
Display(s) Samsung Viewfinity Ultra S6 (34" UW)
Case ghetto CM Cosmos RC-1000
Audio Device(s) ALC1220
Power Supply SeaSonic SSR-550FX (80+ GOLD)
Mouse Logitech G603
Keyboard Modecom Volcano Blade (Kailh choc LP)
VR HMD Google dreamview headset(aka fancy cardboard)
Software Windows 11, Ubuntu 24.04 LTS
It sits in a closet now, but I did a ton of object identification training on this thing.
Haven't had a chance to even see those in person, let alone touch and break one :D
I'd gladly take one off your hands for a worthy cause :toast:
 
Joined
Mar 29, 2014
Messages
512 (0.13/day)
Day late, lmao. AI is gonna be around for the rest of humanity, that is until it kills us lol. AMD missed out on the fleecing of the market. ANd the real pain is Nvidia, knowing that they had a captive market but instead of growing it organically they went the route of fleecing the fuck out of the market. It is so bad that the top 3 major AI players are all well on their way to custom chips, with Google on their 4th gen and Tesla 2nd gen. No company in their right mind is ok with feeding a suppliers gross margin of 75%. Thus Nvidia killed their own dominance, that's the thing you seem to be missing, smh.
They didn't miss out, they elected not to go that route. ;)
 
Top