News Posts matching #AI

Return to Keyword Browsing

AMD "Strix Halo" Zen 5 Mobile Processor Pictured: Chiplet-based, Uses 256-bit LPDDR5X

Enthusiasts on the ChipHell forum scored an alleged image of AMD's upcoming "Strix Halo" mobile processor, and set out to create some highly plausible schematic slides. These are speculative. While "Strix Point" is the mobile processor that succeeds the current "Hawk Point" and "Phoenix" processors; "Strix Halo" is in a category of its own—to offer gaming experiences comparable to discrete GPUs in the ultraportable form-factor where powerful discrete GPUs are generally not possible. "Strix Halo" also goes head on against Apple's M3 Max and M3 Pro processors powering the latest crop of MacBook Pros. It has the same advantages as a single-chip solution, as the M3 Max.

The "Strix Halo" silicon is a chiplet-based processor, although very different from "Fire Range". The "Fire Range" processor is essentially a BGA version of the desktop "Granite Ridge" processor—it's the same combination of one or two "Zen 5" CCDs that talk to a client I/O die, and is meant for performance-thru-enthusiast segment notebooks. "Strix Halo," on the other hand, use the same one or two "Zen 5" CCDs, but with a large SoC die featuring an oversized iGPU, and 256-bit LPDDR5X memory controllers not found on the cIOD. This is key to what AMD is trying to achieve—CPU and graphics performance in the league of the M3 Pro and M3 Max at comparable PCB and power footprints.

Acer Expands Chromebook Plus Laptop Lineup with New 14-Inch Model Powered by Intel Core Processors

Acer today expanded its line of Chromebook Plus laptops with the Acer Chromebook Plus 514 (CB514-4H/T), providing users with a performance-minded, compact and durable model that enables them to do more with the AI-powered capabilities of ChromeOS. "The new Acer Chromebook Plus 514 (CB514-4H/T) delivers the sought-after combination of a portable design, 14-inch Full HD display and performance-minded technology that lets users get the most out of exciting capabilities offered with Chromebook Plus," said James Lin, General Manager, Notebooks, Acer Inc. "Students, businesses, families, and individuals need to be more productive, connected and empowered than ever, and can achieve this using Acer Chromebook Plus devices."

The new Acer Chromebook Plus 514 is the latest addition to Acer's lineup of Chromebook Plus laptops that offer enhanced Chromebook performance and experiences, emphasizing better hardware designs with upgraded displays and cameras paired with powerful productivity, creativity, and multimedia capabilities. Like all Acer Chromebook Plus laptops, users have the power to do more with the new Chromebook Plus 514 (CB514-4H/T). Powered by an Intel Core i3-N305 processor and an ample 8 GB of LPDDR5 RAM, the Acer Chromebook Plus 514 provides 2x the speed, memory, and storage, giving responsive performance and efficient multitasking, whether running built-in AI-powered apps like Google Docs and Photos, watching favorite shows in full HD on a 1080p display, or movie-making with LumaFusion. Plus, the processor ensures all-day enjoyment with up to 11 hours of usage on the fast-charging battery.

Intel Builds World's Largest Neuromorphic System to Enable More Sustainable AI

Today, Intel announced that it has built the world's largest neuromorphic system. Code-named Hala Point, this large-scale neuromorphic system, initially deployed at Sandia National Laboratories, utilizes Intel's Loihi 2 processor, aims at supporting research for future brain-inspired artificial intelligence (AI), and tackles challenges related to the efficiency and sustainability of today's AI. Hala Point advances Intel's first-generation large-scale research system, Pohoiki Springs, with architectural improvements to achieve over 10 times more neuron capacity and up to 12 times higher performance.

"The computing cost of today's AI models is rising at unsustainable rates. The industry needs fundamentally new approaches capable of scaling. For that reason, we developed Hala Point, which combines deep learning efficiency with novel brain-inspired learning and optimization capabilities. We hope that research with Hala Point will advance the efficiency and adaptability of large-scale AI technology." -Mike Davies, director of the Neuromorphic Computing Lab at Intel Labs

Lenovo Prepares Thinkpad T14s and Yoga Slim 14 Laptops with Qualcomm Snapdragon X Processor

Lenovo is putting finishing touches on the Yoga Slim 7 14 2024, one of the first non-reference laptop to feature Qualcomm's latest Snapdragon X processor. Leaked images circulating on X (formerly Twitter) reveal a sleek and stylish design, with a 14-inch or 14.5-inch display encased in a slim and portable form factor. Qualcomm has previously showcased eye-catching demo reference systems in a striking red color scheme, but Lenovo's Yoga Slim 7 14 2024 marks the first time a significant laptop manufacturer with actual product images incorporating the Snapdragon X chip. The Yoga Slim 7 14 2024 is part of Lenovo's popular Slim laptop lineup, including models powered by Intel and AMD processors. The latest "Gen 8" iteration featured options for AMD Ryzen 7040 series and Intel 13th Gen Core i and Core Ultra series CPUs.

One notable addition to the Snapdragon X-powered model is the inclusion of a dedicated Microsoft Copilot button. Qualcomm has heavily touted the Snapdragon X's Neural Processing Unit (NPU) performance and its ability to directly accelerate various AI and machine learning algorithms on the device. There have been a few comparison points between Meteor Lake with Intel's NPU and Snapdragon X Elite with Qualcomm's NPU. The chipmaker's X Elite and X Plus laptop offerings are expected to arrive soon, and there are strong indications that this may happen during the Computex trade show.

Update 17:28 UTC: X user WalkingCat has posted images of Lenovo Thinkpad T14s laptop, which can be seen below.

NVIDIA Launches the RTX A400 and A1000 Professional Graphics Cards

AI integration across design and productivity applications is becoming the new standard, fueling demand for advanced computing performance. This means professionals and creatives will need to tap into increased compute power, regardless of the scale, complexity or scope of their projects. To meet this growing need, NVIDIA is expanding its RTX professional graphics offerings with two new NVIDIA Ampere architecture-based GPUs for desktops: the NVIDIA RTX A400 and NVIDIA RTX A1000.

They expand access to AI and ray tracing technology, equipping professionals with the tools they need to transform their daily workflows. The RTX A400 GPU introduces accelerated ray tracing and AI to the RTX 400 series GPUs. With 24 Tensor Cores for AI processing, it surpasses traditional CPU-based solutions, enabling professionals to run cutting-edge AI applications, such as intelligent chatbots and copilots, directly on their desktops. The GPU delivers real-time ray tracing, so creators can build vivid, physically accurate 3D renders that push the boundaries of creativity and realism.

AMD Expands Commercial AI PC Portfolio to Deliver Leadership Performance Across Professional Mobile and Desktop Systems

Today, AMD announced new products that will expand its commercial mobile and desktop AI PC portfolio, delivering exceptional productivity and premium AI and connectivity experiences to business users. The new AMD Ryzen PRO 8040 Series are the most advanced x86 processors built for business laptops and mobile workstations. In addition, AMD also announced the AMD Ryzen PRO 8000 Series desktop processor, the first AI enabled desktop processor for business users, engineered to deliver cutting-edge performance with low power consumption.

With AMD Ryzen AI built into select models, AMD is further extending its AI PC leadership. By leveraging the CPU, GPU, and dedicated on-chip neural processing unit (NPU), new Ryzen AI-powered processors provide more dedicated AI processing power than previous generations, with up to 16 dedicated NPU TOPS (Trillions of Operations Per Second) and up to 39 total system TOPS. Commercial PCs equipped with new Ryzen AI-enabled processors will help transform user experience, offering next-gen performance for AI-enabled collaboration, content creation, and data and analytics workloads. With the addition of AMD PRO technologies, IT managers can unlock enterprise-grade manageability features to simplify IT operations and complete PC deployment faster across the organization, built-in security features for chip-to-cloud defense from sophisticated attacks, as well as unprecedented stability, reliability and platform longevity for enterprise software.

Lenovo Embraces the AI PC Era with New ThinkCentre Desktops Powered by AMD Ryzen PRO 8000 Series Desktop Processors

Lenovo has unveiled a selection of ThinkCentre desktops powered by AMD Ryzen PRO 8000 Series desktop processors with up to 16 TOPS (trillion operations per second) of integrated NPU capability dedicated to process AI workloads, including the performance focused ThinkCentre M75t Gen 5, the flexible ThinkCentre M75s Gen 5, and the compact ThinkCentre M75q Gen 5. Designed to meet the diverse needs of modern business, the ThinkCentre M75 Gen 5 family of desktops harnesses the AI capability of its component while optimizing its energy efficiency to deliver impressive results.

"The AI PC era is already here and at Lenovo we are embracing it to unlock new possibilities," said Sanjeev Menon, vice president and general manager, Worldwide Desktop Business in Intelligent Devices Group, Lenovo. "The need for businesses to integrate AI into their operations continues to grow and our ThinkCentre M75 family of desktops, with a strong and stable power supply, the ability to upgrade components when needed, and the space to expand memory and optimize thermal management are the ideal options to enhance productivity with AI without heavy investments. Lenovo and AMD have a long-standing partnership focused on delivering value to our customers and we know users will be delighted by the leap in performance of our new desktops."

Demand for NVIDIA's Blackwell Platform Expected to Boost TSMC's CoWoS Total Capacity by Over 150% in 2024

NVIDIA's next-gen Blackwell platform, which includes B-series GPUs and integrates NVIDIA's own Grace Arm CPU in models such as the GB200, represents a significant development. TrendForce points out that the GB200 and its predecessor, the GH200, both feature a combined CPU+GPU solution, primarily equipped with the NVIDIA Grace CPU and H200 GPU. However, the GH200 accounted for only approximately 5% of NVIDIA's high-end GPU shipments. The supply chain has high expectations for the GB200, with projections suggesting that its shipments could exceed millions of units by 2025, potentially making up nearly 40 to 50% of NVIDIA's high-end GPU market.

Although NVIDIA plans to launch products such as the GB200 and B100 in the second half of this year, upstream wafer packaging will need to adopt more complex and high-precision CoWoS-L technology, making the validation and testing process time-consuming. Additionally, more time will be required to optimize the B-series for AI server systems in aspects such as network communication and cooling performance. It is anticipated that the GB200 and B100 products will not see significant production volumes until 4Q24 or 1Q25.

Sony PlayStation 5 Pro Specifications Confirmed, Console Arrives Before Holidays

Thanks for the detailed information obtained by The Verge, today we confirm previously leaked details as Sony gears up to unveil the highly anticipated PlayStation 5 Pro, codenamed "Trinity." According to insider reports, Sony is urging developers to optimize their games for the PS5 Pro, with a primary focus on enhancing ray tracing capabilities. The console is expected to feature an RDNA 3 GPU with 30 WGP running BVH8, capable of 33.5 TeraFLOPS of FP32 single-precision computing power, and a slightly quicker CPU running at 3.85 GHz, enabling it to render games with ray tracing enabled or achieve higher resolutions and frame rates in select titles. Sony anticipates GPU rendering on the PS5 Pro to be approximately 45 percent faster than the standard PlayStation 5. The PS5 Pro GPU will be larger and utilize faster system memory to bolster ray tracing performance, boasting up to three times the speed of the regular PS5.

Additionally, the console will employ a more powerful ray tracing architecture, backed by PlayStation Spectral Super Resolution (PSSR), allowing developers to leverage graphics features like ray tracing more extensively. To support this endeavor, Sony is providing developers with test kits, and all games submitted for certification from August onward must be compatible with the PS5 Pro. Insider Gaming, the first to report the full PS5 Pro specs, suggests a potential release during the 2024 holiday period. The PS5 Pro will also feature modifications for developers regarding system memory, with Sony increasing the memory bandwidth from 448 GB/s to 576 GB/s, enhancing efficiency for an even more immersive gaming experience. To do AI processing, there is an custom AI accelerator capable of 300 8-bit INT8 TOPS and 67 16-bit FP16 TeraFLOPS, in addition to ACV audio codec running up to 35% faster.

Apple Preparing M4 Chips with AI Capabilities to Fight Declining Mac Sales

While everyone has been focused on shipping an AI-enhanced product recently, one tech giant didn't appear to be bothered- Apple. However, according to Mark Gurman from Bloomberg, Apple is readying an overhaul of its Apple Silicon M-series chips to embed AI processing capabilities at the processor level. As the report indicates, Apple is preparing an update for late 2024 and early 2025 with the M4 series of chips, which will reportedly feature AI processing units similar to those found in other commercial chips. There should be three levels of the M4 series, with the entry-level M4 codenamed Donan, the mid-level M4 chip codenamed Brava, and the high-end M4 chip codenamed Hydra.

Sales of Apple Macs peaked in 2022; the following year was a sharp decline, and sales have continued to be flat since. The new AI PCs for Windows-based systems have been generating hype from all major vendors, hoping to introduce AI features to end users. However, Apple wants to be part of the revolution, and the company has already scheduled the World Wide Developer Conference for June 10th. At WWDC this year, Apple is supposed to show a suite of AI-powered solutions to enable better user experience and increase productivity. With M4 chips getting AI enhancement, the WWDC announcements will get extra hardware accelerations. However, we must wait for the exact announcements before making further assumptions.

Meta Announces New MTIA AI Accelerator with Improved Performance to Ease NVIDIA's Grip

Meta has announced the next generation of its Meta Training and Inference Accelerator (MTIA) chip, which is designed to train and infer AI models at scale. The newest MTIA chip is a second-generation design of Meta's custom silicon for AI, and it is being built on TSMC's 5 nm technology. Running at the frequency of 1.35 GHz, the new chip is getting a boost to 90 Watts of TDP per package compared to just 25 Watts for the first-generation design. Basic Linear Algebra Subprograms (BLAS) processing is where the chip shines, and it includes matrix multiplication and vector/SIMD processing. At GEMM matrix processing, each chip can process 708 TeraFLOPS at INT8 (presumably meant FP8 in the spec) with sparsity, 354 TeraFLOPS without, 354 TeraFLOPS at FP16/BF16 with sparsity, and 177 TeraFLOPS without.

Classical vector and processing is a bit slower at 11.06 TeraFLOPS at INT8 (FP8), 5.53 TeraFLOPS at FP16/BF16, and 2.76 TFLOPS single-precision FP32. The MTIA chip is specifically designed to run AI training and inference on Meta's PyTorch AI framework, with an open-source Triton backend that produces compiler code for optimal performance. Meta uses this for all its Llama models, and with Llama3 just around the corner, it could be trained on these chips. To package it into a system, Meta puts two of these chips onto a board and pairs them with 128 GB of LPDDR5 memory. The board is connected via PCIe Gen 5 to a system where 12 boards are stacked densely. This process is repeated six times in a single rack for 72 boards and 144 chips in a single rack for a total of 101.95 PetaFLOPS, assuming linear scaling at INT8 (FP8) precision. Of course, linear scaling is not quite possible in scale-out systems, which could bring it down to under 100 PetaFLOPS per rack.
Below, you can see images of the chip floorplan, specifications compared to the prior version, as well as the system.

ASUS IoT Announces PE8000G

ASUS IoT, the global AIoT solution provider, today announced PE8000G at Embedded World 2024, a powerful edge AI computer that supports multiple GPU cards for high performance—and is expertly engineered to handle rugged conditions with resistance to extreme temperatures, vibration and variable voltage. PE8000G is powered by formidable Intel Core processors (13th and 12th gen) and the Intel R680E chipset to deliver high-octane processing power and efficiency.

With its advanced architecture, PE8000G excels at running multiple neural network modules simultaneously in real-time—and represents a significant leap forward in edge AI computing. With its robust design, exceptional performance and wide range of features, PE8000G series is poised to revolutionize AI-driven applications across multiple industries, elevating edge AI computing to new heights and enabling organizations to tackle mission-critical tasks with confidence and to achieve unprecedented levels of productivity and innovation.

Report: Global PC Shipments Return to Growth and Pre-Pandemic Volumes in the First Quarter of 2024

After two years of decline, the worldwide traditional PC market returned to growth during the first quarter of 2024 (1Q24) with 59.8 million shipments, growing 1.5% year over year, according to preliminary results from the International Data Corporation (IDC) Worldwide Quarterly Personal Computing Device Tracker. Growth was largely achieved due to easy year-over-year comparisons as the market declined 28.7% during the first quarter of 2023, which was the lowest point in PC history. In addition, global PC shipments finally returned to pre-pandemic levels as 1Q24 volumes rivaled those seen in 1Q19 when 60.5 million units were shipped.

With inflation numbers trending down, PC shipments have begun to recover in most regions, leading to growth in the Americas as well as Europe, the Middle East, and Africa (EMEA). However, the deflationary pressures in China directly impacted the global PC market. As the largest consumer of desktop PCs, weak demand in China led to yet another quarter of declines for global desktop shipments, which already faced pressure from notebooks as the preferred form factor.

Google Launches Axion Arm-based CPU for Data Center and Cloud

Google has officially joined the club of custom Arm-based, in-house-developed CPUs. As of today, Google's in-house semiconductor development team has launched the "Axion" CPU based on Arm instruction set architecture. Using the Arm Neoverse V2 cores, Google claims that the Axion CPU outperforms general-purpose Arm chips by 30% and Intel's processors by a staggering 50% in terms of performance. This custom silicon will fuel various Google Cloud offerings, including Compute Engine, Kubernetes Engine, Dataproc, Dataflow, and Cloud Batch. The Axion CPU, designed from the ground up, will initially support Google's AI-driven services like YouTube ads and Google Earth Engine. According to Mark Lohmeyer, Google Cloud's VP and GM of compute and machine learning infrastructure, Axion will soon be available to cloud customers, enabling them to leverage its performance without overhauling their existing applications.

Google's foray into custom silicon aligns with the strategies of its cloud rivals, Microsoft and Amazon. Microsoft recently unveiled its own AI chip for training large language models and an Arm-based CPU called Cobalt 100 for cloud and AI workloads. Amazon, on the other hand, has been offering Arm-based servers through its custom Graviton CPUs for several years. While Google won't sell these chips directly to customers, it plans to make them available through its cloud services, enabling businesses to rent and leverage their capabilities. As Amin Vahdat, the executive overseeing Google's in-house chip operations, stated, "Becoming a great hardware company is very different from becoming a great cloud company or a great organizer of the world's information."

Micron Debuts World's First Quad-Port SSD to Accelerate Data-Rich Autonomous and AI-Enabled Workloads

Micron Technology, Inc., today announced it is sampling the automotive-grade Micron 4150AT SSD, the world's first quad-port SSD, capable of interfacing with up to four systems on chips (SoCs) to centralize storage for software-defined intelligent vehicles. The Micron 4150AT SSD combines market-leading features such as single-root input/output virtualization (SR-IOV), a PCIe Generation 4 interface and ruggedized automotive design. With these features, the automotive-grade SSD provides the ecosystem with data center-level flexibility and power.

"As storage requirements race to keep up with rich in-vehicle experiences featuring AI and advanced algorithms for higher levels of autonomous safety, this era demands a new paradigm for automotive storage to match," said Michael Basca, Micron vice president of embedded products and systems. "Building on our collaboration with the innovators redefining next-generation automotive architectures, Micron has reimagined storage from the ground up to deliver the world's first quad-port SSD - the Micron 4150AT - which provides the industry flexibility and horsepower to roll out the transformative technologies on the horizon."

Intel Launches Gaudi 3 AI Accelerator: 70% Faster Training, 50% Faster Inference Compared to NVIDIA H100, Promises Better Efficiency Too

During the Vision 2024 event, Intel announced its latest Gaudi 3 AI accelerator, promising significant improvements over its predecessor. Intel claims the Gaudi 3 offers up to 70% improvement in training performance, 50% better inference, and 40% better efficiency than Nvidia's H100 processors. The new AI accelerator is presented as a PCIe Gen 5 dual-slot add-in card with a 600 W TDP or an OAM module with 900 W. The PCIe card has the same peak 1,835 TeraFLOPS of FP8 performance as the OAM module despite a 300 W lower TDP. The PCIe version works as a group of four per system, while the OAM HL-325L modules can be run in an eight-accelerator configuration per server. This likely will result in a lower sustained performance, given the lower TDP, but it confirms that the same silicon is used, just finetuned with a lower frequency. Built on TSMC's N5 5 nm node, the AI accelerator features 64 Tensor Cores, delivering double the FP8 and quadruple FP16 performance over the previous generation Gaudi 2.

The Gaudi 3 AI chip comes with 128 GB of HBM2E with 3.7 TB/s of bandwidth and 24 200 Gbps Ethernet NICs, with dual 400 Gbps NICs used for scale-out. All of that is laid out on 10 tiles that make up the Gaudi 3 accelerator, which you can see pictured below. There is 96 MB of SRAM split between two compute tiles, which acts as a low-level cache that bridges data communication between Tensor Cores and HBM memory. Intel also announced support for the new performance-boosting standardized MXFP4 data format and is developing an AI NIC ASIC for Ultra Ethernet Consortium-compliant networking. The Gaudi 3 supports clusters of up to 8192 cards, coming from 1024 nodes comprised of systems with eight accelerators. It is on track for volume production in Q3, offering a cost-effective alternative to NVIDIA accelerators with the additional promise of a more open ecosystem. More information and a deeper dive can be found in the Gaudi 3 Whitepaper.

Intel Unleashes Enterprise AI with Gaudi 3, AI Open Systems Strategy and New Customer Wins

At the Intel Vision 2024 customer and partner conference, Intel introduced the Intel Gaudi 3 accelerator to bring performance, openness and choice to enterprise generative AI (GenAI), and unveiled a suite of new open scalable systems, next-gen products and strategic collaborations to accelerate GenAI adoption. With only 10% of enterprises successfully moving GenAI projects into production last year, Intel's latest offerings address the challenges businesses face in scaling AI initiatives.

"Innovation is advancing at an unprecedented pace, all enabled by silicon - and every company is quickly becoming an AI company," said Intel CEO Pat Gelsinger. "Intel is bringing AI everywhere across the enterprise, from the PC to the data center to the edge. Our latest Gaudi, Xeon and Core Ultra platforms are delivering a cohesive set of flexible solutions tailored to meet the changing needs of our customers and partners and capitalize on the immense opportunities ahead."

Acer Debuts New Predator Helios Neo 14; a Portable AI Gaming Laptop with Intel Core Ultra Processors

Acer unveiled the latest addition to its gaming line with the new Predator Helios Neo 14, a portable powerhouse equipped with the latest Intel Core Ultra processors, unlocking AI experiences on the PC. In addition, the Acer Nitro 16 gaming laptop has been refreshed with Intel Core 14th gen processors, allowing users to do more when gaming or creating. Players of all levels will be amazed by the pristine visuals and AI-driven graphics shown when gaming or streaming thanks to the devices' NVIDIA GeForce RTX 40[1] Series Laptop GPUs, which support the latest AI-powered DLSS 3.5 technology and other AI accelerations across creative applications. Brilliant colors and pixels stand out on the laptops' WQXGA display options, combined with 100% support for sRGB color gamut and NVIDIA G-SYNC compatibility.

To further complement creativity and capabilities on the gaming laptops, these feature Microsoft Copilot in Windows (with a dedicated Copilot key), providing faster access to the everyday AI assistant and Acer's suite of AI-enhanced videoconferencing features. With one month of Xbox Game Pass Ultimate included in every device, players can enjoy hundreds of high-quality PC games. Plus, dual fans, exhaust cooling systems, and liquid metal thermal on the CPUs ensure the gaming laptops' interiors run on full speed even in the most heated gaming sessions.

Acer Launches New Nitro 14 and Nitro 16 Gaming Laptops Powered by AMD Ryzen 8040 Series Processors

Acer today announced the new Nitro 14 and Nitro 16 gaming laptops, powered by AMD Ryzen 8040 Series processors with Ryzen AI[1]. With up to NVIDIA GeForce RTX 4060[2] Laptop GPUs supported by DLSS 3.5 technology, both are backed by NVIDIA's RTX AI platform, providing an array of capabilities in over 500 games and applications, enhanced by AI. Gamers are immersed in their 14- and 16-inch NVIDIA G-SYNC compatible panels with up to WQXGA (2560x1600) resolution.

Whether in call or streaming in-game, Acer PurifiedVoice 2.0 harnesses the power of AI to block out external noises, while Acer PurifiedView keeps users always front and center of all the action. Microsoft Copilot in Windows (with a dedicated Copilot key) helps accelerate everyday tasks on these AI laptops, and with one month of Xbox Game Pass Ultimate included with every device, players will enjoy hundreds of high-quality PC games. To seamlessly take command of device performance and customizations, one click of the NitroSense key directs users to the control center and the library of available AI-related functions through the new Experience Zone.

AMD Extends Leadership Adaptive SoC Portfolio with New Versal Series Gen 2 Devices Delivering End-to-End Acceleration for AI-Driven Embedded Systems

AMD today announced the expansion of the AMD Versal adaptive system on chip (SoC) portfolio with the new Versal AI Edge Series Gen 2 and Versal Prime Series Gen 2 adaptive SoCs, which bring preprocessing, AI inference, and postprocessing together in a single device for end-to-end acceleration of AI-driven embedded systems.

These initial devices in the Versal Series Gen 2 portfolio build on the first generation with powerful new AI Engines expected to deliver up to 3x higher TOPs-per-watt than first generation Versal AI Edge Series devicesi, while new high-performance integrated Arm CPUs are expected to offer up to 10x more scalar compute than first gen Versal AI Edge and Prime series devicesii.

Advantech Unveils Cutting-Edge GPU Card with Intel Arc A380E

Advantech (2395.TW), a global leader in intelligent IoT systems and embedded platforms, is excited to announce the EAI-3101, a brand-new industrial PCIe GPU card powered by the Intel Arc A380E, built for 5-year longevity. Featuring 128 Intel Xe matrix AI engines, this GPU card delivers outstanding AI computing power of 5.018 TFLOPS, surpassing the capabilities of the NVIDIA T1000, 2 times over. With ray tracing technology and Intel XeSS AI-upscaling, the EAI-3101 supports up to 8K UHD resolution and achieves a 50% enhancement in graphics performance over the NVIDIA T1000.

To aid in quickly realizing Vision AI, Advantech provides the Edge AI SDK, a rapid AI development toolkit compatible with Intel OpenVINO, which can process the same workload in 40% less time. This groundbreaking graphics solution, with optimized thermal design and an auto smart fan, is specially engineered for image processing and AI acceleration across gaming, medical analysis, and video surveillance. Advantech will demonstrate the EAI-3101 GPU card from April 9th to 11th at the Embedded World 2024 (Hall 3, booth no. 339) in Nuremberg, Germany.

AIO Workstation Combines 128-Core Arm Processor and Four NVIDIA GPUs Totaling 28,416 CUDA Cores

All-in-one computers are often traditionally seen as lower-powered alternatives to traditional desktop workstations. However, a new offering from Alafia AI, a startup focused on medical imaging appliances, aims to shatter that perception. The company's upcoming Alafia Aivas SuperWorkstation packs serious hardware muscle, demonstrating that all-in-one systems can match the performance of their more modular counterparts. At the heart of the Aivas SuperWorkstation lies a 128-core Ampere Altra processor, running at 3.0 GHz clock speed. This CPU is complemented by not one but three NVIDIA L4 GPUs for compute, and a single NVIDIA RTX 4000 Ada GPU for video output, delivering a combined 28,416 CUDA cores for accelerated parallel computing tasks. The system doesn't skimp on other components, either. It features a 4K touch display with up to 360 nits of brightness, an extensive 2 TB of DDR4 RAM, and storage options up to an 8 TB solid-state drive. This combination of cutting-edge CPU, GPU, memory, and storage is squarely aimed at the demands of medical imaging and AI development workloads.

The all-in-one form factor packs this incredible hardware into a sleek, purposefully designed clinical research appliance. While initially targeting software developers, Alafia AI hopes that institutions that can optimize their applications for the Arm architecture can eventually deploy the Aivas SuperWorkstation for production medical imaging workloads. The company is aiming for application integration in Q3 2024 and full ecosystem device integration by Q4 2024. With this powerful new offering, Alafia AI is challenging long-held assumptions about the performance limitations of all-in-one systems. The Aivas SuperWorkstation demonstrates that the right hardware choices can transform these compact form factors into true powerhouse workstations. Especially with a combined total output of three NVIDIA L4 compute units, alongside RTX 4000 Ada graphics card, the AIO is more powerful than some of the high-end desktop workstations.

X-Silicon Startup Wants to Combine RISC-V CPU, GPU, and NPU in a Single Processor

While we are all used to having a system with a CPU, GPU, and, recently, NPU—X-Silicon Inc. (XSi), a startup founded by former Silicon Valley veterans—has unveiled an interesting RISC-V processor that can simultaneously handle CPU, GPU, and NPU workloads in a chip. This innovative chip architecture, which will be open-source, aims to provide a flexible and efficient solution for a wide range of applications, including artificial intelligence, virtual reality, automotive systems, and IoT devices. The new microprocessor combines a RISC-V CPU core with vector capabilities and GPU acceleration into a single chip, creating a versatile all-in-one processor. By integrating the functionality of a CPU and GPU into a single core, X-Silicon's design offers several advantages over traditional architectures. The chip utilizes the open-source RISC-V instruction set architecture (ISA) for both CPU and GPU operations, running a single instruction stream. This approach promises lower memory footprint execution and improved efficiency, as there is no need to copy data between separate CPU and GPU memory spaces.

Called the C-GPU architecture, X-Silicon uses RISC-V Vector Core, which has 16 32-bit FPUs and a Scaler ALU for processing regular integers as well as floating point instructions. A unified instruction decoder feeds the cores, which are connected to a thread scheduler, texture unit, rasterizer, clipping engine, neural engine, and pixel processors. All is fed into a frame buffer, which feeds the video engine for video output. The setup of the cores allows the users to program each core individually for HPC, AI, video, or graphics workloads. Without software, there is no usable chip, which prompts X-Silicon to work on OpenGL ES, Vulkan, Mesa, and OpenCL APIs. Additionally, the company plans to release a hardware abstraction layer (HAL) for direct chip programming. According to Jon Peddie Research (JPR), the industry has been seeking an open-standard GPU that is flexible and scalable enough to support various markets. X-Silicon's CPU/GPU hybrid chip aims to address this need by providing manufacturers with a single, open-chip design that can handle any desired workload. The XSi gave no timeline, but it has plans to distribute the IP to OEMs and hyperscalers, so the first silicon is still away.

AMD Zen 5 Execution Engine Leaked, Features True 512-bit FPU

AMD "Zen 5" CPU microarchitecture will introduce a significant performance increase for AVX-512 workloads, with some sources reported as high as 40% performance increases over "Zen 4" in benchmarks that use AVX-512. A Moore's Law is Dead report detailing the execution engine of "Zen 5" holds the answer to how the company managed this—using a true 512-bit FPU. Currently, AMD uses a dual-pumped 256-bit FPU to execute AVX-512 workloads on "Zen 4." The updated FPU should significantly improve the core's performance in workloads that take advantage of 512-bit AVX or VNNI instructions, such as AI.

Giving "Zen 5" a 512-bit FPU meant that AMD also had to scale up the ancillaries—all the components that keep the FPU fed with data and instructions. The company therefore increased the capacity of the L1 DTLB. The load-store queues have been widened to meet the needs of the new FPU. The L1 Data cache has been doubled in bandwidth, and increased in size by 50%. The L1D is now 48 KB in size, up from 32 KB in "Zen 4." FPU MADD latency has been reduced by 1 cycle. Besides the FPU, AMD also increased the number of Integer execution pipes to 10, from 8 on "Zen 4." The exclusive L2 cache per core remains 1 MB in size.
Update 07:02 UTC: Moore's Law is Dead reached out to us and said that the slide previously posted by them, which we had used in an earlier version of this article, is fake, but said that the information contained in that slide is correct, and that they stand by the information.

SK hynix Signs Investment Agreement of Advanced Chip Packaging with Indiana

SK hynix Inc., the world's leading producer of High-Bandwidth Memory (HBM) chips, announced today that it will invest an estimated $3.87 billion in West Lafayette, Indiana to build an advanced packaging fabrication and R&D facility for AI products. The project, the first of its kind in the United States, is expected to drive innovation in the nation's AI supply chain, while bringing more than a thousand new jobs to the region.

The company held an investment agreement ceremony with officials from Indiana State, Purdue University, and the U.S. government at Purdue University in West Lafayette on the 3rd and officially announced the plan. At the event, officials from each party including Governor of Indiana Eric Holcomb, Senator Todd Young, Director of the White House Office of Science and Technology Policy Arati Prabhakar, Assistant Secretary of Commerce Arun Venkataraman, Secretary of Commerce State of Indiana David Rosenberg, Purdue University President Mung Chiang, Chairman of Purdue Research Foundation Mitch Daniels, Mayor of city of West Lafayette Erin Easter, Ambassador of the Republic of Korea to the United States Hyundong Cho, Consul General of the Republic of Korea in Chicago Junghan Kim, SK vice chairman Jeong Joon Yu, SK hynix CEO Kwak Noh-Jung and SK hynix Head of Package & Test Choi Woojin, participated.
Return to Keyword Browsing
Feb 23rd, 2025 08:30 EST change timezone

New Forum Posts

Popular Reviews

Controversial News Posts