News Posts matching #NVIDIA

Return to Keyword Browsing

NVIDIA GeForce NOW Gets 40% Off Performance Day Passes, More Games

Level up GeForce NOW experiences this summer with 40% off Performance Day Passes. Enjoy 24 hours of premium cloud gaming with RTX ON, delivering low latency and shorter wait times. The hot deal comes just in time for the cloud's highly anticipated launch of Dune: Awakening - a multiplayer survival game on a massive scale set on the unforgiving sands of Arrakis. It's perfect to pair with the nine games available this week, including the Frosthaven demo announced at Steam Next Fest.

Try Before You Buy
Level up to the cloud, no commitment required. For a limited time, grab a Performance Day Pass at a price that's less than an ice cream sundae and experience premium GeForce NOW gaming for 24 hours. With RTX ON, enjoy shorter wait times and lower latency for supported games, all powered by the cloud. Dive into popular games with upgraded visuals and smoother gameplay over free users, whether exploring vast open worlds or battling in fast-paced arenas. Take the experience even further by applying the value of the Day Pass toward a six-month Performance membership during the limited-time summer sale. It's the perfect way to try out premium cloud gaming before jumping into a longer-term membership.

NVIDIA TensorRT Boosts Stable Diffusion 3.5 Performance on NVIDIA GeForce RTX and RTX PRO GPUs

Generative AI has reshaped how people create, imagine and interact with digital content. As AI models continue to grow in capability and complexity, they require more VRAM, or video random access memory. The base Stable Diffusion 3.5 Large model, for example, uses over 18 GB of VRAM - limiting the number of systems that can run it well. By applying quantization to the model, noncritical layers can be removed or run with lower precision. NVIDIA GeForce RTX 40 Series and the Ada Lovelace generation of NVIDIA RTX PRO GPUs support FP8 quantization to help run these quantized models, and the latest-generation NVIDIA Blackwell GPUs also add support for FP4.

NVIDIA collaborated with Stability AI to quantize its latest model, Stable Diffusion (SD) 3.5 Large, to FP8 - reducing VRAM consumption by 40%. Further optimizations to SD3.5 Large and Medium with the NVIDIA TensorRT software development kit (SDK) double performance. In addition, TensorRT has been reimagined for RTX AI PCs, combining its industry-leading performance with just-in-time (JIT), on-device engine building and an 8x smaller package size for seamless AI deployment to more than 100 million RTX AI PCs. TensorRT for RTX is now available as a standalone SDK for developers.

TSMC Prepares "CoPoS": Next-Gen 310 × 310 mm Packages

As demand for ever-growing AI compute power continues to rise and manufacturing advanced nodes becomes more difficult, packaging is undergoing its golden era of development. Today's advanced accelerators often rely on TSMC's CoWoS modules, which are built on wafer cuts measuring no more than 120 × 150 mm in size. In response to the need for more space, TSMC has unveiled plans for CoPoS, or "Chips on Panel on Substrate," which could expand substrate dimensions to 310 × 310 mm and beyond. By shifting from round wafers to rectangular panels, CoPoS offers more than five times the usable area. This extra surface makes it possible to integrate additional high-bandwidth memory stacks, multiple I/O chiplets and compute dies in a single package. It also brings panel-level packaging (PLP) to the fore. Unlike wafer-level packaging (WLP), PLP assembles components on large, rectangular panels, delivering higher throughput and lower cost per unit. Systems with PLP will be actually viable for production runs and allow faster iterations over WLP.

TSMC will establish a CoPoS pilot line in 2026 at its Visionchip subsidiary. In 2027, the pilot facility will focus on refining the process, to meet partner requirements by the end of the year. Mass production is projected to begin between the end of 2028 and early 2029 at TSMC's Chiayi AP7 campus. That site, chosen for its modern infrastructure and ample space, is also slated to host production of multi-chip modules and System-on-Wafer technologies. NVIDIA is expected to be the launch partner for CoPoS. The company plans to leverage the larger panel area to accommodate up to 12 HBM4 chips alongside several GPU chiplets, offering significant performance gains for AI workloads. At the same time, AMD and Broadcom will continue using TSMC's CoWoS-L and CoWoS-R variants for their high-end products. Beyond simply increasing size, CoPoS and PLP may work in tandem with other emerging advances, such as glass substrates and silicon photonics. If development proceeds as planned, the first CoPoS-enabled devices could reach the market by late 2029.

MAINGEAR Unleashes ULTIMA 18 - The Ultimate 18" 4K Gaming Laptop

MAINGEAR, the leader in premium-quality, high-performance gaming PCs, today announced its most powerful laptop to date, the 18-inch ULTIMA 18. Developed in collaboration with CLEVO, ULTIMA 18 redefines what a gaming laptop can be by offering desktop-level specs, like a 4K@200 Hz G-SYNC display, Intel Core Ultra 9 275HX processor, and up to an NVIDIA GeForce RTX 5090 mobile GPU, all inside a sleek chassis outfitted with metal lid and palm rest.

Designed for elite gamers and creators who demand top-tier performance without compromise, ULTIMA 18 is MAINGEAR's first laptop to support modern dual-channel DDR5 memory, PCIe Gen 5 SSDs, dual Thunderbolt 5 ports, and Wi-Fi 7. Whether plugged in or on the move, this system delivers unprecedented power, quiet efficiency, and immersive visuals for the most demanding workloads and graphics-rich game titles.

AMD Instinct MI355X Draws up to 1,400 Watts in OAM Form Factor

Tomorrow evening, AMD will host its "Advancing AI" livestream to introduce the Instinct MI350 series, a new line of GPU accelerators designed for large-scale AI training and inference. First shown in prototype form at ISC 2025 in Hamburg just a day ago, each MI350 card features 288 GB of HBM3E memory, delivering up to 8 TB/s of sustained bandwidth. Customers can choose between the single-card MI350X and the higher-clocked MI355X or opt for a full eight-GPU platform that aggregates to over 2.3 TB of memory. Both chips are built on the CDNA 4 architecture, which now supports four different precision formats: FP16, FP8, FP6, and FP4. The addition of FP6 and FP4 is designed to boost throughput in modern AI workloads, where models of tomorrow with tens of trillions of parameters are trained on FP6 and FP4.

In half-precision tests, the MI350X achieves 4.6 PetaFLOPS on its own and 36.8 PetaFLOPS in eight-GPU platform form, while the MI355X surpasses those numbers, reaching 5.03 PetaFLOPS and just over 40 PetaFLOPS. AMD is also aiming to improve energy efficiency by a factor of thirty compared with its previous generation. The MI350X card runs within a 1,000 Watt power envelope and relies on air cooling, whereas the MI355X steps up to 1,400 Watts and is intended for direct-liquid cooling setups. That 400 Watt increase puts it right at NVIDIA's upcoming GB300 "Grace Blackwell Ultra" superchip, which is also a 1,400 W design. With memory capacity, raw computing, and power efficiency all pushed to new heights, the question remains whether real-world benchmarks will match these ambitious specifications. AMD now only lacks platform scaling beyond eight GPUs, which the Instinct MI400 series will address.

NVIDIA NVL72 GB200 Systems Accelerate the Journey to Useful Quantum Computing

The integration of quantum processors into tomorrow's supercomputers promises to dramatically expand the problems that can be addressed with compute—revolutionizing industries including drug and materials development.

In addition to being part of the vision for tomorrow's hybrid quantum-classical supercomputers, accelerated computing is dramatically advancing the work quantum researchers and developers are already doing to achieve that vision. And in today's development of tomorrow's quantum technology, NVIDIA GB200 NVL72 systems and their fifth-generation multinode NVIDIA NVLink interconnect capabilities have emerged as the leading architecture.

Europe Builds AI Infrastructure With NVIDIA to Fuel Region's Next Industrial Transformation

NVIDIA today announced it is working with European nations, and technology and industry leaders, to build NVIDIA Blackwell AI infrastructure that will strengthen digital sovereignty, support economic growth and position the continent as a leader in the AI industrial revolution. France, Italy, Spain and the U.K. are among the nations building domestic AI infrastructure with an ecosystem of technology and cloud providers, including Domyn, Mistral AI, Nebius and Nscale, and telecommunications providers, including Orange, Swisscom, Telefónica and Telenor.

These deployments will deliver more than 3,000 exaflops of NVIDIA Blackwell compute resources for sovereign AI, enabling European enterprises, startups and public sector organizations to securely develop, train and deploy agentic and physical AI applications. NVIDIA is establishing and expanding AI technology centers in Germany, Sweden, Italy, Spain, the U.K. and Finland. These centers build on NVIDIA's history of collaborating with academic institutions and industry through the NVIDIA AI Technology Center program and NVIDIA Deep Learning Institute to develop the AI workforce and scientific discovery throughout the regions.

NVIDIA Partners With Europe Model Builders and Cloud Providers to Accelerate Region's Leap Into AI

NVIDIA GTC Paris at VivaTech -- NVIDIA today announced that it is teaming with model builders and cloud providers across Europe and the Middle East to optimize sovereign large language models (LLMs), providing a springboard to accelerate enterprise AI adoption for the region's industries.

Model builders and AI consortiums Barcelona Supercomputing Center (BSC), Bielik.AI, Dicta, H Company, Domyn, LightOn, the National Academic Infrastructure for Supercomputing in Sweden (NAISS) together with KBLab at the National Library of Sweden, the Slovak Republic, the Technology Innovation Institute (TII), the University College of London, the University of Ljubljana and UTTER are teaming with NVIDIA to optimize their models with NVIDIA Nemotron techniques to maximize cost efficiency and accuracy for enterprise AI workloads, including agentic AI.

Pegatron Unveils AI-Optimized Server Innovations at GTC Paris 2025

PEGATRON, a globally recognized Design, Manufacturing, and Service (DMS) provider, is showcasing its latest AI server solutions at GTC Paris 2025. Built on NVIDIA Blackwell architecture, PEGATRON's cutting-edge systems are tailored for AI training, reasoning, and enterprise-scale deployment.

NVIDIA GB300 NVL72
At the forefront is the RA4802-72N2, built on the NVIDIA GB300 NVL72 rack system, featuring 72 NVIDIA Blackwell Ultra GPUs and 36 NVIDIA Grace CPUs. Designed for AI factories, it boosts output by up to 50X. PEGATRON's in-house developed Coolant Distribution Unit (CDU) delivers 310 kW of cooling capacity with redundant hot-swappable pumps, ensuring performance and reliability for mission-critical workloads.

NVIDIA DLSS 4 With Multi Frame Generation Coming To Another 8 Games, Including Stellar Blade and Dune: Awakening

More than 700 games and applications feature RTX technologies, and each week new games integrating NVIDIA DLSS, NVIDIA Reflex, and advanced ray-traced effects are released or announced, delivering the definitive PC experience for GeForce RTX players. DLSS 4 with Multi Frame Generation's adoption continues at rapid pace, and is the fastest-adopted NVIDIA gaming technology in our history.

This week, we can reveal another 8 games equipped with the frame rate multiplying technology, including Dragonkin: The Banished, Dune: Awakening, Marvel's Spider-Man 2, Mecha BREAK, MindsEye, REMATCH, and Stellar Blade. And Indiana Jones and the Great Circle: The Order of Giants DLC is launching September 4th, featuring full ray tracing and DLSS 4 with Multi Frame Generation and DLSS Ray Reconstruction. Each title also features DLSS Frame Generation, boosting frame rates for GeForce RTX 40 Series gamers, and DLSS Super Resolution, accelerating performance on every GeForce RTX GPU.

Lenovo Announces the All-New Workstations Solutions and Updates to the ThinkStation Desktop Portfolio

Lenovo, today at NXT BLD, announced its new portfolio of Workstation Solutions, a series of purpose-built, expertly-curated industry solutions that meet and exceed the rigorous performance and workflow requirements of engineers, designers, architects, data scientists, researchers, and creators so these power users can work smarter, faster, and more cost-effectively. Lenovo also unveiled the latest editions of its newest ThinkStation P2 and P3 desktop workstations designed to maximize performance and value.

Lenovo Workstations Solutions—Your Workflow, Perfected
Businesses need more than just powerful hardware—they need complete workflow solutions tailored to real-world industry challenges. Developed by Lenovo engineering experts through research and customer engagement to understand workflow bottlenecks and pain points, Lenovo Workstation Solutions deliver easily deployable blueprints—scalable and secure reference architectures powered by the state-of-the-art Lenovo Workstations—featuring superior hardware, software and services.

MSI Powers AI's Next Leap for Enterprises at ISC 2025

MSI, a global leader in high-performance server solutions, is showcasing its enterprise-grade, high-performance server platforms at ISC 2025, taking place June 10-12 at booth #E12. Built on standardized and modular architectures, MSI's AI servers are designed to power next-generation AI and accelerated computing workloads, enabling enterprises to rapidly advance their AI innovations.

"As AI workloads continue to grow and evolve toward inference-driven applications, we're seeing a significant shift in how enterprises approach AI deployment," said Danny Hsu, General Manager of Enterprise Platform Solutions at MSI. "With modular and standards-based architectures, enterprise data centers can now adopt AI technologies more quickly and cost-effectively than ever before. This marks a new era where AI is not only powerful but also increasingly accessible to businesses of all sizes.

ASUS Announces Key Milestone with Nebius and Showcases NVIDIA GB300 NVL72 System at GTC Paris 2025

ASUS today joined GTC Paris at VivaTech 2025 as a Gold Sponsor, highlighting its latest portfolio of AI infrastructure solutions and reinforcing its commitment to advancing the AI Factory vision with a full range of NVIDIA Blackwell Ultra solutions, delivering breakthrough performance from large-scale datacenter to personal desktop.

ASUS is also excited to announce a transformative partnership milestone in its partnership with Nebius. Together, the two companies are enabling a new era of AI innovation built on NVIDIA's advanced platforms. Building on the success of the NVIDIA GB200 NVL72 platform deployment, ASUS and Nebius are now moving forward with strategic collaborations featuring the next-generation NVIDIA GB300 NVL72 platform. This ongoing initiative underscores ASUS's role as a key enabler in AI infrastructure, committed to delivering scalable, high-performance solutions that help enterprises accelerate AI adoption and innovation.

Micron Ships HBM4 Samples: 12-Hi 36 GB Modules with 2 TB/s Bandwidth

Micron has achieved a significant advancement of the HBM4 architecture, which will stack 12 DRAM dies (12-Hi) to provide 36 GB of capacity per package. According to company representatives, initial engineering samples are scheduled to ship to key partners in the coming weeks, paving the way for full production in early 2026. The HBM4 design relies on Micron's established 1β ("one-beta") process node for DRAM tiles, in production since 2022, while it prepares to introduce EUV-enabled 1γ ("one-gamma") later this year for DDR5. By increasing the interface width from 1,024 to 2,048 bits per stack, each HBM4 chip can achieve a sustained memory bandwidth of 2 TB/s, representing a 20% efficiency improvement over the existing HBM3E standard.

NVIDIA and AMD are expected to be early adopters of Micron's HBM4. NVIDIA plans to integrate these memory modules into its upcoming Rubin-Vera AI accelerators in the second half of 2026. AMD is anticipated to incorporate HBM4 into its next-generation Instinct MI400 series, with further information to be revealed at the company's Advancing AI 2025 conference. The increased capacity and bandwidth of HBM4 will address growing demands in generative AI, high-performance computing, and other data-intensive applications. Larger stack heights and expanded interface widths enable more efficient data movement, a critical factor in multi-chip configurations and memory-coherent interconnects. As Micron begins mass production of HBM4, major obstacles to overcome will be thermal performance and real-world benchmarks, which will determine how effectively this new memory standard can support the most demanding AI workloads.
Micron HBM4 Memory

NVIDIA and HPE Join Forces to Construct Advanced Supercomputer in Germany

NVIDIA and Hewlett Packard Enterprise announced Tuesday at a supercomputing conference in Hamburg their partnership with Germany's Leibniz Supercomputing Centre to build a new supercomputer called Blue Lion which will deliver approximately 30 times more computing power than the current SuperMUC-NG system. The Blue Lion supercomputer will run on NVIDIA's upcoming Vera Rubin architecture. This setup combines the Rubin GPU with NVIDIA's first custom CPU Vera. The integrated system aims to unite simulation, data processing, and AI into one high-bandwidth low-latency platform. Optimized to support scientific research it boasts shared memory coherent compute abilities, and in-network acceleration.

HPE will build the system using its next-gen Cray technology by including NVIDIA GPUs along with cutting-edge storage and interconnect systems. The Blue Lion supercomputer will use HPE's 100% fanless direct liquid-cooling setup. This design circulates warm water through pipes for efficient cooling while the generated system's heat output will be reused to warm buildings nearby. The Blue Lion project comes after NVIDIA said Lawrence Berkeley National Lab in the US will also set up a Vera Rubin-powered system called Doudna next year. Scientists will have access to the Blue Lion supercomputer beginning in early 2027. The Blue Lion supercomputer, based in Germany will be utilized by researchers working on climate, physics, and machine learning. In contrast, Doudna, the U.S. Department of Energy's next supercomputer, will get its data from telescopes, genome sequencers, and fusion experiments.

El Capitan Retains Top Spot in 65th TOP500 List as Exascale Era Expands

The 65th edition of the TOP500 showed that the El Capitan system retains the No. 1 position. With El Capitan, Frontier, and Aurora, there are now 3 Exascale systems leading the TOP500. All three are installed at Department of Energy (DOE) laboratories in the United States.

The El Capitan system at the Lawrence Livermore National Laboratory, California, remains the No. 1 system on the TOP500. The HPE Cray EX255a system was measured with 1.742 EFlop/s on the HPL benchmark. LLNL now also submitted a measurement for the HPCG benchmark, achieving 17.41 Petaflop/s, which makes the system the new No. 1 on this ranking as well.

NVIDIA N1x is the Company's Arm Notebook Superchip

We've known since 2023 that NVIDIA is working on an Arm-based notebook SoC, and now we're seeing the first signs of the chip. A processor labelled "NVIDIA N1x" surfaced on the Geekbench 6.2.2 online database, where it scored 3096 points in the single-threaded benchmark, and 18837 points in the multithreaded benchmark. This chip is shown powering an HP-branded prototype notebook, labelled "HP 8EA3," which is running Geekbench on Ubuntu 24.04.1 LTS. The processor is identified by Geekbench as having 20 logical processors, which means it has a core-count of 20. This could be a multi-tiered big.LITTLE configuration making up those 20 cores. The clock speed being reported is 2.81 GHz. The company could implement reference Arm cores, such as the Cortex-X925 P-cores, and Cortex A725 E-cores. The HP testbed used for the Geekbench run has a whopping 128 GB of RAM.

NVIDIA has been eyeing a specific slice of the PC pie that's addressed by Qualcomm with its Snapdragon Elite line of processors for Windows-on-Arm notebooks, complete with an NPU accelerating Microsoft Copilot+ on device. The N1x could also compete with Apple's M3 or M4 chips powering its iPad Pro and MacBooks. For now, Microsoft has confined Arm-based Copilot+ to Snapdragon processors, but NVIDIA will probably work with Microsoft to open up this platform to its chips. NVIDIA has been an Arm SoC maker for decades, its first rodeo with Arm-based client-segment SoCs has been under the Tegra brand, powering Android smartphones and tablets. The company has been making Arm CPUs all this while, but for the enterprise segment (eg: Grace CPU).

Samsung 3 GB GDDR7 Chips Sold in Chinese Retail, Attracts Memory Modders

Samsung-branded 24 Gbit (3 GB) GDDR7 memory chips are being loosely sold in the Chinese retail market. The chips are being sold in OEM packaging meant for PCB pick-and-place machines, although the seller offers to sell these chips on a per-unit basis, priced at RMB ¥72.50, or approximately USD $10 per chip. This opens the door for graphics card memory modders to, in theory, give the GeForce RTX 5090 a 50% memory boost with $160 worth of memory chips. An RTX 5090 with 48 GB of memory would accelerate AI models with larger parameter counts than a stock card.

Given that off-brand RTX 5090 cards are being sold in China in board designs meant for commercial AI acceleration farms, this sale opens the door for off-brand RTX 5090 cards with 48 GB memory. Much of the RTX 5090 gray-market in China runs in complete disregard of U.S. export controls that prevent NVIDIA from selling the RTX 5090 in the Chinese domestic market, instead selling the RTX 5090D, a card with nearly identical gaming performance to RTX 5090, but with reduced AI acceleration and blockchain performance.

NVIDIA GeForce RTX 5050 Uses Slower GDDR6 Memory, Based on GB207 Silicon

NVIDIA is preparing to launch a new lower mid-range graphics card SKU in July, the GeForce RTX 5050. Positioned below the RTX 5060, the RTX 5050 possibly targets a price-point under the $250 mark, looking for a slice of the pie commanded by the Intel Arc B580. We are now learning that NVIDIA is making design choices that enable it to sell this card with an aggressive price, specifically, the choice of older generation GDDR6 memory. The card will likely feature 8 GB of GDDR6 memory across a 128-bit memory interface. At this point, we don't know the memory speeds, but if we were to hazard a guess, it could be 18 Gbps, for 288 GB/s of memory bandwidth.

The RTX 5050 is also expected to debut and max out the new "GB207" silicon, the smallest chip based on the GeForce Blackwell graphics architecture. This chip is expected to come with 20 SM, for 2,560 CUDA cores, 80 Tensor cores, 20 RT cores, 80 TMUs, and an unknown number of ROPs. The RTX 5050 is expected to be given a total graphics power (TGP) value of 130 W. It will be possible to build cards with 6-pin PCIe power connectors (75 W from connector, 75 W from the PCIe slot), although we expect single 8-pin PCIe to be the standard. The 130 W TGP will make it possible to build low-profile or compact, ITX-friendly cards.

NVIDIA GeForce RTX 5050 Max-Q/Mobile GPU Turns up in PC ID List

Post-Computex 2025, another hint about NVIDIA's next rung of current-gen mobile graphics cards has emerged online. Mid-week, harukaze5719 shared an intriguing screenshot of PCI IDs, accompanied by a brief message: "2d80-2dff = GB207." Most of the captured information is common knowledge, but an unreleased model was present. The "GeForce RTX 5050 Max-Q/Mobile" entry will likely translate into real life laptop/notebook form by mid-summer.

harukaze5719's leaked list seems to confirm the future deployment of a "GB207M" GPU die—likely the smallest and least potent iteration of Team Green's "Blackwell" gaming line. According to fresh reportage, a desktop sibling is supposedly being lined up for launch in July. A late April theory suggested AMD's preparing of RDNA 4 mobile "Radeon RX 9000M" GPUs; the least powerful options—reportedly based on a Navi 44 die—could compete with NVIDIA's GeForce RTX 5060 Laptop design. Rival timelines are unclear, but manufacturing partners (Lenovo, LG, MSI, Razer etc.) seem to be prepping a plethora of Team Green-based portable gaming devices.

NVIDIA Grabs Market Share, AMD Loses Ground, and Intel Disappears in Latest dGPU Update

Within the discrete graphics card sector, NVIDIA achieved a remarkable 92% share of the add-in board (AIB) GPU market in the first quarter of 2025, according to data released by Jon Peddie Research (JPR). This represents an 8.5% increase compared to NVIDIA's previous position. By contrast, AMD's share contracted to just 8%, down 7.3 points, while Intel's presence effectively disappeared, falling to 0% after losing 1.2 points. JPR reported that AIB shipments reached 9.2 million units during Q1 2025 despite desktop CPU shipments declining to 17.8 million units. The firm projects that the AIB market will face a compound annual decline of 10.3% from 2024 to 2028, although the installed base of discrete GPUs is expected to grow to 130 million units by the end of the forecast period. By 2028, an estimated 86% of desktop PCs are expected to feature a dedicated graphics card.

NVIDIA's success this quarter can be attributed to its launch of the RTX 50 series GPUs. In contrast, AMD's RDNA 4 GPUs were released significantly later in Q1. Additionally, Intel's Battlemage Arc GPUs, which were launched in Q4 2024, have struggled to gain traction, likely due to limited availability and low demand in the mainstream market. The broader PC GPU market, which includes integrated solutions, contracted by 12% from the previous quarter, with a total of 68.8 million units shipped. Desktop graphics unit sales declined by 16%, while notebook GPUs decreased by 10%. Overall, NVIDIA's total GPU share rose by 3.6 points, AMD's dipped by 1.6 points, and Intel's declined by 2.1 points. Meanwhile, data center GPUs bucked the overall downward trend, rising by 9.6% as enterprises continue to invest in artificial intelligence applications. On the CPU side, notebook processors accounted for 71% of shipments, with desktop CPUs comprising the remaining 29%.

NVIDIA Blackwell Delivers Breakthrough Performance in Latest MLPerf Training Results

NVIDIA is working with companies worldwide to build out AI factories—speeding the training and deployment of next-generation AI applications that use the latest advancements in training and inference. The NVIDIA Blackwell architecture is built to meet the heightened performance requirements of these new applications. In the latest round of MLPerf Training—the 12th since the benchmark's introduction in 2018—the NVIDIA AI platform delivered the highest performance at scale on every benchmark and powered every result submitted on the benchmark's toughest large language model (LLM)-focused test: Llama 3.1 405B pretraining.

The NVIDIA platform was the only one that submitted results on every MLPerf Training v5.0 benchmark—underscoring its exceptional performance and versatility across a wide array of AI workloads, spanning LLMs, recommendation systems, multimodal LLMs, object detection and graph neural networks. The at-scale submissions used two AI supercomputers powered by the NVIDIA Blackwell platform: Tyche, built using NVIDIA GB200 NVL72 rack-scale systems, and Nyx, based on NVIDIA DGX B200 systems. In addition, NVIDIA collaborated with CoreWeave and IBM to submit GB200 NVL72 results using a total of 2,496 Blackwell GPUs and 1,248 NVIDIA Grace CPUs.

Chinese Tech Firms Reportedly Unimpressed with Overheating of Huawei AI Accelerator Samples

Mid-way through last month, Tencent's President—Martin Lau—confirmed that this company had stockpiled a huge quantity of NVIDIA H20 AI GPUs, prior to new trade restrictions coming into effect. According to earlier reports, China's largest tech firms have collectively spent $16 billion on hardware acquisitions in Q1'25. Team Green engineers are likely engaged in the creation of "nerfed" enterprise-grade chip designs—potentially ready for deployment later on in 2025. Huawei leadership is likely keen to take advantage of this situation, although it will be difficult to compete with the sheer volume of accumulated H20 units. The Shenzhen, Guangdong-based giant's Ascend AI accelerator family is considered to be a valid alternative to equivalent "sanction-conformant" NVIDIA products.

The controversial 910C model and a successor seem to be worthy candidates; as demonstrated by preliminary performance data, but fresh industry murmurs suggest teething problems. The Information has picked up inside track chatter from unnamed moles at ByteDance and Alibaba. During test runs, staffers noted the overheating of Huawei Ascend 910C trial samples. Additionally, they highlighted limitations within the Huawei Compute Architecture for Neural Networks (CANN) software platform. NVIDIA's extremely mature CUDA ecosystem holds a significant advantage here. Several of China's prime AI players—including DeepSeek—are reportedly pursuing in-house AI chip development projects; therefore positioning themselves as competing with Huawei, in a future scenario.

NVIDIA Marketplace Briefly Lists "Acme GeForce RTX 4040 8 GB BRICK Edition"

Perusers of the NVIDIA US Marketplace noticed a listing of an "Acme GeForce RTX 4040 8 GB BRICK Edition" product. Prior to a fresh removal, this mystery item was not in stock—reflecting the short supply of many GeForce RTX 50-series offerings. An advertised £$0.00" price indicates internal placeholder or jokey origins. After all, graphics card-related scams have involved the swiping of genuine articles—craftily replaced with similarly weighted substitutes. In some cases actual bricks, or—very recently—multiple cross-body backpacks stuffed into retail packages. According to news reports, NVIDIA's past-gen BRICK SKU is equipped with an impressive set of amusing features: "durable casing, passive cooling, absolute silence, and zero dependency on drivers." On a serious note, Team Green's desktop "Ada Lovelace" gaming card generation never dipped below the GeForce RTX 4060 tier. TechPowerUp's GPU database lists an unreleased GeForce RTX 4050 model—leakers believe that a direct successor will turn up in July.

EUROCOM Intros Nightsky RX515 Laptop Series Powered by NVIDIA Blackwell GPU

EUROCOM launches the Nightsky RX515 - a remarkable fusion of power, portability, and precision, designed to cater demanding professionals, students creators and gamers. EUROCOM Nightsky RX515 is powered by NVIDIA latest Blackwell RTX 5060 architecture, Intel Core Ultra 9 275HX processor and supports up to 4 active displays, 128 GB of DDR5 memory and up to 24 TB of SSD RAID 0/1/5 NVMe drives.

EUROCOM Nightsky RX515 is a 15.6-inch customizable laptop engineered for users who desire maximum flexibility and connectivity in a lightweight, travel-ready form factor. Whether user is a student, professional, gamer, home user or creator on the move, the Nightsky RX515 offers the freedom to do more wherever the work or pleasure takes user to.
Return to Keyword Browsing
Jun 12th, 2025 23:01 CDT change timezone

New Forum Posts

Popular Reviews

TPU on YouTube

Controversial News Posts