News Posts matching #Enterprise

Return to Keyword Browsing

Swissbit Intros iShield Key 2 Hardware Authenticator

Swissbit, a leading manufacturer of storage and security solutions, announces the launch of the iShield Key 2, the next generation of its hardware authenticators. Powered by an NXP chip, the new product family delivers phishing-resistant authentication for enterprises and public sector organizations worldwide, offering a wide range of models and feature sets.

At its core, the iShield Key 2 series is built on the FIDO2 standard, supporting passwordless authentication with market-leading capacity for up to 300 Passkeys. The Pro versions expand functionality to include HOTP and TOTP for generating one-time passwords and PIV for smart card integration. Setting new standards, the iShield Key 2 MIFARE variant uniquely combines digital and physical access control in a single device. This enables efficient authentication for both IT systems and physical infrastructure such as buildings or secure printing environments. For customers with the highest security requirements, a FIPS 140-3 Level 3 certified version is also available. All models are housed in a compact, lightweight, and rugged industrial-grade casing - engineered and manufactured at Swissbit's facility in Berlin, Germany.

NVIDIA & ServiceNow CEOs Jointly Present "Super Genius" Open-source Apriel Nemotron 15B LLM

ServiceNow is accelerating enterprise AI with a new reasoning model built in partnership with NVIDIA—enabling AI agents that respond in real time, handle complex workflows and scale functions like IT, HR and customer service teams worldwide. Unveiled today at ServiceNow's Knowledge 2025—where NVIDIA CEO and founder Jensen Huang joined ServiceNow chairman and CEO Bill McDermott during his keynote address—Apriel Nemotron 15B is compact, cost-efficient and tuned for action. It's designed to drive the next step forward in enterprise large language models (LLMs).

Apriel Nemotron 15B was developed with NVIDIA NeMo, the open NVIDIA Llama Nemotron Post-Training Dataset and ServiceNow domain-specific data, and was trained on NVIDIA DGX Cloud running on Amazon Web Services (AWS). The news follows the April release of the NVIDIA Llama Nemotron Ultra model, which harnesses the NVIDIA open dataset that ServiceNow used to build its Apriel Nemotron 15B model. Ultra is among the strongest open-source models at reasoning, including scientific reasoning, coding, advanced math and other agentic AI tasks.

Cooler Master HQ Experience Coinciding with Computex 2025

Join us at Cooler Master HQ for an exclusive COMPUTEX experience. From cooling the demands of AI and high-performance systems to enabling everyday reliability and modular design, discover how we're shaping the future through smarter, more adaptable solutions. Cooler Master invites you to experience the future of tech—live from our headquarters. As the tech landscape evolves, we're pushing cooling further: powering the rise of AI, delivering smarter performance for everyday systems, and unlocking new possibilities through modular design. This year's showcase brings it all together under one vision.

One Cooler Master—Where Cooling Drives the Future
It's the fusion of our industrial and consumer expertise. For over 30 years, we've built advanced cooling solutions for businesses, including: AI data centers, telecom, and mobile—then brought that same innovation to gamers, creators, and everyday users. One Cooler Master reflects our unified vision: engineering that drives performance at every level.

IBM Intros LinuxONE Emperor 5 Mainframe with Telum II Processor

IBM has introduced the LinuxONE Emperor 5, its newest Linux computing platform that runs on the Telum II processor with built-in AI acceleration features. This launch aims to tackle three key issues for tech leaders: better security measures, reduced costs, and smooth AI incorporation into business systems. The heart of the system, the Telum II processor, includes a second-generation on-chip AI accelerator. This component is designed to boost predictive AI abilities and large language models for instant transaction handling. The upcoming IBM Spyre Accelerator (set to arrive in late 2025) via PCIe card will boost generative AI functions. The platform comes with an updated AI Toolkit fine-tuned for the Telum II processor. It also offers early looks at Red Hat OpenShift AI and Virtualization allowing unified control of both standard virtual machines and containerized workloads.

The platform provides wide-ranging security measures. These include confidential computing strong cryptographic abilities, and NIST-approved post-quantum algorithms. These safeguard sensitive AI models and data from current risks and expected post-quantum attacks. When it comes to productivity, companies can combine several server workloads on one high-capacity system. This might cut ownership expenses by up to 44% compared to x86 options over five years. At the same time, it keeps exceptional 99.999999% uptime rates according to IBM. The LinuxOne Emperor 5 will run Linux Red Hat Enterprise Linux (RHEL), SUSE Linux Enterprise Server (SLES) and Canonical Ubuntu Server. Tina Tarquinio, chief product officer at IBM Z and LinuxONE, said: "IBM LinuxONE 5 represents the next evolution of our Linux infrastructure strategy. It is designed to help clients unlock the full potential of Linux and AI while optimizing their datacenters, simplifying their operations, and addressing risk. Whether you're building intelligent applications, deploying regulated workloads, consolidating infrastructure, or preparing for the next wave of transformation, IBM LinuxONE offers an exciting path forward."

AMD Discusses Importance of AI PC Initiative - Ryzen AI PRO 300 Series Ideal for Enterprise

One of the interesting long-term trends in the commercial market is the shift in how people are functionally using PCs. Over the past few years, AI has gone from a dinner party conversation piece to practical use cases. Hand-waved discussions of future benefits are now concrete benefits in the here-and-now, from writing and validating software code to shaping customer communication. While many of these early AI workloads were executed in the cloud, we are seeing new use cases that require new local PC capabilities - capabilities AMD has been working to develop as part of its larger AI PC initiative. We've seen AI deployment accelerate in PCs, particularly as Microsoft builds these functions right into the operating system via its Windows 11 Copilot+ PC program.

In just two years, we've moved from an environment where the NPU was an unsupported, unknown proposition, to integrating NPU support into first and third-party applications. Commercial OEMs are expected to introduce AI PCs and Copilot+ PCs at a variety of price points throughout 2025, making local AI support more available and affordable. That's an important factor for corporate customers, many of which will be simultaneously negotiating the end of Windows 10 support and a historically large, pandemic-related system refresh wave. AI presents a further variable. The nature of the AI question has evolved over the last couple of years from "Does AI have a future in the enterprise?" to "What are the best practices for enterprise AI and AI PC deployment?" That's a significant shift, and it's worth unpacking in a little more detail.

IBM Cloud is First Service Provider to Deploy Intel Gaudi 3

IBM is the first cloud service provider to make Intel Gaudi 3 AI accelerators available to customers, a move designed to make powerful artificial intelligence capabilities more accessible and to directly address the high cost of specialized AI hardware. For Intel, the rollout on IBM Cloud marks the first major commercial deployment of Gaudi 3, bringing choice to the market. By leveraging Intel Gaudi 3 on IBM Cloud, the two companies aim to help clients cost-effectively test, innovate and deploy GenAI solutions.

According to a recent forecast by research firm Gartner, worldwide generative AI (GenAI) spending is expected to total $644 billion in 2025, an increase of 76.4% from 2024. The research found "GenAI will have a transformative impact across all aspects of IT spending markets, suggesting a future where AI technologies become increasingly integral to business operations and consumer products."

NVIDIA Anticipates Another Leap Forward for Cybersecurity - Enabled by Agentic AI

Agentic AI is redefining the cybersecurity landscape—introducing new opportunities that demand rethinking how to secure AI while offering the keys to addressing those challenges. Unlike standard AI systems, AI agents can take autonomous actions—interacting with tools, environments, other agents and sensitive data. This provides new opportunities for defenders but also introduces new classes of risks. Enterprises must now take a dual approach: defend both with and against agentic AI.

Building Cybersecurity Defense With Agentic AI
Cybersecurity teams are increasingly overwhelmed by talent shortages and growing alert volume. Agentic AI offers new ways to bolster threat detection, response and AI security—and requires a fundamental pivot in the foundations of the cybersecurity ecosystem. Agentic AI systems can perceive, reason and act autonomously to solve complex problems. They can also serve as intelligent collaborators for cyber experts to safeguard digital assets, mitigate risks in enterprise environments and boost efficiency in security operations centers. This frees up cybersecurity teams to focus on high-impact decisions, helping them scale their expertise while potentially reducing workforce burnout. For example, AI agents can cut the time needed to respond to software security vulnerabilities by investigating the risk of a new common vulnerability or exposure in just seconds. They can search external resources, evaluate environments and summarize and prioritize findings so human analysts can take swift, informed action.

NVIDIA Bringing Cybersecurity Platform to Every AI Factory

As enterprises increasingly adopt AI, securing AI factories—where complex, agentic workflows are executed—has never been more critical. NVIDIA is bringing runtime cybersecurity to every AI factory with a new NVIDIA DOCA software framework, part of the NVIDIA cybersecurity AI platform. Running on the NVIDIA BlueField networking platform, NVIDIA DOCA Argus operates on every node to immediately detect and respond to attacks on AI workloads, integrating seamlessly with enterprise security systems to deliver instant threat insights. The DOCA Argus framework provides runtime threat detection by using advanced memory forensics to monitor threats in real time, delivering detection speeds up to 1,000x faster than existing agentless solutions—without impacting system performance.

Unlike conventional tools, Argus runs independently of the host, requiring no agents, integration or reliance on host-based resources. This agentless, zero-overhead design enhances system efficiency and ensures resilient security in any AI compute environment, including containerized and multi-tenant infrastructures. By operating outside the host, Argus remains invisible to attackers—even in the event of a system compromise. Cybersecurity professionals can seamlessly integrate the framework with their SIEM, SOAR and XDR security platforms, enabling continuous monitoring and automated threat mitigation and extending their existing cybersecurity capabilities for AI infrastructure.

Innodisk Announces its First PCIe Gen5 SSD Series for Data Centers

Innodisk, a leading industrial-grade flash storage provider, announces the launch of its latest PCIe Gen 5 SSD series, designed to meet the OCP Data Center NVMe SSD spec v2.0 and the increasing demands of AI model training, big data analytics, and data-intensive environments.

The new series supports multiple form factors, including U.2, as well as EDSFF E1.S, E3.S, and the brand-new E3.L for data center applications, catering to the diverse needs of enterprise and data center environments. This marks a new chapter for Innodisk in further fulfilling enterprise demands.

MSI Servers Power the Next-Gen Datacenters at the 2025 OCP EMEA Summit

MSI, a leading global provider of high-performance server solutions, unveiled its latest ORv3-compliant and high-density multi-node server platforms at the 2025 OCP EMEA Summit, held April 29-30 at booth A19. Built on OCP-recognized DC-MHS architecture and supporting the latest AMD EPYC 9005 Series processors, these next-generation platforms are engineered to deliver outstanding compute density, energy efficiency, and scalability—meeting the evolving demands of modern, data-intensive datacenters.

"We are excited to be part of open-source innovation and sustainability through our contributions to the Open Compute Project," said Danny Hsu, General Manager of Enterprise Platform Solutions. "We remain committed to advancing open standards, datacenter-focused design, and modular server architecture. Our ability to rapidly develop products tailored to specific customer requirements is central to enabling next-generation infrastructure, making MSI a trusted partner for scalable, high-performance solutions."

Oracle Cloud Infrastructure Bolstered by Thousands of NVIDIA Blackwell GPUs

Oracle has stood up and optimized its first wave of liquid-cooled NVIDIA GB200 NVL72 racks in its data centers. Thousands of NVIDIA Blackwell GPUs are now being deployed and ready for customer use on NVIDIA DGX Cloud and Oracle Cloud Infrastructure (OCI) to develop and run next-generation reasoning models and AI agents. Oracle's state-of-the-art GB200 deployment includes high-speed NVIDIA Quantum-2 InfiniBand and NVIDIA Spectrum-X Ethernet networking to enable scalable, low-latency performance, as well as a full stack of software and database integrations from NVIDIA and OCI.

OCI, one of the world's largest and fastest-growing cloud service providers, is among the first to deploy NVIDIA GB200 NVL72 systems. The company has ambitious plans to build one of the world's largest Blackwell clusters. OCI Superclusters will scale beyond 100,000 NVIDIA Blackwell GPUs to meet the world's skyrocketing need for inference tokens and accelerated computing. The torrid pace of AI innovation continues as several companies including OpenAI have released new reasoning models in the past few weeks.

MSI Presenting AI's Next Leap at Japan IT Week Spring 2025

MSI, a leading global provider of high-performance server solutions, is bringing AI-driven innovation to Japan IT Week Spring 2025 at Booth #21-2 with high-performance server platforms built for next-generation AI and cloud computing workloads. MSI's NVIDIA MGX AI Servers deliver modular GPU-accelerated computing to optimize AI training and inference, while the Core Compute line of Multi-Node Servers maximize compute density and efficiency for AI inference and cloud service provider workloads. MSI's Open Compute line of ORv3 Servers enhance scalability and thermal efficiency in hyperscale AI deployments. MSI's Enterprise Servers provide balanced compute, storage, and networking for seamless AI workloads across cloud and edge. With deep expertise in system integration and AI-driven infrastructure, MSI is advancing the next generation of intelligent computing solutions to power AI's next leap.

"AI's advancement hinges on performance efficiency, compute density, and workload scalability. MSI's server platforms are engineered to accelerate model training, optimize inference, and maximize resource utilization—ensuring enterprises have the processing power to turn AI potential into real-world impact," said Danny Hsu, General Manager of MSI Enterprise Platform Solutions.

AMD Announces Press Conference & Livestream at Computex 2025

AMD today announced that it will be hosting a press conference during Computex 2025. The in-person and livestreamed press conference will take place on Wednesday, May 21, 2025, at 11 a.m. UTC+8, Taipei, at the Grand Hyatt, Taipei. The event will showcase the advancements AMD has driven with AI in gaming, PCs and professional workloads.

AMD senior vice president and general manager of the Computing and Graphics Group Jack Huynh, along with industry partners, will discuss how AMD is expanding its leadership across gaming, workstations, and AI PCs, and highlight the breadth of the company's high-performance computing and AI product portfolio. The livestream will start at 8 p.m. PT/11 p.m. ET on Tuesday, May 20 on AMD.com, with replay available after the conclusion of the livestream event.

NVIDIA Blackwell Platform Boosts Water Efficiency by Over 300x - "Chill Factor" for AI Infrastructure

Traditionally, data centers have relied on air cooling—where mechanical chillers circulate chilled air to absorb heat from servers, helping them maintain optimal conditions. But as AI models increase in size, and the use of AI reasoning models rises, maintaining those optimal conditions is not only getting harder and more expensive—but more energy-intensive. While data centers once operated at 20 kW per rack, today's hyperscale facilities can support over 135 kW per rack, making it an order of magnitude harder to dissipate the heat generated by high-density racks. To keep AI servers running at peak performance, a new approach is needed for efficiency and scalability.

One key solution is liquid cooling—by reducing dependence on chillers and enabling more efficient heat rejection, liquid cooling is driving the next generation of high-performance, energy-efficient AI infrastructure. The NVIDIA GB200 NVL72 and the NVIDIA GB300 NVL72 are rack-scale, liquid-cooled systems designed to handle the demanding tasks of trillion-parameter large language model inference. Their architecture is also specifically optimized for test-time scaling accuracy and performance, making it an ideal choice for running AI reasoning models while efficiently managing energy costs and heat.

NVIDIA Will Bring Agentic AI Reasoning to Enterprises with Google Cloud

NVIDIA is collaborating with Google Cloud to bring agentic AI to enterprises seeking to locally harness the Google Gemini family of AI models using the NVIDIA Blackwell HGX and DGX platforms and NVIDIA Confidential Computing for data safety. With the NVIDIA Blackwell platform on Google Distributed Cloud, on-premises data centers can stay aligned with regulatory requirements and data sovereignty laws by locking down access to sensitive information, such as patient records, financial transactions and classified government information. NVIDIA Confidential Computing also secures sensitive code in the Gemini models from unauthorized access and data leaks.

"By bringing our Gemini models on premises with NVIDIA Blackwell's breakthrough performance and confidential computing capabilities, we're enabling enterprises to unlock the full potential of agentic AI," said Sachin Gupta, vice president and general manager of infrastructure and solutions at Google Cloud. "This collaboration helps ensure customers can innovate securely without compromising on performance or operational ease." Confidential computing with NVIDIA Blackwell provides enterprises with the technical assurance that their user prompts to the Gemini models' application programming interface—as well as the data they used for fine-tuning—remain secure and cannot be viewed or modified. At the same time, model owners can protect against unauthorized access or tampering, providing dual-layer protection that enables enterprises to innovate with Gemini models while maintaining data privacy.

Industry's First-to-Market Supermicro NVIDIA HGX B200 Systems Demonstrate AI Performance Leadership

Super Micro Computer, Inc. (SMCI), a Total IT Solution Provider for AI/ML, HPC, Cloud, Storage, and 5G/Edge, has announced first-to-market industry leading performance on several MLPerf Inference v5.0 benchmarks, using the 8-GPU. The 4U liquid-cooled and 10U air-cooled systems achieved the best performance in select benchmarks. Supermicro demonstrated more than 3 times the tokens per second (Token/s) generation for Llama2-70B and Llama3.1-405B benchmarks compared to H200 8-GPU systems. "Supermicro remains a leader in the AI industry, as evidenced by the first new benchmarks released by MLCommons in 2025," said Charles Liang, president and CEO of Supermicro. "Our building block architecture enables us to be first-to-market with a diverse range of systems optimized for various workloads. We continue to collaborate closely with NVIDIA to fine-tune our systems and secure a leadership position in AI workloads." Learn more about the new MLPerf v5.0 Inference benchmarks here.

Supermicro is the only system vendor publishing record MLPerf inference performance (on select benchmarks) for both the air-cooled and liquid-cooled NVIDIA HGX B200 8-GPU systems. Both air-cooled and liquid-cooled systems were operational before the MLCommons benchmark start date. Supermicro engineers optimized the systems and software to showcase the impressive performance. Within the operating margin, the Supermicro air-cooled B200 system exhibited the same level of performance as the liquid-cooled B200 system. Supermicro has been delivering these systems to customers while we conducted the benchmarks. MLCommons emphasizes that all results be reproducible, that the products are available and that the results can be audited by other MLCommons members. Supermicro engineers optimized the systems and software, as allowed by the MLCommons rules.

MangoBoost Achieves Record-Breaking MLPerf Inference v5.0 Results with AMD Instinct MI300X

MangoBoost, a provider of cutting-edge system solutions designed to maximize AI data center efficiency, has set a new industry benchmark with its latest MLPerf Inference v5.0 submission. The company's Mango LLMBoost AI Enterprise MLOps software has demonstrated unparalleled performance on AMD Instinct MI300X GPUs, delivering the highest-ever recorded results for Llama2-70B in the offline inference category. This milestone marks the first-ever multi-node MLPerf inference result on AMD Instinct MI300X GPUs. By harnessing the power of 32 MI300X GPUs across four server nodes, Mango LLMBoost has surpassed all previous MLPerf inference results, including those from competitors using NVIDIA H100 GPUs.

Unmatched Performance and Cost Efficiency
MangoBoost's MLPerf submission demonstrates a 24% performance advantage over the best-published MLPerf result from Juniper Networks utilizing 32 NVIDIA H100 GPUs. Mango LLMBoost achieved 103,182 tokens per second (TPS) in the offline scenario and 93,039 TPS in the server scenario on AMD MI300X GPUs, outperforming the previous best result of 82,749 TPS on NVIDIA H100 GPUs. In addition to superior performance, Mango LLMBoost + MI300X offers significant cost advantages. With AMD MI300X GPUs priced between $15,000 and $17,000—compared to the $32,000-$40,000 cost of NVIDIA H100 GPUs (source: Tom's Hardware—H100 vs. MI300X Pricing)—Mango LLMBoost delivers up to 62% cost savings while maintaining industry-leading inference throughput.

Forget Reboots, Live Patches are Coming to Windows 11 Enterprise Clients

Microsoft is introducing live patch updates for Windows 11 Enterprise, version 24H2, that allow critical security fixes to be applied without interrupting users. These updates, known as hotpatches, are available for x64 devices running on AMD or Intel CPUs. Hotpatch updates are designed to install quickly and take effect immediately. Unlike standard monthly security updates that require a system restart, hotpatch updates provide instant protection against vulnerabilities while allowing users to continue working. This new process can reduce the number of restarts from twelve per year to just four. The update schedule follows a quarterly cycle. In January, April, July, and October, devices install a complete security update with new features and fixes that do require a restart. In the two months that follow each of these baseline updates, devices receive hotpatch updates that only include security fixes and do not need a reboot. This approach ensures that essential protections are applied quickly without impacting daily work.

To use hotpatch updates, organizations need a Microsoft subscription that includes Windows 11 Enterprise (or Windows 365 Enterprise) and devices running build 26100.2033 or later. These devices must also be managed using Microsoft Intune, where IT administrators can set up a hotpatch-enabled quality update policy. The Intune admin center automatically detects eligible devices and manages the update process. Hotpatch updates are currently available on Intel and AMD-powered devices. For Arm64 devices, hotpatch updates are still in public preview and require an extra configuration step: disabling CHPE support via a registry key or the upcoming DisableCHPE CSP. This update system represents a more efficient way to secure Windows client devices. By minimizing the need for restarts and delivering updates in a predictable, quarterly cycle, Microsoft aims to help organizations protect their systems with minimal disruption. We expect these live patches to trickle down to more Windows 11 versions, like Home and Pro editions.

AAEON Launches UP 710S Edge, Its Smallest Mini PC Powered by Intel N Series Processor

AAEON's UP brand, has released the UP 710S Edge, the company's smallest Intel Processor N-powered Mini PC with Wi-Fi support, measuring just 92 mm x 77 mm x 38 mm.

Positioned as a compact platform for companies to upgrade industrial automation setups, the UP 710S Edge is available in models featuring the full Intel Processor N series family (formerly Alder Lake-N), as well as offering an 8-bit GPIO with optional SPI, I2C, and PWM, a first for the product line.

IBM & Intel Announce the Availability of Gaudi 3 AI Accelerators on IBM Cloud

Yesterday, at Intel Vision 2025, IBM announced the availability of Intel Gaudi 3 AI accelerators on IBM Cloud. This offering delivers Intel Gaudi 3 in a public cloud environment for production workloads. Through this collaboration, IBM Cloud aims to help clients more cost-effectively scale and deploy enterprise AI. Intel Gaudi 3 AI accelerators on IBM Cloud are currently available in Frankfurt (eu-de) and Washington, D.C. (us-east) IBM Cloud regions, with future availability for the Dallas (us-south) IBM Cloud region in Q2 2025.

IBM's AI in Action 2024 report found that 67% of surveyed leaders reported revenue increases of 25% or more due to including AI in business operations. Although AI is demonstrating promising revenue increases, enterprises are also balancing the costs associated with the infrastructure needed to drive performance. By leveraging Intel's Gaudi 3 on IBM Cloud, the two companies are aiming to help clients more cost effectively test, innovate and deploy generative AI solutions. "By bringing Intel Gaudi 3 AI accelerators to IBM Cloud, we're enabling businesses to help scale generative AI workloads with optimized performance for inferencing and fine-tuning. This collaboration underscores our shared commitment to making AI more accessible and cost-effective for enterprises worldwide," said Saurabh Kulkarni, Vice President, Datacenter AI Strategy and Product Management, Intel.

Supermicro Ships Over 20 New Systems that Redefine Single-Socket Performance

Super Micro Computer, Inc., a Total IT Solution Provider for AI/ML, HPC, Cloud, Storage, and 5G/Edge, is announcing the availability of new single-socket servers capable of supporting applications that required dual-socket servers for a range of data center workloads. By leveraging a single-socket architecture, enterprises and data center operators can reduce initial acquisition costs, ongoing operational costs such as power and cooling, and reduce the physical footprint of server racks compared to previous generations of systems based on older processors.

"We are entering a new era of compute where energy-efficient and thermally optimized single-socket architectures are becoming a viable alternative to traditional dual-processor servers," said Charles Liang, president and CEO of Supermicro. "Our new single-socket servers support 100% more cores per system than previous generations and have been designed to maximize acceleration, networking, and storage flexibility. Supporting up to 500-watt TDP processors, these new systems can be configured to fulfill a wide range of workload requirements."

Japanese Retailer Reportedly Prepping NVIDIA RTX PRO 6000 96 GB Stock For Sale in May, Leak Indicates $8435+ Pricing

During GTC 2025, NVIDIA unveiled the professional (PRO) side of its "Blackwell" GPU line—headlined by a monstrous GDDR7 96 GB option, that unleashes the full potential of their GB202 die. Industry watchdogs anticipated sky-high pricing, as befits such a potent specification sheet/feature set. As reported by VideoCardz over the past weekend, a North American enterprise PC hardware store—Connection—has populated its webshop with several of Team Green's brand-new RTX PRO Blackwell Series SKUs. The publication received tip-offs from a portion of its readership; including some well-heeled individuals who have already claimed pre-orders. Starting off, the investigation highlighted upper crust offerings: "the flagship model, called the RTX PRO 6000 with 96 GB of VRAM, will launch at $8435 (bulk) to $8565 (box), and this price seemingly applies to both models: the Workstation Edition and a sub-variant called Max-Q. Both are equipped with the same specs, but the latter is capped at 300 W TDP while retaining 88% of the Al performance, claimed NVIDIA."

Connection has removed its RTX PRO 6000 Blackwell and RTX PRO 6000 Blackwell Max-Q product pages, but the rest of Team Green's professional stack is still visible (see relevant screenshot below). The RTX PRO 5000 Blackwell 48 GB card is priced at $4569.24 (or $4439.50 for bulk). The cheapest offering is a $696.54 RTX PRO 2000 Blackwell 8 GB model. Officially, NVIDIA and its main professional series board partner—PNY—only revealed 4500, 5000 and 6000 product tiers. VideoCardz put a spotlight on some of these unannounced options, including: "the RTX 4000 non-SFF version, while this retailer has six listings for such SKUs (two SFF and two non-SFF, both in bulk and box variants). Presumably, this would suggest that NVIDIA may launch a non-SFF version later. However, the company didn't put 'SFF' in the official card's name, so perhaps this information is no longer valid, and there's only one model." According to a GDM/Hermitage AkiHabara Japan press release, a local reseller—Elsa—is preparing NVIDIA RTX PRO 6000 Blackwell Workstation Edition and RTX PRO 6000 Blackwell Max-Q Workstation Edition stock for scheduled release "in May 2025, while the other models are scheduled for release around summer." Additionally, another retailer (ASK Co., Ltd.): "has stated that the price and release date are subject to inquiry."

NVIDIA & Storage Industry Leaders Unveil New Class of Enterprise Infrastructure for the AI Era

At GTC 2025, NVIDIA announced the NVIDIA AI Data Platform, a customizable reference design that leading providers are using to build a new class of AI infrastructure for demanding AI inference workloads: enterprise storage platforms with AI query agents fueled by NVIDIA accelerated computing, networking and software. Using the NVIDIA AI Data Platform, NVIDIA-Certified Storage providers can build infrastructure to speed AI reasoning workloads with specialized AI query agents. These agents help businesses generate insights from data in near real time, using NVIDIA AI Enterprise software—including NVIDIA NIM microservices for the new NVIDIA Llama Nemotron models with reasoning capabilities—as well as the new NVIDIA AI-Q Blueprint.

Storage providers can optimize their infrastructure to power these agents with NVIDIA Blackwell GPUs, NVIDIA BlueField DPUs, NVIDIA Spectrum-X networking and the NVIDIA Dynamo open-source inference library. Leading data platform and storage providers—including DDN, Dell Technologies, Hewlett Packard Enterprise, Hitachi Vantara, IBM, NetApp, Nutanix, Pure Storage, VAST Data and WEKA—are collaborating with NVIDIA to create customized AI data platforms that can harness enterprise data to reason and respond to complex queries. "Data is the raw material powering industries in the age of AI," said Jensen Huang, founder and CEO of NVIDIA. "With the world's storage leaders, we're building a new class of enterprise infrastructure that companies need to deploy and scale agentic AI across hybrid data centers."

Server Market Revenue Increased 91% During the Q4 2024, NVIDIA Continues Dominating the GPU Server Space

According to the International Data Corporation (IDC) Worldwide Quarterly Server Tracker, the server market reached a record $77.3 billion dollars in revenue during the last quarter of the year. This quarter showed the second highest growth rate since 2019 with a year-over-year increase of 91% in vendor revenue. Revenue generated from x86 servers increased 59.9% in 2024Q4 to $54.8 billion while Non-x86 servers increased 262.1% year over year to $22.5 billion.

Revenue for servers with an embedded GPU in the fourth quarter of 2024 grew 192.6% year-over-year and for the full year 2024, more than half of the server market revenue came from servers with an embedded GPU. NVIDIA continues dominating the server GPU space with over 90% of the total shipments with and embedded GPU in 2024Q4. The fast pace at which hyperscalers and cloud service providers have been adopting servers with embedded GPUs has fueled the server market growth which has more than doubled in size since 2020 with revenue of $235.7 billion dollars for the full year 2024.

MSI Powers the Future of Cloud Computing at CloudFest 2025

MSI, a leading global provider of high-performance server solutions, unveiled its next-generation server platforms—ORv3 Servers, DC-MHS Servers, and NVIDIA MGX AI Servers—at CloudFest 2025, held from March 18-20 at booth H02. The ORv3 Servers focus on modularity and standardization to enable seamless integration and rapid scalability for hyperscale growth. Complementing this, the DC-MHS Servers emphasize modular flexibility, allowing quick reconfiguration to adapt to diverse data center requirements while maximizing rack density for sustainable operations. Together with NVIDIA MGX AI Servers, which deliver exceptional performance for AI and HPC workloads, MSI's comprehensive solutions empower enterprises and hyperscalers to redefine cloud infrastructure with unmatched flexibility and performance.

"We're excited to present MSI's vision for the future of cloud infrastructure." said Danny Hsu, General Manager of MSI's Enterprise Platform Solutions. "Our next-generation server platforms address the critical needs of scalability, efficiency, and sustainability. By offering modular flexibility, seamless integration, and exceptional performance, we empower businesses, hyperscalers, and enterprise data centers to innovate, scale, and lead in this cloud-powered era."
Return to Keyword Browsing
May 9th, 2025 17:40 EDT change timezone

New Forum Posts

Popular Reviews

Controversial News Posts