News Posts matching #Benchmark

Return to Keyword Browsing

AMD's Radeon RX 7900 GRE Gets Benchmarked

AMD's China exclusive Radeon RX 7900 GRE has been put through its paces by Expreview and the US$740 equivalent card should in short not carry the 7900-series moniker. In most of the tests, the card performs like a Raden RX 6950 XT or worse, with it being beaten by the Radeon RX 6800 XT in 3D Mark Fire Strike, even if it's only by the tiniest amount. Expreview has done a fairly limited comparison, mainly pitching the Radeon RX 7900 GRE against the Radeon RX 7900 XT and NVIDIA's GeForce RTX 4070, where it loses by a mile towards AMD's higher-end GPU, which by no means was unexpected as this is a lower tier product.

However, when it comes the GeForce RTX 4070, AMD struggles to keep up at 1080p, where NVIDIA takes home the win in games like The Last of US Part 1 and Diablo 4. In games like F1 22 and Assassin's Creed Hall of Valor, AMD is only ahead by a mere percentage point or less. Once ray tracing is enabled, AMD only wins in F1 22 and it's by less than one percent again and Far Cry 6, where AMD is almost three percent faster. Moving up in resolution, the Radeon RX 7900 GRE ends up being a clear winner, most likely partially due to having 16 GB of VRAM and at 1440p the GeForce RTX 4070 also falls behind in most of the ray traced game tests, if only just in most of them. At 4K the NVIDIA card can no longer keep up, but the Radeon RX 7900 GRE isn't really a 4K champion either, dropping under 60 FPS in more resource heavy games like Cyberpunk 2077 and The Last of Us Part 1. Considering the GeForce RTX 4070 Ti only costs around US$50 more, it seems like it would be the better choice, despite having less VRAM. AMD appears to have pulled an NVIDIA with this card, which at least performance wise, seems to belong in the Radeon RX 7800 segment. The benchmark figures also suggests that the actual Radeon RX 7800 cards won't be worth the wait, unless AMD prices them very competitively.

Update 11:45 UTC: [Editor's note: The official MSRP from AMD appears to be US$649 for this card, which is more reasonable, but the performance still places this in in a category lower than the model name suggests.]

AMD's Ryzen 5 7500F Gets Benchmarked, Available Globally

AMD's recently added Ryzen 5 7500F for the AM5 socket was initially said to only be available in the PRC, but according to AMD, it will apparently be available globally. That said, AMD apparently only seeded review units to select Asian media, among them Quasar Zone in Korea, who put the six core, 12 thread CPU through its paces. Overall performance is very close to the Ryzen 5 7600, which isn't really all that strange, considering the two only differ by 100 MHz in both base and boost clock. In most of the benchmarks, the Ryzen 5 7500F is around two to three percent slower than the Ryzen 5 7600 on average.

When compared to the slightly more pricey Intel Core i5-13400, AMD falls behind multithreaded apps but comes out on top in most of the games tested, with the usual odd exception as would be expected. On average, the Ryzen 5 7500F is some 13 percent faster in the game benchmarks at 1080p, although this is using an NVIDIA GeForce RTX 4090 graphics card. It even beats the overall much faster Intel Core i5-13500 in gaming by around nine percent on average. However, the Ryzen 5 7500F system loses out to the two Intel systems when it comes to power efficiency, drawing around 20 Watts more on average when gaming. At US$179.99 it seems like AMD finally has a budget friendly CPU for the AM5 platform, if you're willing to lose the integrated GPU. It's unknown when the CPU will be available outside of Asia at this point in time.

Leaked AMD Radeon RX 7700 & RX 7800 GPU Benchmarks Emerge

A set of intriguing 3DMark Time Spy benchmark results have been released by hardware leaker All_The_Watts!!—these are alleged to have been produced by prototype Radeon RX 7700 and Radeon RX 7800 graphics cards (rumored to be based on variants of the Navi 32 GPU). The current RDNA 3 lineup of mainstream GPUs is severely lacking in middle ground representation, but Team Red is reported to be working on a number of models to fill in the gap. We expect a number of leaks to emerge as we get closer to a rumored product reveal scheduled for late August (to coincide with Gamescon).

The recently released 3DMark Time Spy scores reveal that the alleged Radeon RX 7700 candidate scored 15,465 points, while the RX 7800 achieved 18,197 points—both running on an unspecified test system. The results (refer to the Tom's Hardware-produced chart placed below) are not going to generate a lot of excitement at this stage when compared to predecessors and some of the competition—evaluation samples are not really expected to be optimized to a great degree. We hope to see finalized products with decent drivers putting in a good appearance and performing better later on this year.

Denuvo Setting Up Benchmarking System, Attempting to Disprove Performance Shortfalls

Irdeto is the current owner of Denuvo Software Solutions—the Austrian development team behind the infamous anti-tamper technology and digital rights management (DRM) system. According to Ars Technica neither of these organizations have made great efforts (in the past) to engage in discussion about the controversial anti-piracy and anti-cheat suites—but Steeve Huin, Irdeto's Chief Operating Officer of Video Games—agreed to grant the publication an exclusive interview. The article is titled "Denuvo wants to convince you its DRM isn't evil," which sums up a lot of the public perception regarding Denuvo technologies—having received plenty of flak for high CPU usage and causing excessive activity within storage components. Some users propose that the latter scenario has resulted in shorter lifespans for their solid-state drives. Ars Technica has a long history of Denuvo-related coverage, so a company representative has been sent in for some damage control.

Off the bat, Huin acknowledges that he and his colleagues are aware of Denuvo's reputation: "In the pirating/cracking community, we're seen as evil because we're helping DRM exist and we're ensuring people make money out of games." He considers the technology to be a positive force: "Anti-piracy technologies is to the benefit of the game publishers, [but also] is of benefit to the players in that it protects the [publisher's] investment and it means the publishers can then invest in the next game...But people typically don't think enough of that...Whether people want to believe it or not, we are all gamers, we love gaming, we love being part of it. We develop technologies with the intent to make the industry better and stronger."

AMD Ryzen 5 7500F CPU Gets Benchmarked

The Puget Systems benchmark database outed AMD's Ryzen 5 7500F 6-core/12-thread processor last week—industry experts proposed that it was the first example of a Ryzen 7000 SKU with a disabled iGPU. A South Korean retailer indicated unit pricing of around $170-180, with a possible local launch date on July 7. It seems that retail units have not hit the market (at the time of writing), but Geekbench 6.1 results have since appeared online. According to an entry on the Geekbench database—that was spotted by Olrak29 earlier today—the Ryzen 5 7500F has a base clock of 3.7 GHz. It can boost up to 5.0 GHz on a single core, while all cores can reach a maximum of 4.8 GHz. The listing confirms that this new SKU sits firmly in the AMD "Raphael" CPU family.

The processor was tested on a system running Microsoft Windows 11—partial specifications of the evaluation build include an ASUS TUF Gaming A620M-PLUS WIFI motherboard and 32 GB of DDR5-6000 RAM. The tested Ryzen 5 7500F CPU achieved scores of 2782 points (single-core) and 13323 points (multi-threaded), which places it slightly ahead of the Ryzen 5 7600X in multi-thread performance. It trails slightly behind with its single-core result, but these figures are impressive considering that the Ryzen 5 7500F will likely be offered at a more budget friendly price when compared to its closest iGPU-enabled siblings.

NVIDIA H100 GPUs Set Standard for Generative AI in Debut MLPerf Benchmark

In a new industry-standard benchmark, a cluster of 3,584 H100 GPUs at cloud service provider CoreWeave trained a massive GPT-3-based model in just 11 minutes. Leading users and industry-standard benchmarks agree: NVIDIA H100 Tensor Core GPUs deliver the best AI performance, especially on the large language models (LLMs) powering generative AI.

H100 GPUs set new records on all eight tests in the latest MLPerf training benchmarks released today, excelling on a new MLPerf test for generative AI. That excellence is delivered both per-accelerator and at-scale in massive servers. For example, on a commercially available cluster of 3,584 H100 GPUs co-developed by startup Inflection AI and operated by CoreWeave, a cloud service provider specializing in GPU-accelerated workloads, the system completed the massive GPT-3-based training benchmark in less than eleven minutes.

MLCommons Shares Intel Habana Gaudi2 and 4th Gen Intel Xeon Scalable AI Benchmark Results

Today, MLCommons published results of its industry AI performance benchmark, MLPerf Training 3.0, in which both the Habana Gaudi2 deep learning accelerator and the 4th Gen Intel Xeon Scalable processor delivered impressive training results.

"The latest MLPerf results published by MLCommons validates the TCO value Intel Xeon processors and Intel Gaudi deep learning accelerators provide to customers in the area of AI. Xeon's built-in accelerators make it an ideal solution to run volume AI workloads on general-purpose processors, while Gaudi delivers competitive performance for large language models and generative AI. Intel's scalable systems with optimized, easy-to-program open software lowers the barrier for customers and partners to deploy a broad array of AI-based solutions in the data center from the cloud to the intelligent edge." - Sandra Rivera, Intel executive vice president and general manager of the Data Center and AI Group

Geekbench Leak Suggests NVIDIA GeForce RTX 4060 Nearly 20% Faster than RTX 3060

NVIDIA is launching its lower end GeForce RTX 4060 graphics card series next week, but has kept schtum about the smaller Ada Lovelace AD107 GPU's performance level. This more budget-friendly offering (MSRP $299) is rumored to have 3,072 CUDA cores, 24 RT cores, 96 Tensor cores, 96 TMUs, and 32 ROPs. It will likely sport 8 GB of GDDR6 memory across a 128-bit wide memory bus. Benchleaks has discovered the first set of test results via a database leak, and posted these details on social media earlier today. Two Geekbench 6 runs were conducted on a test system comprised of an Intel Core i5-13600K CPU, ASUS Z790 ROG APEX motherboard, DDR5-6000 memory and the aforementioned GeForce card.

The GPU Compute test utilizing the Vulkan API resulted in a score of 99419, and another using OpenCL achieved 105630. We are looking at a single sample here, so expect variations when other units get tested in Geekbench prior to the June 29 launch. The RTX 4060 is about 12% faster (in Vulkan) than its direct predecessor—RTX 3060. The gap widens with its Open CL performance, where it offers an almost 20% jump over the older card. The RTX 3060 Ti presents around 3-5% faster performance over the RTX 4060. We hope to see actual in-game benchmarking carried out soon.

NVIDIA H100 Hopper GPU Tested for Gaming, Slower Than Integrated GPU

NVIDIA's H100 Hopper GPU is a device designed for pure AI and other compute workloads, with the least amount of consideration for gaming workloads that involve graphics processing. However, it is still interesting to see how this 30,000 USD GPU fairs in comparison to other gaming GPUs and whether it is even possible to run games on it. It turns out that it is technically feasible but not making much sense, as the Chinese YouTube channel Geekerwan notes. Based on the GH100 GPU SKU with 14,592 CUDA, the H100 PCIe version tested here can achieve 204.9 TeraFLOPS at FP16, 51.22 TeraFLOPS at FP32, and 25.61 TeraFLOPS at FP64, with its natural power laying in accelerating AI workloads.

However, how does it fare in gaming benchmarks? Not very well, as the testing shows. It scored 2681 points in 3DMark Time Spy, which is lower than AMD's integrated Radeon 680M, which managed to score 2710 points. Interestingly, the GH100 has only 24 ROPs (render output units), while the gaming-oriented GA102 (highest-end gaming GPU SKU) has 112 ROPs. This is self-explanatory and provides a clear picture as to why the H100 GPU is used for computing only. Since it doesn't have any display outputs, the system needed another regular GPU to provide the picture, while the computation happened on the H100 GPU.

Primate Labs Rolls Out Geekbench 6.1

Primate Labs has released the newest update to its cross-platform CPU and GPU benchmark that measures your system's performance, Geekbench 6.1. The latest version brings new features and improvements, including the upgrade to Clang 16, an increased workload gap that should minimize thermal throttling on some devices, as well as introduces support for SVE and AVX 512- FP 16 instructions, and support for fixed-point math. The update also improves multi-core performance.

These changes result in Geekbench 6.1 single-scores to be up to 5 percent higher and multi-core scores up to 10 percent higher, compared to Geekbench 6.0 scores. Due to these differences, Primate Labs recommends that users do not compare scores between Geekbench 6.0 and Geekbench 6.1. Geekbench 6.1 is also a recommended update, according to Primate Labs.

Capcom Releases Street Fighter 6 PC Benchmark Tool

Capcom has kindly provided a new benchmarking tool for folks who are wondering whether their souped-up PC gaming rigs can run the upcoming seventh (not sixth, despite the title) main entry in the Street Fighter series with aplomb - the testing suite can be downloaded from here. The development team's introductory message states: "The Street Fighter 6 Benchmark Tool monitors gameplay on the Street Fighter 6 demo and calculates a score that indicates your PC's performance. The results of the benchmarking will be shown as follows, with a score of 91 or above demonstrating that your PC can play the game with ease."

The explanation continues: "If your PC does not meet the system requirements needed to run this benchmarking software, it may not launch properly. If that happens, please reconfirm that you satisfy the criteria listed on this page under System Requirements." Street Fighter 6 is arriving this Friday (June 2), so Capcom's benchmarking tool only gives a little bit of advanced notice - an unfortunate few who "cannot operate the game" (with a 0-30 score) will need to make necessary PC upgrades in time for launch day action. Or they could simply buy the bare minimum point of entry on console: a PlayStation 4 Slim or the cheapest current generation system - Xbox Series S.

Intel Core Ultra 7 1003H CPU Benchmark Results Appear Online

A hardware tipster - Benchleaks - has highlighted an interesting new entry for an Intel Meteor Lake Client Platform CPU on the PugetBench site - it seems that early platform validation results have been uploaded by mistake (prior to embargo restrictions). The MTL-P CPU in question appears to be a laptop/mobile variant given its "H" designation. We are also looking at another example of Team Blue's new SKU naming system with this Core Ultra 7 1003H processor - the company has confirmed that revised branding will debut as part of the Meteor Lake family.

The previously leaked Core model (5 1003H) also sported an "Ultra" tag, so it is possible that only high-end examples have been outed online over the past month. Puget System Lightroom Classic benchmark results that were produced by the Core Ultra 7 1003H CPU were not exactly next level - scoring only 534.5 points overall - this could indicate that a prototype unit was benched. An older Core i7-8665U laptop processor only lagged behind by 32.5 points. The test platform was fitted with 16 GB (2 x 8 GB) of DDR5-5600 memory, and ran in a Windows 11 Enterprise (22621) OS environment. Intel's latest marketing spiel is bigging up the potential of Meteor Lake's AI acceleration capabilities, via the built-in neural VPU.

ASUS ROG Ally's Ryzen Z1 Extreme Custom APU Verified by Benchmark Info

An intriguing entry has appeared on the Geekbench Browser site, the information was uploaded with a timestamp from this morning (11:07 am on April 20 to be specific) pointing to a mobile ASUS device that was tested in GeekBench 5. The archived info dump reveals that the subject matter of the benchmark was the ASUS ROG Ally handheld gaming console, which has received a lot of attention in recent weeks - with it being touted as a very serious alternative to Valve's Steam Deck, a handheld gaming PC that is quite popular with enthusiasts. The ROG Ally will need to offer a potent hardware package if it stands to compete directly with the Steam Deck, and the latest information confirms that this new contender is very promising in that department. Geekbench 5 awarded an impressive OpenCL score of 35498 to the RC71L variant of the ROG Ally, an RC71X-assigned model is known to exist but details of its exact nature have not been revealed. This particular ROG Alloy unit was running Windows 11 Home (64-bit) under the operating system's performance power plan.

The new entry on Geekbench Browser shows that the Ally is packing an AMD Ryzen Z1 Extreme APU, which appears to be a customized version of the Ryzen 7 7840U APU mobile platform chipset - previous rumors have indicated that the latter would be in the driving seat. Both Phoenix range SoCs share the basic 8 cores and 16 thread makeup, but the Z1 Extreme is capable of boosting up to 5.062 GHz from a base frequency of 3.30 GHz. AMD's Radeon 780M iGPU (RDNA 3) is expected to deal with the Ally's graphical aspect, but the benchmark info dump only provides scant details about the GPU (codenamed "gfx1103") - most notably the presence of six computer units, an 800 MHz max frequency, and access to 8.20 GB of video memory. Number crunching boffins have calculated that the Ally could field 768 FP32 cores, courtesy of the dual issue SIMD design inherent to RDNA 3.

3DMark Gets AMD FidelityFX Super Resolution 2 (FSR 2) Feature Test

UL Benchmarks today released an update to 3DMark that adds a Feature Test for AMD FidelityFX Super Resolution 2 (FSR 2), the company's popular upscaling-based performance enhancement. This was long overdue, as 3DMark has had a Feature Test for DLSS for years now; and as of October 2022, it even got one for Intel XeSS. The new FSR 2 Feature Test uses a scene from the Speed Way DirectX 12 Ultimate benchmark, where it compares fine details of a vehicle and a technic droid between native resolution with TAA and FSR 2, and highlights the performance uplift. To use the feature test, you'll need any GPU that supports DirectX 12 and FSR 2 (that covers AMD, NVIDIA, and Intel Arc). For owners of 3DMark who purchased it before October 12, 2022, they'll need to purchase the Speed Way upgrade to unlock the AMD FSR feature test.

OpenAI Unveils GPT-4, Claims to Outperform Humans in Certain Academic Benchmarks

We've created GPT-4, the latest milestone in OpenAI's effort in scaling up deep learning. GPT-4 is a large multimodal model (accepting image and text inputs, emitting text outputs) that, while less capable than humans in many real-world scenarios, exhibits human-level performance on various professional and academic benchmarks. For example, it passes a simulated bar exam with a score around the top 10% of test takers; in contrast, GPT-3.5's score was around the bottom 10%. We've spent 6 months iteratively aligning GPT-4 using lessons from our adversarial testing program as well as ChatGPT, resulting in our best-ever results (though far from perfect) on factuality, steerability, and refusing to go outside of guardrails.

Over the past two years, we rebuilt our entire deep learning stack and, together with Azure, co-designed a supercomputer from the ground up for our workload. A year ago, we trained GPT-3.5 as a first "test run" of the system. We found and fixed some bugs and improved our theoretical foundations. As a result, our GPT-4 training run was (for us at least!) unprecedentedly stable, becoming our first large model whose training performance we were able to accurately predict ahead of time. As we continue to focus on reliable scaling, we aim to hone our methodology to help us predict and prepare for future capabilities increasingly far in advance—something we view as critical for safety.

AMD Ryzen 9 7945HX Beats Core i9-13950HX In Gaming Performance, Dragon Range Equipped Laptops Available Now

AMD has announced the immediate availability of its Ryzen 7045 HX-series (Dragon Range) processors for high performance laptop devices. In a Youtube video released on March 10, AMD's Jason Banta has announced the availability of the world's most powerful mobile processor, the Ryzen 9 7945HX. He listed the OEM partners who have integrated the 7945HX into flagship level laptop models. He also declared that this range topping CPU is a competition beater. Gaming benchmark tests have demonstrated that the Ryzen 9 7945HX beats Intel's Raptor Lake Core i9-13950HX by an average margin of 10%.

Intel Xeon W-3400/2400 "Sapphire Rapids" Processors Run First Benchmarks

Thanks to the attribution of Puget Systems, we have a preview of Intel's latest Xeon W-3400 and Xeon W-2400 workstation processors based on Sapphire Rapids core technology. Delivering up to 56 cores and 112 threads, these CPUs are paired with up to eight TeraBytes of eight-channel DDR5-4800 memory. For expansion, they offer up to 112 PCIe 5.0 lanes come with up to 350 Watt TDP; some models are unlocked for overclocking. This interesting HEDT family for workstation usage comes at a premium with an MSRP of $5,889 for the top-end SKU, and motherboard prices are also on the pricey side. However, all of this should come as no surprise given the expected performance professionals expect from these chips. Puget Systems has published test results that include: Photoshop, After Effects, Premiere Pro, DaVinci Resolve, Unreal Engine, Cinebench R23.2, Blender, and V-Ray. Note that Puget Systems said that: "While this post has been an interesting preview of the new Xeon processors, there is still a TON of testing we want to do. The optimizations Intel is working on is of course at the top, but there are several other topics we are highly interested in." So we expect better numbers in the future.
Below, you can see the comparison with AMD's competing Threadripper Pro HEDT SKUs, along with power usage using different Windows OS power profiles:

Intel Publishes Sorting Library Powered by AVX-512, Offers 10-17x Speed Up

Intel has recently updated its open-source C++ header file library for high-performance SIMD-based sorting to support the AVX-512 SIMD instruction set. Extending the capability of regular AVX2 support, the sorting functions now implement 512-bit extensions to offer greater performance. According to Phoronix, the NumPy Python library for mathematics that underpins a lot of software has updated its software base to use the AVX-512 boosted sorting functionality that yields a fantastic uplift in performance. The library uses AVX-512 to vectorize the quicksort for 16-bit and 64-bit data types using the extended instruction set. Benchmarked on an Intel Tiger Lake system, the NumPy sorting saw a 10-17x increase in performance.

Intel's engineer Raghuveer Devulapalli changed the NumPy code, which was merged into the NumPy codebase on Wednesday. Regarding individual data types, the new implementation increases 16-bit int sorting by 17x and 32-bit data type sorting by 12-13x, while float 64-bit sorting for random arrays has experienced a 10x speed up. Using the x86-simd-sort code, this speed-up shows the power of AVX-512 and its capability to enhance the performance of various libraries. We hope to see more implementations of AVX-512, as AMD has joined the party by placing AVX-512 processing elements on Zen 4.

Alleged NVIDIA AD106 GPU Tested in 3DMark and AIDA64

Benchmarks and specifications of an alleged NVIDIA AD106 GPU have tipped up on Chiphell, although the original poster has since removed all the details. Thanks to @harukaze5719 on Twitter, who posted the details, we still get an insight into what we might be able to expect from NVIDIA's upcoming mid-range cards. All these details should be taken as is, as the original source isn't exactly what we'd call trustworthy. Based on the data in the TPU GPU database, the GPU in question should be the GeForce RTX 4070 Mobile with much higher clock speeds or an equivalent desktop part that offers more CUDA cores than the RTX 4060 Ti. Whatever the specific AD106 GPU is, it's being compared to the GeForce RTX 2080 Super and the RTX 3070 Ti.

The GPU was tested in AIDA64 and 3DMark and it beats the RTX 2080 Super in all of the tests, while drawing some 55 W less power at the same time. In some of the benchmarks the wins are within the margin of testing error, for example when it comes to the memory performance in AIDA64. However, we're looking at a GPU connected to only half the memory bandwidth here, as the AD106 GPU only has a 128-bit memory bus, compared to 256-bit for the RTX 2080 Super, although the memory clocks are much higher, but the overall memory bandwidth is still nearly 36 percent higher in the RTX 2080 Super. Yet, the AD106 GPU manages to beat the RTX 2080 Super in all of the memory benchmarks in AIDA64.

BAPCo Releases SYSmark 30, the Latest Generation of the Premier PC Performance Metric Featuring New Applications and Scenarios

BAPCo, a non-profit consortium of leading PC hardware manufacturers, released SYSmark 30, the latest generation of the premier PC benchmark that measures and compares system performance using real-world applications and workloads.

The Office Application scenario features updated workloads for popular office suite-style applications. The General Productivity scenario features tasks like web browsing, file compression, and application installation. The new Photo Editing scenarios measure the responsiveness of creative photo management and manipulation usage models. The Advanced Content Creation scenario heavily uses photo and video editing applications, including multitasking.

First Alleged AMD Radeon RX 7900-series Benchmarks Leaked

With only a couple of days to go until the AMD RX 7900-series benchmarks go live, some alleged benchmarks from both the RX 7900 XTX and RX 7900 XT have leaked on Twitter. The two cards are being compared to a NVIDIA RTX 4080 card in no less than seven different game titles, all running at 4K resolution. The games are God of War, Cyberpunk 2077, Assassin's Creed Valhalla, Watchdogs Legion, Red Dead Redemption 2, Doom Eternal and Horizon Zero Dawn. The cards were tested on a system with a Core i9-12900K CPU which was paired with 32 GB of RAM of unknown type.

It's too early to draw any real conclusions from this test, but in general, the RX 7900 XTX comes out on top, ahead of the RTX 4080, so no surprises here. The RX 7900 XT is either tied with the RTX 4080 or a fair bit slower, with the exception being Red Dead Redemption 2, where the RTX 4080 is the slowest card, although it also appears to have some issues, since the one percent lows are hitting 2 FPS. Soon, the reviews will be out and everything will become more clear, but it appears that AMD's RX 7900 XTX will give NVIDIA's RTX 4080 a run for its money, if these benchmarks are anything to go by.

Update Dec 11th: The original tweet has been removed, for unknown reasons. It could be because the numbers were fake, or because they were in breach of AMD's NDA.

AMD 4th Generation EPYC "Genoa" Processors Benchmarked

Yesterday, AMD announced its latest addition to the data center family of processors called EPYC Genoa. Named the 4th generation EPYC processors, they feature a Zen 4 design and bring additional I/O connectivity like PCIe 5.0, DDR5, and CXL support. To disrupt the cloud, enterprise, and HPC offerings, AMD decided to manufacture SKUs with up to 96 cores and 192 threads, an increase from the previous generation's 64C/128T designs. Today, we are learning more about the performance and power aspects of the 4th generation AMD EPYC Genoa 9654, 9554, and 9374F SKUs from 3rd party sources, and not the official AMD presentation. Tom's Hardware published a heap of benchmarks consisting of rendering, compilation, encoding, parallel computing, molecular dynamics, and much more.

In the comparison tests, we have AMD EPYC Milan 7763, 75F3, and Intel Xeon Platinum 8380, a current top-end Intel offering until Sapphire Rapids arrives. Comparing 3rd-gen EPYC 64C/128T SKUs with 4th-gen 64C/128T EPYC SKUs, the new generation brings about a 30% increase in compression and parallel compute benchmarks performance. When scaling to the 96C/192T SKU, the gap is widened, and AMD has a clear performance leader in the server marketplace. For more details about the benchmark results, go here to explore. As far as comparison to Intel offerings, AMD leads the pack as it has a more performant single and multi-threaded design. Of course, beating the Sapphire Rapids to market is a significant win for team red, so we are still waiting to see how the 4th generation Xeon stacks up against Genoa.

Intel Delivers Leading AI Performance Results on MLPerf v2.1 Industry Benchmark for DL Training

Today, MLCommons published results of its industry AI performance benchmark in which both the 4th Generation Intel Xeon Scalable processor (code-named Sapphire Rapids) and Habana Gaudi 2 dedicated deep learning accelerator logged impressive training results.


"I'm proud of our team's continued progress since we last submitted leadership results on MLPerf in June. Intel's 4th gen Xeon Scalable processor and Gaudi 2 AI accelerator support a wide array of AI functions and deliver leadership performance for customers who require deep learning training and large-scale workloads." Sandra Rivera, Intel executive vice president and general manager of the Datacenter and AI Group

NVIDIA RTX 4080 20-30% Slower than RTX 4090, Still Smokes the RTX 3090 Ti: Leaked Benchmarks

Benchmarks of NVIDIA's upcoming GeForce RTX 4080 (formerly known as the RTX 4080 16 GB) are already out as the leaky taps in the Asian tech forumscape know no bounds. Someone with access to an RTX 4080 sample and drivers on ChipHell forums, put it through a battery of synthetic and gaming tests. The $1,200 MSRP graphics card was tested on 3DMark Time Spy, Port Royal, and games that include Forza Horizon 5, Call of Duty Modern Warfare II, Cyberpunk 2077, Borderlands 3, and Shadow of the Tomb Raider.

The big picture: the RTX 4080 is found to be halfway between the RTX 3090 Ti and the RTX 4090. At stock settings, and in 3DMark Time Spy Extreme (4K), it has 71% the performance of an RTX 4090, whereas the RTX 3090 Ti is 55% that of the RTX 4090. With its "power limit" slider maxed out, the RTX 4080 inches 2 percentage-points closer to the RTX 4090 (73% that of the RTX 4090), and with a bit of manual OC, it adds another 4 percentage-points. Things change slightly with 3DMark Port Royal, where the RTX 4080 is 69% the performance of the RTX 4090 in a test where the RTX 3090 Ti does 58% that of the RTX 4090.

Basemark Debuts a Unique Benchmark for Comparisons Between Android, iOS, Linux, MacOS and Windows Devices

Basemark launched today GPUScore Sacred Path. It is the world's only cross-platform GPU benchmark that includes the latest GPU technologies like Variable Rate Shading (VRS). Sacred Path supports all the relevant device categories - ranging from premium mobile phones to high-end gaming PCs and discrete graphics cards, including full support of the major operating systems, such as Android, iOS, Linux, macOS and Windows.

This benchmark is of great importance for application vendors, device manufacturers, GPU vendors and IT Media. Game developers need a thorough understanding of performance across the device range to optimize the use of the same assets across a maximum device range. GPU vendors and device manufacturers can compare their products with competitor products, which allows them to develop new product ranges with the correct targeting. In addition, Sacred Path is a true asset for media reviewing any GPU-equipped devices.
Return to Keyword Browsing
Dec 4th, 2024 03:35 EST change timezone

New Forum Posts

Popular Reviews

Controversial News Posts