News Posts matching #GPU

Return to Keyword Browsing

Microsoft Working on Custom AI Processor Codenamed Project Athena

According to The Information, Microsoft has been working on creating custom processors for processing AI with a project codenamed Athena. Based on TSMC's 5 nm process, these chips are designed to accelerate AI workloads and scale to hundreds or even thousands of chips. With the boom of Large Language Models (LLMs) that require billions of parameters, training them requires a rapid increase of computational power to a point where companies purchase hundreds of thousands of GPUs from the likes of NVIDIA. However, creating custom processors is a familiar feat for a company like Microsoft. Hyperscalers like AWS, Google, and Meta are already invested in the creation of processors for AI training, and Microsoft is just joining as well.

While we don't have much information about these processors, we know that Microsoft started the project in 2019, and today these processors are in the hands of select employees of Microsoft and OpenAI that work with AI projects and need computational horsepower. Interestingly, some projections assume that if Microsoft could match NVIDIA's GPU performance, the cost would only be a third of NVIDIA's offerings. However, it is challenging to predict that until more information is provided. Microsoft plans to make these chips more widely available as early as next year; however, there is no specific information on when and how, but Azure cloud customers would be the most logical place to start.

HBM Supply Leader SK Hynix's Market Share to Exceed 50% in 2023 Due to Demand for AI Servers

A strong growth in AI server shipments has driven demand for high bandwidth memory (HBM). TrendForce reports that the top three HBM suppliers in 2022 were SK hynix, Samsung, and Micron, with 50%, 40%, and 10% market share, respectively. Furthermore, the specifications of high-end AI GPUs designed for deep learning have led to HBM product iteration. To prepare for the launch of NVIDIA H100 and AMD MI300 in 2H23, all three major suppliers are planning for the mass production of HBM3 products. At present, SK hynix is the only supplier that mass produces HBM3 products, and as a result, is projected to increase its market share to 53% as more customers adopt HBM3. Samsung and Micron are expected to start mass production sometime towards the end of this year or early 2024, with HBM market shares of 38% and 9%, respectively.

AI server shipment volume expected to increase by 15.4% in 2023
NVIDIA's DM/ML AI servers are equipped with an average of four or eight high-end graphics cards and two mainstream x86 server CPUs. These servers are primarily used by top US cloud services providers such as Google, AWS, Meta, and Microsoft. TrendForce analysis indicates that the shipment volume of servers with high-end GPGPUs is expected to increase by around 9% in 2022, with approximately 80% of these shipments concentrated in eight major cloud service providers in China and the US. Looking ahead to 2023, Microsoft, Meta, Baidu, and ByteDance will launch generative AI products and services, further boosting AI server shipments. It is estimated that the shipment volume of AI servers will increase by 15.4% this year, and a 12.2% CAGR for AI server shipments is projected from 2023 to 2027.

Colorful Custom RTX 4060 Ti GPU Clocks Outed, 8 GB VRAM Confirmed

Resident TechPowerUp hardware database overseer T4C Fantasy has divulged some early information about a custom version of the NVIDIA GeForce RTX 4060 Ti GPU card - Colorful's catchily named iGame RTX 4060 Ti Ultra White OC model has been added to the TPU GPU database, and T4C Fantasy has revealed a couple of tidbits on Twitter. The GPU has been tuned to have a maximum boost clock of 2580 MHz, jumping from a base clock of 2310 MHz. According to past leaks the reference version of the GeForce RTX 4060 Ti has a default boost clock of 2535 MHz, so Colorful's engineers have managed to add another 45 MHz on top of that with their custom iteration - so roughly 2% more than the reference default.

T4C Fantasy also confirmed that the Colorful iGame RTX 4060 Ti Ultra W OC will be appointed with 8 GB of VRAM, which also matches the reference model's rumored memory spec. T4C Fantasy points out that brands have the option to produce RTX 4060 Ti cards with a larger pool of attached video memory, but launch models will likely stick with the standard allotment of 8 GB of VRAM. The RTX 4060 Ti is listed as being based on the Ada Lovelace GPU architecture (GPU variant AD106-350-A1), and T4C Fantasy expects that Team Green will stick with a 5 nm process size - contrary to reports of a transition to manufacturing on 4 nm (chez TSMC foundries).

AMD Brings ROCm to Consumer GPUs on Windows OS

AMD has published an exciting development for its Radeon Open Compute Ecosystem (ROCm) users today. Now, ROCm is coming to the Windows operating system, and the company has extended ROCm support for consumer graphics cards instead of only supporting professional-grade GPUs. This development milestone is essential for making AMD's GPU family more competent with NVIDIA and its CUDA-accelerated GPUs. For those unaware, AMD ROCm is a software stack designed for GPU programming. Similarly to NVIDIA's CUDA, ROCm is designed for AMD GPUs and was historically limited to Linux-based OSes and GFX9, CDNA, and professional-grade RDNA GPUs.

However, according to documents obtained by Tom's Hardware (which are behind a login wall), AMD has brought support for ROCm to Radeon RX 6900 XT, Radeon RX 6600, and R9 Fury GPU. What is interesting is not the inclusion of RX 6900 XT and RX 6600 but the support for R9 Fury, an eight-year-old graphics card. Also, what is interesting is that out of these three GPUs, only R9 Fury has full ROCm support, the RX 6900 XT has HIP SDK support, and RX 6600 has only HIP runtime support. And to make matters even more complicated, the consumer-grade R9 Fury GPU has full ROCm support only on Linux and not Windows. The reason for this strange selection of support has yet to be discovered. However, it is a step in the right direction, as AMD has yet to enable more functionality on Windows and more consumer GPUs to compete with NVIDIA.

Snapdragon 8 Gen 3 GPU Could be 50% More Powerful Than Current Gen Adreno 740

An online tipster, posting on the Chinese blog site Weibo, has let slip that Qualcomm's upcoming Snapdragon 8 Gen 3 mobile chipset is touted to pack some hefty graphical capabilities. The suggested Adreno "750" smartphone and tablet GPU is touted to offer a 50% increase over the present generation Adreno 740 - as featured on the recently released and cutting-edge Snapdragon 8 Gen 2 chipset. The current generation top-of-the-range Snapdragon is no slouch when it comes to graphics benchmarks, where it outperforms Apple's prime contender - the Bionic A16 SoC.

The Snapdragon 8 Gen 3 SoC is expected to launch in the last quarter of 2023, but details of the flagship devices that it will power are non-existent at the time of writing. The tipster suggests that Qualcomm has decided to remain on TSMC's 4 nm process for its next generation mobile chipset - perhaps an all too safe decision when you consider that Apple has upped the stakes with the approach of its Bionic A17 SoC. It has been reported that the Cupertino, California-based company has chosen to fabricate via TSMC's 3 nm process, although the Taiwanese foundry is said to be struggling with its N3 production line. The engineers at Qualcomm's San Diego headquarters are alleged to be experimenting with increased clock speeds running on the next gen Adreno GPU - as high as 1.0 GHz - in order to eke out as much performance as possible, in anticipation of besting the Bionic A17 in graphics benchmarks. The tipster theorizes that Qualcomm will still have a hard time matching Apple in terms of pure CPU throughput, so the consolation prize will lie with a superior GPU getting rigged onto the Snapdragon 8 Gen 3.

Alphacool Expands Enterprise Solutions with Water Blocks for A100 80 GB PCIe, RTX A4000, and RTX 6000 Ada 48 GB SKUs

Alphacool expands the portfolio of the Enterprise Solutions series for GPU water coolers and presents the new ES NV A100 80 GB PCIe, ES RTX A4000 with backplate and ES RTX 6000 Ada 48 GB.

To best dissipate the enormous waste heat of this GPU generation, the cooler is positioned close to the components to be cooled in an exemplary manner. The fin structure has been adapted and allows a very good water flow while increasing the cooling surface. The modified jetplate with improved inflow engine ensures optimal distribution of water on the cooling fins. The fully chromed copper base is resistant to acids, scratches and damages. The matte carbon finish gives the cooler a noble appearance. At the same time, this makes it interesting for private users who want to do without aRGB lighting.

AMD Makes Radeon Pro W7900 & W7800 Workstation GPUs Official

AMD unveils the most powerful AMD Radeon Pro graphics cards, Offering unique features and leadership performance to tackle heavy to Extreme Professional Workloads - AMD today announced the AMD Radeon PRO W7000 Series graphics, its most-powerful workstation graphics cards to date. The AMD Radeon PRO W7900 and AMD Radeon PRO W7800 graphics cards are built on groundbreaking AMD RDNA 3 architecture, delivering significantly higher performance than the previous generation and exceptional performance-per-dollar compared to the competitive offering. The new graphics cards are designed for professionals to create and work with high-polygon count models seamlessly, deliver incredible image fidelity and color accuracy, and run graphics and compute-based applications concurrently without disruption to workflows.

AMD Radeon PRO W7000 Series graphics cards feature the world's first workstation GPU architecture based on AMD's advanced chiplet design, providing real-world multi-tasking performance and incredible power efficiency. The new graphics cards are also the first professional workstation GPUs to offer the new AMD Radiance Display Engine featuring DisplayPort 2.1 that delivers a superior visual experience, higher resolutions and more available colors than ever before.

Bulk Order of GPUs Points to Twitter Tapping Big Time into AI Potential

According to Business Insider, Twitter has made a substantial investment into hardware upgrades at its North American datacenter operation. The company has purchased somewhere in the region of 10,000 GPUs - destined for the social media giant's two remaining datacenter locations. Insider sources claim that Elon Musk has committed to a large language model (LLM) project, in an effort to rival OpenAI's ChatGPT system. The GPUs will not provide much computational value in the current/normal day-to-day tasks at Twitter - the source reckons that the extra processing power will be utilized for deep learning purposes.

Twitter has not revealed any concrete plans for its relatively new in-house artificial intelligence project but something was afoot when, earlier this year, Musk recruited several research personnel from Alphabet's DeepMind division. It was theorized that he was incubating a resident AI research lab at the time, following personal criticisms levelled at his former colleagues at OpenAI, ergo their very popular and much adopted chatbot.

Intel Discontinues Brand New Max 1350 Data Center GPU, Successor Targets Alternative Markets

Intel has decided to re-organize its Max series of Data Center GPUs (codenamed Ponte Vecchio), as revealed to Tom's Hardware this week, with a particular model - the Data Center Max GPU 1350 set for removal from the lineup. Industry experts are puzzled by this decision, given that the 1350 has been officially "available" on the market since January 2023, following soon after the announcement of the entire Max range in November 2022. Intel has removed listings and entries for the Data Center GPU Max 1350 from its various web presences.

A (sort of) successor is in the works, Intel has lined up the Data Center Max GPU 1450 for release later in the year. This model will have a trimmed I/O bandwidth - this modification is likely targeting companies in China, where performance standards are capped at a certain level (via U.S. sanctions on GPU exports). An Intel spokesperson provided further details and reasons for rearranging the Max product range: "We launched the Intel Data Center Max GPU 1550 (600 W), which was initially targeted for liquid-cooled solutions only. We have since expanded our support by offering Intel Data Center Max GPU 1550 (600 W) to include air-cooled solutions."

Most Popular Graphics Card: NVIDIA GeForce RTX 3060, Steam Hardware Survey

Steam's latest March survey has put NVIDIA's GeForce RTX 3060 at the top, reaching over 10 percent and surpassing both the GTX 1060 and the RTX 2060. NVIDIA has been holding the crown with over 80 percent of users running on their GPUs, while AMD held just over 10 percent. This means that the NVIDIA RTX 3060 almost has more users on Steam than all AMD Radeon graphics cards combined. Intel holds just over 6 percent. Bear in mind that Intel and AMD numbers also include integrated GPUs.

When it comes to CPUs, there are 74.46 percent running on Intel CPUs and 25.54 percent on AMD. Most users use a 6-core CPU, 45.76, with 8-core CPUs taking 18.45 percent. The memory amount has obviously risen, as 56.92 percent run on 16 GB, and 22.41 percent have 32 GB systems. When it comes to OS, most users are running on Windows 10, 73.95 percent, while Windows 11 OS takes 22.41 percent. While some might argue that the Steam Survey is not exactly precise as it is apparently based on a random survey, it does give a general idea and shows the big picture.

NVIDIA H100 AI Performance Receives up to 54% Uplift with Optimizations

On Wednesday, the MLCommons team released the MLPerf 3.0 Inference numbers, and there was an exciting submission from NVIDIA. Reportedly, NVIDIA has used software optimization to improve the already staggering performance of its latest H100 GPU by up to 54%. For reference, NVIDIA's H100 GPU first appeared on MLPerf 2.1 back in September of 2022. In just six months, NVIDIA engineers worked on AI optimizations for the MLPerf 3.0 release to find that basic software optimization can catalyze performance increases anywhere from 7-54%. The workloads for measuring the inferencing speed suite included RNN-T speech recognition, 3D U-Net medical imaging, RetinaNet object detection, ResNet-50 object classification, DLRM recommendation, and BERT 99/99.9% natural language processing.

What is interesting is that NVIDIA's submission is a bit modified. There are open and closed categories that vendors have to compete in, where closed is the mathematical equivalent of a neural network. In contrast, the open category is flexible and allows vendors to submit results based on optimizations for their hardware. The closed submission aims to provide an "apples-to-apples" hardware comparison. Given that NVIDIA opted to use the closed category, performance optimization of other vendors such as Intel and Qualcomm are not accounted for here. Still, it is interesting that optimization can lead to a performance increase of up to 54% in NVIDIA's case with its H100 GPU. Another interesting takeaway is that some comparable hardware, like Qualcomm Cloud AI 100, Intel Xeon Platinum 8480+, and NeuChips's ReccAccel N3000, failed to finish all the workloads. This is shown as "X" on the slides made by NVIDIA, stressing the need for proper ML system software support, which is NVIDIA's strength and an extensive marketing claim.

Intel's Next Generation GPUs to be Made by TSMC, Celestial Set for 3 nm Process

Intel has awarded TSMC with some big contracts for future manufacturing of next generation GPUs, according to Taiwan's Commercial Times. As previously covered on TPU, the second generation Battlemage graphics processing units will get fabricated via a 4 nm process. According to insider sources at both partnering companies, Intel is eyeing a release date in the second half of 2024 for this Xe2-based architecture. The same sources pointed to the third generation Celestial graphics processing units being ready in time for a second half of 2026 launch window. Arc Celestial, which is based on the Xe3 architecture, is set for manufacture in the coming years courtesy of TSMC's N3X (3 nm) process node.

One of the sources claim that Intel is quietly confident about its future prospects in the GPU sector, despite mixed critical and commercial reactions to the first generation line-up of Arc Alchemist discrete graphics cards. The company is said to be anticipating great demand for more potent versions of its graphics products in the future, and internal restructuring efforts have not dulled the will of a core team of engineers. The restructuring process resulted in the original AXG graphics division being divided into two sub-groups - CCG and DCAI. The pioneer of the entire endeavor, Raja Koduri, departed Intel midway through last month, to pursue new opportunities with an AI-focused startup.

Antec Unveils Full-Tower Performance 1 FT Flagship Case with Temperature-Control Display and High Cooling Performance

[Editor's note: Our in-depth review of Antec Performance 1 FT Case is now live]

With outstanding cooling performance and lots of useful features, Antec Inc. presents the latest flagship of full towers, the new Performance 1 FT. Featuring an airflow-enhanced front panel design, magnetic front filter, and four pre-installed PWM fans, this new chassis provides efficient airflow and great cooling performance. The case is now commercially available from MSRP US$159.99.

The new Antec flagship supports the latest RTX 40 Series GPUs. Considering the increasing demands for CPU and GPU cooling, Antec designed the Performance 1 FT to enhance the air intake, improve the cable routing, and enable an easy installation with various options. The new temperature display function allows to keep control of the components. The display screen located on the case top allows users to check the temperature of the GPU and CPU at a glance. It addresses the need for temperature monitoring without purchasing expensive cooling kits.

DirectX 12 API New Feature Set Introduces GPU Upload Heaps, Enables Simultaneous Access to VRAM for CPU and GPU

Microsoft has implemented two new features into its DirectX 12 API - GPU Upload Heaps and Non-Normalized sampling have been added via the latest Agility SDK 1.710.0 preview, and the former looks to be the more intriguing of the pair. The SDK preview is only accessible to developers at the present time, since its official introduction on Friday 31 March. Support has also been initiated via the latest graphics drivers issued by NVIDIA, Intel, and AMD. The Microsoft team has this to say about the preview version of GPU upload heaps feature in DirectX 12: "Historically a GPU's VRAM was inaccessible to the CPU, forcing programs to have to copy large amounts of data to the GPU via the PCI bus. Most modern GPUs have introduced VRAM resizable base address register (BAR) enabling Windows to manage the GPU VRAM in WDDM 2.0 or later."

They continue to describe how the update allows the CPU to gain access to the pool of VRAM on the connected graphics card: "With the VRAM being managed by Windows, D3D now exposes the heap memory access directly to the CPU! This allows both the CPU and GPU to directly access the memory simultaneously, removing the need to copy data from the CPU to the GPU increasing performance in certain scenarios." This GPU optimization could offer many benefits in the context of computer games, since memory requirements continue to grow in line with an increase in visual sophistication and complexity.

NVIDIA Ramps Up Battle Against Makers of Unlicensed GeForce Cards

NVIDIA is stepping up to manufacturers of counterfeit graphics card in China according to an article published by MyDrivers - the hardware giant is partnering up with a number of the nation's major e-commerce companies in order to eliminate inventories of bogus GPUs. It is claimed that these online retail platforms, including JD.com and Douyin, are partway into removing a swathe of dodgy stock from their listings. NVIDIA is seeking to disassociate itself from the pool of unlicensed hardware and the brands responsible for flooding the domestic and foreign markets with so-called fake graphics cards. The company is reputed to be puzzled about the murky origins of this bootlegging of their patented designs.

The market became saturated with fake hardware during the Ethereum mining boom - little known cottage companies such as 51RSIC, Corn, Bingying and JieShuoMllse were pushing rebadged cheap OEM cards to domestic e-tail sites. The knock-off GPUs also crept outside of that sector, and import listings started to appear on international platforms including Ebay, AliExpress, Amazon and Newegg. NVIDIA is also fighting to stop the sale of refurbished cards - these are very likely to have been utilized in intensive cryptocurrency mining activities. A flood of these hit the market following an extreme downturn in crypto mining efforts, and many enthusiast communities have warned against acquiring pre-owned cards due to the high risk of component failure.

HP Launches New Laptops and Accessories for Hybrid Work

Today at the Amplify Partner Conference, HP Inc. announced new products and solutions to usher in the next era of hybrid work for everyone with a comprehensive set of computing solutions for hybrid flexibility. With only 22 percent of workers describing themselves as 'thriving' in hybrid work, it's clear companies are still figuring out how to make hybrid work. "Most companies want to move past the 'forced return' to the office era of hybrid work," said Alex Cho, President of Personal Systems, HP Inc. "The challenge is, they're not sure how to. We believe the future is hybrid flexibility, which delivers the best of the home and the office to workers everywhere."

According to HP's Future of Work study, 80 percent of workers want to be in the office some of the time, but many companies continue to struggle to get workers back in the office. HP research suggests that the most significant barrier to a return to office is a sub-optimal technology experience. In fact, 89 percent say technology is the most important factor driving return to office decisions. Similarly, of those who report thriving in hybrid work, 90 percent believe that access to the right technology and tools leads to a positive work experience. To accelerate employees' return to work, the right technology is required for optimal work setups, enabling success for companies and their employees.

HP Boosts Gaming Solutions for Awe-Inspiring Experiences

Today at the Amplify Partner Conference, HP Inc announced its latest line-up of gaming hardware and software designed to bring gamers everything they need to enjoy the games they love. The new OMEN Transcend 16 Laptop, OMEN 16 Laptop, Victus 16 Laptop, and a vast range of stunning OMEN monitors offer casual, hobbyist, lifestyle, and hardcore gamers the power and flexibility to play and work hard. To bring everything together, new enhancements in OMEN Gaming Hub offer a variety of performance and personalization features.

People everywhere seek devices that can adapt to hybrid play and work. Sixty-two percent of gamers prefer a PC that fits their entire life instead of being used for gaming only. With a billion new gamers entering the space in the last seven years and 84 percent of them using games to connect with others with similar interests, gaming devices enable access to countless games and corresponding communities.

NVIDIA Enables More Encoding Streams on GeForce Consumer GPUs

NVIDIA has quietly removed some video encoding limitations on its consumer GeForce graphics processing units (GPUs), allowing encoding of up to five simultaneous streams. Previously, NVIDIA's consumer GeForce GPUs were limited to three simultaneous NVENC encodes. The same limitation did not apply to professional GPUs.

According to NVIDIA's own Video Encode and Decode GPU Support Matrix document, the number of concurrent NVENC encodes on consumer GPUs have been increased from three to five. This includes certain GeForce GPUs based on Maxwell 2nd Gen, Pascal, Turing, Ampere, and Ada Lovelace GPU architectures. While the number of concurrent NVDEC decodes were never limited, there is a limitation on how many streams you can encode by certain GPU, depending on the resolution of the stream and the codec.

AMD FSR 3 FidelityFX Super Resolution Technology Unveiled at GDC 2023

AMD issued briefing material earlier this month, teasing an upcoming reveal of its next generation FidelityFX at GDC 2023. True to form, today the hardware specialist has announced that FidelityFX Super Resolution 3.0 is incoming. The company is playing catch up with rival NVIDIA, who have already issued version 3.0 of its DLSS graphics enhancer/upscaler for a small number of games. AMD says that FSR 3.0 is in an early stage of development, but it is hoped that its work on temporal upscaling will result in a number of improvements over the previous generation.

The engineering team is aiming for a 2x frame performance improvement over the existing FSR 2.0 technique, which it claims is already capable of: "computing more pixels than we have samples in the current frame." This will be achieved by generating a greater number of pixels in a current frame, via the addition of interpolated frames. It is highly likely that the team will reach a point in development where one sample, at least, will be created for every interpolated pixel. The team wants to prevent feedback loops from occurring - an interpolated frame will only be shown once, and any interpolation artifact would only remain for one frame.

NVIDIA RTX 3080 Ti Owners Reporting Bricked Cards During Diablo IV Closed Beta Play Sessions

A combination of the Diablo IV Closed Beta and NVIDIA RTX 3080 Ti graphics card is proving lethal for the latter - community feedback has alerted Blizzard to take action, and they will be investigating the issue in the coming days, with assistance from NVIDIA. It is speculated that the game is exposing underlying hardware faults within an affected card, but it is odd that a specific model is generating the largest number of issues. Unlucky 3080 Ti owners participating in the closed beta are said to be experiencing unpleasant or inconsistent in-game performance at best, and BSODs followed by non-functional GPUs at worst.

A Blizzard forumite, ForANge, chimed in with their experience: "My graphics card also burned out. While playing, the fans of my Gigabyte GeForce RTX 3080 Ti suddenly started running at maximum speed. At the same time, the signal from the monitor disappeared. After turning off the power and trying to turn it back on, I couldn't get it to work anymore. The card is just under a year old. It happened during a cutscene with flowers when they showed a snowy landscape."

NVIDIA Prepares H800 Adaptation of H100 GPU for the Chinese Market

NVIDIA's H100 accelerator is one of the most powerful solutions for powering AI workloads. And, of course, every company and government wants to use it to power its AI workload. However, in countries like China, shipment of US-made goods is challenging. With export regulations in place, NVIDIA had to get creative and make a specific version of its H100 GPU for the Chinese market, labeled the H800 model. Late last year, NVIDIA also created a China-specific version of the A100 model called A800, with the only difference being the chip-to-chip interconnect bandwidth being dropped from 600 GB/s to 400 GB/s.

This year's H800 SKU also features similar restrictions, and the company appears to have made similar sacrifices for shipping its chips to China. From the 600 GB/s bandwidth of the regular H100 PCIe model, the H800 is gutted to only 300 GB/s of bi-directional chip-to-chip interconnect bandwidth speed. While we have no data if the CUDA or Tensor core count has been adjusted, the sacrifice of bandwidth to comply with export regulations will have consequences. As the communication speed is reduced, training large models will increase the latency and slow the workload compared to the regular H100 chip. This is due to the massive data size that needs to travel from one chip to another. According to Reuters, an NVIDIA spokesperson declined to discuss other differences, stating that "our 800 series products are fully compliant with export control regulations."

Halo Infinite's Latest PC Patch Shifts Minimum GPU Spec Requirements, Below 4 GB of VRAM Insufficient

The latest patch for Halo Infinite has introduced an undesired side effect for a select portion of its PC platform playerbase. Changes to minimum system specification requirements were not clarified by 343 Industries in their patch notes, but it appears that the game now refuses to launch for owners of older GPU hardware. A limit of 4 GB of VRAM has been listed as the bare minimum since Halo Infinite's launch in late 2021, with the AMD Radeon RX 570 and Nvidia GTX GeForce 1050 Ti cards representing the entry level GPU tier, basic versions of both were fitted with 4 GB of VRAM as standard.

Apparently users running the GTX 1060 3 GB model were able to launch and play the game just fine prior to the latest patch, due to it being more powerful than the entry level cards, but now it seems that the advertised hard VRAM limit has finally gone into full effect. The weaker RX 570 and GTX 1050 Ti cards are still capable of running Halo Infinite after the introduction of season 3 content, but a technically superior piece of hardware cannot, which is unfortunate for owners of the GTX 1060 3 GB model who want to play Halo Infinite in its current state.

NVIDIA Preparing RTX 5000 Ada Generation Workstation GPU

In addition to the RTX 4000 SFF Ada Generation workstation GPU launched at the GTC 2023, NVIDIA is apparently also working on the NVIDIA RTX 5000 Ada Generation, which should fit between the previously available AD102-based RTX 6000 Ada Generation workstation graphics card and the new AD104-based RTX 4000 SFF Ada Generation.

According to a fresh report coming from kopite7kimi, the NVIDIA RTX 5000 Ada Generation workstation GPU packs 15,360 CUDA cores and 32 GB of GDDR6 memory. If these specifications are spot on, the RTX 5000 Ada Generation GPU should be also based on the AD102 GPU, with a memory interface cut-down to 256-bit to match the 32 GB of GDDR6 memory. NVIDIA also has enough room to fill the rest of the lineup, but judging from this information, there will be a pretty big gap between the RTX 6000 and RTX 5000 Ada Generation workstation GPUs.

Gigabyte Joins NVIDIA GTC 2023 and Supports New NVIDIA L4 Tensor Core GPU and NVIDIA OVX 3.0

GIGABYTE Technology, an industry leader in high-performance servers and workstations, today announced participation in the global AI conference, NVIDIA GTC, and will share an AI session and other resources to educate attendees. Additionally, with the release of the NVIDIA L4 Tensor Core GPU, GIGABYTE has already begun qualifying its G-series servers to support it with validation. Last, as the NVIDIA OVX architecture has reached a new milestone, GIGABYTE has begun production of purpose-built GIGABYTE servers based on the OVX 3.0 architecture to handle the performance and scale needed for real-time, physically accurate simulations, expansive 3D worlds, and complex digital twins.

NVIDIA Session (S52463) "Protect and Optimize AI Models on Development Platform"
GTC is a great opportunity for researchers and industries to share what they have learned in AI to help further discoveries. This time around, GIGABYTE has a talk by one of MyelinTek's senior engineers that is responsible for the research and development of MLOps technologies. The session demonstrates an AI solution using a pipeline function to quickly retrain new AI models and encrypt them.

Apple A17 Bionic SoC Performance Targets Could be Lowered

Apple's engineering team is rumored to be adjusting performance targets set for its next generation mobile SoC - the A17 Bionic - due to issues at the TSMC foundry. The cutting edge 3 nm process is proving difficult to handle, according to industry tipsters on Twitter. The leaks point to the A17 Bionic's overall performance goals being lowered by 20%, mainly due to the TSMC N3B node not meeting production targets. The factory is apparently lowering its yield and execution targets due to ongoing problems with FinFET limitations.

The leakers have recently revealed more up-to-date A17 Bionic's Geekbench 6 scores, with single thread performance at 3019, and multi-thread at 7860. Various publications have been hyping the mobile SoC's single thread performance as matching that of desktop CPUs from Intel and AMD, more specifically 13th-gen Core i7 and 'high-end' Ryzen models. Naturally the A17 Bionic cannot compete with these CPUs in terms of multi-thread performance.
Return to Keyword Browsing
Jul 4th, 2025 01:15 CDT change timezone

New Forum Posts

Popular Reviews

TPU on YouTube

Controversial News Posts