News Posts matching #IBM

Return to Keyword Browsing

IBM to Replace 7,800 Jobs Using Automation and AI

With the rise of generative artificial intelligence (AI) and other AI tools, the new models pose a threat to workers around the globe. Today, as reported by Bloomberg News, we find out that International Business Machines, or IBM shortly, is planning to replace thousands of jobs using automation and AI technology. As the report suggests, IBM's Chief Executive Officer Arvind Krishna is expecting to pause the hiring for roles AI could replace in the coming years. The main department that will see a job cut is the non-customer-facing roles like the back office. There are 26,000 people at IBM working in the back office, and CEO noted that "I could easily see 30% of that getting replaced by AI and automation over a five-year period."

This translates to roughly 7,800 jobs impacted by AI over the next five years. The CEO also added that departments such as human resources would see a hiring slowdown or even suspension of hiring further. IBM has 260,000 employees, and the company continues to look for engineering and customer-facing roles.

Rapidus to Start Production of 2 nm Fab in Chitose, Gets Cash Injection from Japanese Government

Future Japanese chipmaker Rapidus has announced that their first fab will be located in Chitose, Hokkaido, located in northern Japan. The planned 2 nm fab will be one of the most advanced fabs in the world once it's ready and construction is said to be starting in September, thanks to approval by the related Japanese government agencies. So far, the Japanese government has approved 330 billion yen for Rapidus, with the most recent investment being 260 billion yen or the equivalent of US$1.94 billion.

However, the total investment into the 2 nm fab is expected to end up somewhere around 5 trillion yen (~US$37.5 billion) in total investments before the fab is ready for mass production. Rapidus is collaborating with IBM and has already sent a group of researchers to its Albany Nanotech facility in upstate New York, which is one of the world's most advanced semiconductor research facilities. At the same time, Japan is working on building a local talent pool of researchers and semiconductor plant workers, by spearheading specialised training for select university students from Japan's top universities. Time will tell if this gamble pays off for Japan, as it's going to be a huge investment before the new fab stands ready in early 2025.

GlobalFoundries Files Lawsuit Against IBM to Protect its Intellectual Property and Trade Secrets

GlobalFoundries (GF) today sued IBM for trade secret misappropriation. The complaint asserts the former semiconductor manufacturing company has unlawfully disclosed GF's confidential IP and trade secrets, after IBM sold its microelectronics business to GF in 2015. The technology at issue was collaboratively developed, over decades, by the companies in Albany, New York and the sole and exclusive right to license and disclose that technology was transferred to GF upon the sale.

In the legal action filed in federal court in the Southern District of New York, GF asserts that IBM unlawfully disclosed GF IP and trade secrets to IBM partners including Intel and Japan's Rapidus, a newly formed advanced logic foundry, and by doing so, IBM is unjustly receiving potentially hundreds of millions of dollars in licensing income and other benefits.

IBM z16 and LinuxONE 4 Get Single Frame and Rack Mount Options

IBM today unveiled new single frame and rack mount configurations of IBM z16 and IBM LinuxONE 4, expanding their capabilities to a broader range of data center environments. Based on IBM's Telum processor, the new options are designed with sustainability in mind for highly efficient data centers, helping clients adapt to a digitized economy and ongoing global uncertainty.

Introduced in April 2022, the IBM z16 multi frame has helped transform industries with real-time AI inferencing at scale and quantum-safe cryptography. IBM LinuxONE Emperor 4, launched in September 2022, features capabilities that can reduce both energy consumption and data center floor space while delivering the scale, performance and security that clients need. The new single frame and rack mount configurations expand client infrastructure choices and help bring these benefits to data center environments where space, sustainability and standardization are paramount.

ORNL's Exaflop Machine Frontier Keeps Top Spot, New Competitor Leonardo Breaks the Top10 List

The 60th edition of the TOP500 reveals that the Frontier system is still the only true exascale machine on the list.

With an HPL score of 1.102 EFlop/s, the Frontier machine at Oak Ridge National Laboratory (ORNL) did not improve upon the score it reached on the June 2022 list. That said, Frontier's near-tripling of the HPL score received by second-place winner is still a major victory for computer science. On top of that, Frontier demonstrated a score of 7.94 EFlop/s on the HPL-MxP benchmark, which measures performance for mixed-precision calculation. Frontier is based on the HPE Cray EX235a architecture and it relies on AMD EPYC 64C 2 GHz processor. The system has 8,730,112 cores and a power efficiency rating of 52.23 gigaflops/watt. It also relies on gigabit ethernet for data transfer.

IBM Osprey Processor Brings 433 Qubits to Power Modular Quantum Supercomputers

IBM is one of the frontiers for using the natural properties of quantum particles to process the information on an enterprise scale. With constant advances in quantum information processing, the company is using newly found discoveries to double the size of its quantum processors. Using quantum properties instead of the conventional on/off switching of bits in the regular processors, quantum processors can process the information on a much larger scale. Last year, IBM unveiled the Eagle quantum processor with 127 qubits. This year, the company is bringing in 433 qubits to the table to power the next generation of enterprise and data center infrastructure.

Called IBM Osprey, it features IBM's 433 qubits cooled to cryogenic temperatures and in a controlled environment. While the computational power of the processor seems to be rather impressive, it is still a noisy quantum implementation that is sensitive to outside noise and requires exceptionally low temperatures to operate, such as -273 Degrees Celcius. To combat some of those obstacles, Osprey adds multi-level wiring to provide flexibility for signal routing and device layout while also adding integrated filtering to reduce noise and improve stability. Concurrently, IBM developed new signal delivery wiring that is 70% cheaper and produces the same result, driving up the ability to commercialize this design. For performance, IBM managed to increase quantum volume four times from 128 to 512 and a 10x improvement in Driving quantum performance from 1.4k to 15k Circuit Layer Operations Per Second (CLOPS).

IBM Artificial Intelligence Unit (AIU) Arrives with 23 Billion Transistors

IBM Research has published information about the company's latest development of processors for accelerating Artificial Intelligence (AI). The latest IBM processor, called the Artificial Intelligence Unit (AIU), embraces the problem of creating an enterprise solution for AI deployment that fits in a PCIe slot. The IBM AIU is a half-height PCIe card with a processor powered by 23 Billion transistors manufactured on a 5 nm node (assuming TSMC's). While IBM has not provided many details initially, we know that the AIU uses an AI processor found in the Telum chip, a core of the IBM Z16 mainframe. The AIU uses Telum's AI engine and scales it up to 32 cores and achieve high efficiency.

The company has highlighted two main paths for enterprise AI adoption. The first one is to embrace lower precision and use approximate computing to drop from 32-bit formats to some odd-bit structures that hold a quarter as much precision and still deliver similar result. The other one is, as IBM touts, that "AI chip should be laid out to streamline AI workflows. Because most AI calculations involve matrix and vector multiplication, our chip architecture features a simpler layout than a multi-purpose CPU. The IBM AIU has also been designed to send data directly from one compute engine to the next, creating enormous energy savings."

Announcing IBM z16: Real-time AI for Transaction Processing at Scale and Industry's First Quantum-Safe System

IBM today unveiled IBM z16, IBM's next-generation system with an integrated on-chip AI accelerator—delivering latency-optimized inferencing. This innovation is designed to enable clients to analyze real-time transactions, at scale -- for mission-critical workloads such as credit card, healthcare and financial transactions. Building on IBM's history of security leadership, IBM z16 also is specifically designed to help protect against near-future threats that might be used to crack today's encryption technologies.

IBM innovations, including the IBM z16, have formed the technology backbone of the global economy for decades. Today's modern IBM mainframe is central to hybrid cloud environments, valued by two thirds of the Fortune 100, 45 of the world's top 50 banks, 8 of the top 10 insurers, 7 of the top 10 global retailers and 8 out of the top 10 telcos as a highly secured platform for running their most mission critical workloads. For example, according to a recent IBM commissioned study by Celent "Operationalizing Fraud Prevention on IBM Z," IBM zSystems run 70% of global transactions, on a value basis. "IBM is the gold standard for highly secured transaction processing. Now with IBM z16 innovations, our clients can increase decision velocity with inferencing right where their mission critical data lives," said Ric Lewis, SVP, IBM Systems. "This opens up tremendous opportunities to change the game in their respective industries so they will be positioned to deliver better customer experiences and more powerful business outcomes.

IBM Announces the Cyber Vault Flash Storage System Resilient to Ransomware and other Cyber Attacks

IBM today unveiled IBM FlashSystem Cyber Vault to help companies better detect and recover quickly from ransomware and other cyberattacks. The company also announced new FlashSystem storage models, based on IBM Spectrum Virtualize to provide a single and consistent operating environment, that are designed to increase cyber resilience and application performance within a hybrid cloud environment.

According to the IBM Cyber Resilient Organization study, 46 percent of respondents surveyed reported experiencing a ransomware attack over the past two years. With cyberattacks continuing to grow, and with average recovery time lasting days or even weeks, business and reputational risks are unprecedented. Even with prevention and detection tactics in place, organizations also must be ready to recover their operations quickly to minimize loss of business and other costs.

IBM Welcomes LG Electronics to the IBM Quantum Network to Advance Industry Applications of Quantum Computing

IBM today announced that LG Electronics has joined the IBM Quantum Network to advance the industry applications of quantum computing. By joining the IBM Quantum Network, IBM will provide LG Electronics access to IBM's quantum computing systems, as well as to IBM's quantum expertise and Qiskit, IBM's open-source quantum information software development kit.

LG Electronics aims to explore applications of quantum computing in industry to support big data, artificial intelligence, connected cars, digital transformation, IoT, and robotics applications - all of which require processing a large amount of data. With IBM Quantum, LG can leverage quantum computing hardware and software advances and applications as they emerge, in accordance with IBM's quantum roadmap. By leveraging IBM Quantum technology, LG will provide workforce training to its employees, permitting LG to investigate how potential breakthroughs can be applied to its industry.

IBM and Samsung Unveil Semiconductor Breakthrough That Defies Conventional Design

Today, IBM and Samsung Electronics jointly announced a breakthrough in semiconductor design utilizing a new vertical transistor architecture that demonstrates a path to scaling beyond nanosheet, and has the potential to reduce energy usage by 85 percent compared to a scaled fin field-effect transistor (finFET)1. The global semiconductor shortage has highlighted the critical role of investment in chip research and development and the importance of chips in everything from computing, to appliances, to communication devices, transportation systems, and critical infrastructure.

The two companies' semiconductor innovation was produced at the Albany Nanotech Complex in Albany, NY, where research scientists work in close collaboration with public and private sector partners to push the boundaries of logic scaling and semiconductor capabilities. This collaborative approach to innovation makes the Albany Nanotech Complex a world-leading ecosystem for semiconductor research and creates a strong innovation pipeline, helping to address manufacturing demands and accelerate the growth of the global chip industry.

IBM Unveils Breakthrough 127-Qubit Quantum Processor

IBM today announced its new 127-quantum bit (qubit) 'Eagle' processor at the IBM Quantum Summit 2021, its annual event to showcase milestones in quantum hardware, software, and the growth of the quantum ecosystem. The 'Eagle' processor is a breakthrough in tapping into the massive computing potential of devices based on quantum physics. It heralds the point in hardware development where quantum circuits cannot be reliably simulated exactly on a classical computer. IBM also previewed plans for IBM Quantum System Two, the next generation of quantum systems.

Quantum computing taps into the fundamental quantum nature of matter at subatomic levels to offer the possibility of vastly increased computing power. The fundamental computational unit of quantum computing is the quantum circuit, an arrangement of qubits into quantum gates and measurements. The more qubits a quantum processor possesses, the more complex and valuable the quantum circuits that it can run.

TOP500 Update Shows No Exascale Yet, Japanese Fugaku Supercomputer Still at the Top

The 58th annual edition of the TOP500 saw little change in the Top10. The Microsoft Azure system called Voyager-EUS2 was the only machine to shake up the top spots, claiming No. 10. Based on an AMD EPYC processor with 48 cores and 2.45GHz working together with an NVIDIA A100 GPU and 80 GB of memory, Voyager-EUS2 also utilizes a Mellanox HDR Infiniband for data transfer.

While there were no other changes to the positions of the systems in the Top10, Perlmutter at NERSC improved its performance to 70.9 Pflop/s. Housed at the Lawrence Berkeley National Laboratory, Perlmutter's increased performance couldn't move it from its previously held No. 5 spot.

IBM Cloud Selects 3rd Gen AMD EPYC Processors for New Bare Metal Offering for Compute-Intensive Workloads

AMD announced today that IBM Cloud has chosen 3rd Gen AMD EPYC processors to expand its bare metal service offerings designed to power customers' demanding workloads and solutions. The new servers, featuring 128 cores, up to 4 TB of memory and 10 NVMe drives per server, give users full access to high-end, dual-socket performance with AMD EPYC 7763 processors; a first for IBM Cloud in a dual-socket platform.

"Our customers have a high demand for computing processing power and the new 3rd Gen AMD EPYC processors provide the high levels of performance and scalability we were looking for," said Suresh Gopalakrishnan, vice president, IBM Cloud. "Our collaboration with AMD has helped us deliver our highest core counts and bandwidth ever available for IBM Cloud customers, to offer top market performance for today and tomorrow's demanding workloads."

"IBM Cloud customers regularly running compute-intensive workloads can see an immediate benefit to speed and scalability by upgrading to 3rd Gen AMD EPYC processors, while helping to deliver a secure experience for end-users," said Lynn Comp, corporate vice president, Cloud Business Group, AMD. "Our continued collaboration with IBM Cloud is further validation of the strong standing AMD holds in the market as we deliver topline solutions that promote a seamless experience for cloud partners and their customers."

NVIDIA Quantum-2 Takes Supercomputing to New Heights, Into the Cloud

NVIDIA today announced NVIDIA Quantum-2, the next generation of its InfiniBand networking platform, which offers the extreme performance, broad accessibility and strong security needed by cloud computing providers and supercomputing centers.

The most advanced end-to-end networking platform ever built, NVIDIA Quantum-2 is a 400 Gbps InfiniBand networking platform that consists of the NVIDIA Quantum-2 switch, the ConnectX-7 network adapter, the BlueField-3 data processing unit (DPU) and all the software that supports the new architecture.

IBM Unveils New Generation of IBM Power Servers for Frictionless, Scalable Hybrid Cloud

IBM (NYSE: IBM) today announced the new IBM Power E1080 server, the first in a new family of servers based on the new IBM Power10 processor, designed specifically for hybrid cloud environments. The IBM Power10-equipped E1080 server is engineered to be one of the most secured server platforms and is designed to help clients operate a secured, frictionless hybrid cloud experience across their entire IT infrastructure.

The IBM Power E1080 server is launching at a critical time for IT. As organizations around the world continue to adapt to unpredictable changes in consumer behaviors and needs, they need a platform that can deliver their applications and insights securely where and when they need them. The IBM Institute of Business Value's 2021 CEO Study found that, of the 3,000 CEOs surveyed, 56% emphasized the need to enhance operational agility and flexibility when asked what they'll most aggressively pursue over the next two to three years.

IBM Unveils On-Chip Accelerated Artificial Intelligence Processor

At the annual Hot Chips conference, IBM (NYSE: IBM) today unveiled details of the upcoming new IBM Telum Processor, designed to bring deep learning inference to enterprise workloads to help address fraud in real-time. Telum is IBM's first processor that contains on-chip acceleration for AI inferencing while a transaction is taking place. Three years in development, the breakthrough of this new on-chip hardware acceleration is designed to help customers achieve business insights at scale across banking, finance, trading, insurance applications and customer interactions. A Telum-based system is planned for the first half of 2022.

Today, businesses typically apply detection techniques to catch fraud after it occurs, a process that can be time consuming and compute-intensive due to the limitations of today's technology, particularly when fraud analysis and detection is conducted far away from mission critical transactions and data. Due to latency requirements, complex fraud detection often cannot be completed in real-time - meaning a bad actor could have already successfully purchased goods with a stolen credit card before the retailer is aware fraud has taken place.

Intel Wins US Government Project to Develop Leading-Edge Foundry Ecosystem

The U.S. Department of Defense, through the NSTXL consortium-based S2MARTS OTA, has awarded Intel an agreement to provide commercial foundry services in the first phase of its multi-phase Rapid Assured Microelectronics Prototypes - Commercial (RAMP-C) program. The RAMP-C program was created to facilitate the use of a U.S.-based commercial semiconductor foundry ecosystem to fabricate the assured leading-edge custom and integrated circuits and commercial products required for critical Department of Defense systems. Intel Foundry Services, Intel's dedicated foundry business launched this year, will lead the work.

"One of the most profound lessons of the past year is the strategic importance of semiconductors, and the value to the United States of having a strong domestic semiconductor industry. Intel is the sole American company both designing and manufacturing logic semiconductors at the leading edge of technology. When we launched Intel Foundry Services earlier this year, we were excited to have the opportunity to make our capabilities available to a wider range of partners, including in the U.S. government, and it is great to see that potential being fulfilled through programs like RAMP-C." -Pat Gelsinger, Intel CEO.

IDC Forecasts Companies to Spend Almost $342 Billion on AI Solutions in 2021

Worldwide revenues for the artificial intelligence (AI) market, including software, hardware, and services, is estimated to grow 15.2% year over year in 2021 to $341.8 billion, according to the latest release of the International Data Corporation (IDC) Worldwide Semiannual Artificial Intelligence Tracker. The market is forecast to accelerate further in 2022 with 18.8% growth and remain on track to break the $500 billion mark by 2024. Among the three technology categories, AI Software occupied 88% of the overall AI market. However, in terms of growth, AI Hardware is estimated to grow the fastest in the next several years. From 2023 onwards, AI Services is forecast to become the fastest growing category.

Within the AI Software category, AI Applications has the lion's share at nearly 50% of revenues. In terms of growth, AI Platforms is the strongest with a five-year compound annual growth rate (CAGR) of 33.2%. The slowest will be AI System Infrastructure Software with a five-year CAGR of 14.4% while accounting for roughly 35% of all AI Software revenues. Within the AI Applications market, AI ERM is expected to grow slightly stronger than AI CRM over the next five years. Meanwhile, AI Lifecycle Software is forecast to grow the fastest among the markets within AI Platforms.

NVIDIA and Global Partners Launch New HGX A100 Systems to Accelerate Industrial AI and HPC

NVIDIA today announced it is turbocharging the NVIDIA HGX AI supercomputing platform with new technologies that fuse AI with high performance computing, making supercomputing more useful to a growing number of industries.

To accelerate the new era of industrial AI and HPC, NVIDIA has added three key technologies to its HGX platform: the NVIDIA A100 80 GB PCIe GPU, NVIDIA NDR 400G InfiniBand networking, and NVIDIA Magnum IO GPUDirect Storage software. Together, they provide the extreme performance to enable industrial HPC innovation.

COLORFUL Launches the First GPU History Museum

Colorful Technology Company Limited, a professional manufacturer of graphics cards, motherboards, all-in-one gaming and multimedia solutions, and high-performance storage, announces the launch of the GPU History Museum in partnership with NVIDIA. COLORFUL has recently relocated to Shenzhen New Generation Industrial Park. With that, COLORFUL is proud to announce the launch of the first GPU History Museum in China. The museum will showcase the beginnings of the Graphics Processing Unit (GPU), to the development and evolution of graphics cards up to the present generation.

Intel to Detail "Alder Lake" and "Sapphire Rapids" Microarchitectures at Hot Chips 33, This August

Intel will detail its 12th Gen Core "Alder Lake" client and "Sapphire Rapids" server CPU microarchitectures at the Hot Chips 33 conclave, this August. In fact, Intel's presentation leads the CPU sessions on the opening day of August 23. "Alder Lake" will be the session opener, followed by AMD's presentation of the already-launched "Zen 3," and IBM's 5 GHz Z processor powering its next-gen mainframes. A talk on Xeon "Sapphire Rapids" follows this. Hot Chips is predominantly an engineering conclave, where highly technical sessions are presented by engineers from major semiconductor firms; and so the sessions on "Alder Lake" and "Sapphire Rapids" are expected to be very juicy.

"Alder Lake" is Intel's attempt at changing the PC ecosystem by introducing hybrid CPU cores, a concept introduced to the x86 machine architecture with "Lakefield." The processor will also support next-generation I/O, such as DDR5 memory. The "Sapphire Rapids" server CPU microarchitecture will see an increase in CPU core counts, next-gen I/O such as PCI-Express 5.0, CXL 1.1, DDR5 memory, and more.

IBM Announces World's First 2nm Chip Technology

IBM today unveiled a breakthrough in semiconductor design and process with the development of the world's first chip announced with 2 nanometer (nm) nanosheet technology. Semiconductors play critical roles in everything from computing, to appliances, to communication devices, transportation systems, and critical infrastructure.

Demand for increased chip performance and energy efficiency continues to rise, especially in the era of hybrid cloud, AI, and the Internet of Things. IBM's new 2 nm chip technology helps advance the state-of-the-art in the semiconductor industry, addressing this growing demand. It is projected to achieve 45 percent higher performance, or 75 percent lower energy use, than today's most advanced 7 nm node chips.

Commodore 64 Modded To Mine Bitcoin

We saw the modified Nintendo Game Boy last month which could crank out Bitcoins at a blistering 0.8 hashes per second or ~125 trillion times slower than a modern Bitcoin ASIC miner. If you are searching for something a bit more modest than the Game Boy take a look at the Commodore 64 which has been modded to achieve a Bitcoin mining rate of 0.3 hashes per second. The Commodore 64 was released by IBM in 1982 featuring the MOS Technology 6510 processor clocked at 1.023 MHz and paired with 64 KB RAM and 20 KB ROM.

While the Commodore currently falls behind the Game Boy there is hope on the horizon with the creator of the program claiming a 10x performance improvement to over 3 hashes per second is possible by re-writing the code in machine language. The commodore 64 can be further upgraded with the SuperCPU upgrade which boosts mining speeds to over 60 hashes per second completely destroying the Game Boy but still falling just short of the latest ASIC miners at ~18,000,000,000,000 hashes per second. Obviously, this demonstration was not meant as a practical application but it is interesting to see how cryptocurrency mining can be implemented on older hardware and the amazing rate of technological advancement we have seen over the last 40 years.

IBM and Fujifilm Set a Record: 580 TB Data Capacity in a Single Cartridge

Magnetic tape storage is one of the oldest technologies used for storing data. The technology was invented way back in 1928, and it is almost 100 years old. By today's standards, the technology is considered to be slow, however, it offers something that no modern HDD or SDD offers. Today, in collaboration with Fujifilm, IBM has developed a Strontium Ferrite (SrFe) magnetic layer for LTO-8 tapes that are capable of storing an amazing 580 TeraBytes of data in a single cartridge. The new technology will enable the modern world to store ever-increasing data sizes we are now counting in zettabytes. To store all of that data, one would need a high-capacity storage device to store all of the "cold data" that doesn't need real-time processing and has information of value.

That is exactly why IBM and Fujifilm have been developing the LTO-8 tape drives that are capable of 580 TB of capacity in a single cartridge. The technology can achieve that capacity thanks to the Strontium Ferrite (SrFe), which is capable of 317Gb/in2 recording density. With 1255 meters of the tape, IBM and Fujifilm have been able to achieve this density metric.
Return to Keyword Browsing
May 21st, 2024 13:18 EDT change timezone

New Forum Posts

Popular Reviews

Controversial News Posts