Intel reveals 4th Gen Xeon Scalable processors

Intel has introduced 4th Gen Intel  Xeon Scalable processors (code-named Sapphire Rapids), the Intel Xeon CPU Max Series (code-named Sapphire Rapids HBM) and the Intel Data Center GPU Max Series (code-named Ponte Vecchio).
4th Gen Intel Xeon Scalable processors Sapphire Rapids
“The launch of 4th Gen Xeon Scalable processors and the Max Series product family is a pivotal moment in fueling Intel’s turnaround, reigniting our path to leadership in the data center and growing our footprint in new arenas,” said Sandra Rivera, Intel executive vice president and general manager of the Data Center and AI Group.

Today, there are over 100 million Xeons installed in the market – from on-prem servers running IT services, including new as-a-service business models, to networking equipment managing Internet traffic, to wireless base station computing at the edge, to cloud services.

4th Gen Xeon processors deliver leading performance with the most built-in accelerators of any CPU in the world to tackle customers’ most important computing challenges across AI, analytics, networking, security, storage and HPC.

4th Gen Intel Xeon customers can expect a 2.9x average performance per watt efficiency improvement for targeted workloads when utilizing built-in accelerators, up to 70-watt power savings per CPU in optimized power mode with minimal performance loss, and a 52 percent to 66 percent lower TCO.
Features of 4th Gen Intel Xeon Scalable processors
The new Optimized Power Mode can deliver up to 20 percent socket power savings with a less than 5 percent performance impact for selected workloads.

4th Gen Xeon processors achieve up to 10x higher PyTorch real-time inference and training performance with built-in Intel Advanced Matrix Extension (Intel AMX) accelerators.

The Xeon CPU Max Series expands on these capabilities for natural language processing, with customers seeing up to 20x speed-up on large language models.

Xeon CPU Max Series is the only x86-based processor with high bandwidth memory, accelerating many HPC workloads without the need for code changes.

Xeon CPU Max Series offers 64 gigabytes of high bandwidth memory (HBM2e) on the package, significantly increasing data throughput for HPC and AI workloads. Xeon CPU Max Series provides up to 3.7 times more performance on a range of real-world applications like energy and earth systems modeling.

Data Center GPU Max Series packs over 100 billion transistors into a 47-tile package, bringing new levels of throughput to challenging workloads like physics, financial services and life sciences. When paired with the Xeon CPU Max Series, the combined platform achieves up to 12.8 times greater performance than the prior generation when running the LAMMPS molecular dynamics simulator.

Intel’s new virtual-machine (VM) isolation technology, Intel Trust Domain Extensions (Intel TDX), will debut with Microsoft Azure, Alibaba Cloud, Google Cloud and IBM Cloud.