TheNFAPost Podcast

Intel has unveiled during its Six Five Summit, its vision for the infrastructure processing unit (IPU), a programmable networking device designed to enable cloud and communication service providers to reduce overhead and free up performance for central processing units (CPUs).

With an IPU, customers will better utilise resources with a secure, programmable, stable solution that enables them to balance processing and storage.

Intel Data Platforms Group chief technology officer Guido Appenzeller said the IPU is a new category of technologies and is one of the strategic pillars of our cloud strategy.

“It expands upon our SmartNIC capabilities and is designed to address the complexity and inefficiencies in the modern data center. At Intel, we are dedicated to creating solutions and innovating alongside our customers and partners – the IPU exemplifies this collaboration,” said Guido Appenzeller.

The IPU is a programmable network device that intelligently manages system-level infrastructure resources by securely accelerating those functions in a data center. It allows cloud operators to shift to a fully virtualized storage and network architecture while maintaining high performance and predictability, as well as a high degree of control.

The IPU has dedicated functionality to accelerate modern applications that are built using a microservice-based architecture in the data center. Research from Google and Facebook has shown 22%1 to 80%2 of CPU cycles can be consumed by microservices communication overhead.

With the IPU, a cloud provider can securely manage infrastructure functions while enabling its customer to entirely control the functions of the CPU and system memory.

Accelerate infrastructure functions, including storage virtualisstion, network virtualisation and security with dedicated protocol accelerators.

Free up CPU cores by shifting storage and network virtualisation functions that were previously done in software on the CPU to the IPU. Improve data center utilization by allowing for flexible workload placement. Enable cloud service providers to customize infrastructure function deployments at the speed of software.

Andrew Putnam, principal Hardware Engineering manager at Microsoft, said since before 2015, Microsoft pioneered the use of reconfigurable SmartNICs across multiple Intel server generations to offload and accelerate networking and storage stacks through services like Azure Accelerated Networking.

“The SmartNIC enables us to free up processing cores, scale to much higher bandwidths and storage IOPS, add new capabilities after deployment, and provide predictable performance to our cloud customers. Intel has been our trusted partner since the beginning, and we are pleased to see Intel continue to promote a strong industry vision for the data center of the future with the infrastructure processing unit,” said Andrew Putnam.

Previous articleHOP Electric Mobility Adds Two New e-Scooters To Its Portfolio
Next articlePaytm Launches Vaccine Appointment Booking On app


Please enter your comment!
Please enter your name here