New AI servers can boost performance into supercomputing territory with a smaller footprint. Credit: dny59 / kentoh / Getty Images Graphcore, the British semiconductor company that develops accelerators for AI and machine learning, has greatly increased the performance of its massively parallel Intelligence Processing Unit (IPU) servers. Graphcore sells its AI-oriented chips in rack-mounted designs called IPU-PODs. Up until now, the maximum per cabinet has been 64 units (in the IPU-POD64). The newest racks are twice and four times as large: the IPU-POD128 and IPU-POD256. With 32 petaFLOPS of AI compute in the IPU-POD 128 and 64 petaFLOPS in the IPU-POD 256, Graphcore says it can now extend its reach into AI supercomputer territory, and with a much smaller footprint than the typical supercomputer, which could fill a basketball court. The IPU-POD systems disaggregate the AI compute from the servers, which means different types of AI workloads requiring different levels of performance can all be run on the same POD. For example, a POD can be allocated for faster training of large Transformer-based language models across an entire system, or the system can be divided into smaller, flexible vPODs to give more developers IPU access. This is done through Graphcore’s Poplar software stack, which includes an SDK, the Graphcore Communication Library (GCL) for managing communication and synchronization between IPUs, and PopRun and PopDist, which allow developers to run their applications across multiple IPU-POD systems. For intra-rack IPU communication, earlier PODs used 64Gb/s IPU-Links. Now, the IPU-POD128 and IPU-POD256 use new Gateway Links, a horizontal, rack-to-rack connection that extends IPU-Links using tunneling over regular 100Gb Ethernet. Both systems have been developed for cloud hyperscalers, national scientific computing labs and enterprise companies with large AI teams in markets like financial services or pharmaceutical. Graphcore’s initial customers include the U.S. Department of Energy’s Lawrence Berkeley National Laboratory and J.P. Morgan. IPU-POD16, IPU-POD64, IPU-POD128 and IPU-POD256 are shipping to customers today from French IT giant ATOS and other systems integrator partners around the world and are available to buy in the cloud from Cirrascale. Related content news HPE, Dell launch another round of AI servers HPE unveils one server, and Dell launches several compute and storage products. By Andy Patrizio Oct 17, 2024 3 mins Servers Data Center news Dell updates servers with AMD AI accelerators The Dell PowerEdge XE9680 ships with Instinct MI300X accelerators, AMD's high-end GPU accelerators that are designed to compete with Nvidia Hopper processors. By Andy Patrizio Jul 17, 2024 3 mins Servers Data Center news Gartner: AI spurs 25% surge in data center systems spending Worldwide IT spending is expected to surpass $5 trillion in 2024, a 7.5% increase over 2023. Gartner's forecast reflects 'ravenous demand' for data center infrastructure. By Denise Dubie Jul 16, 2024 3 mins Generative AI Servers Data Center news Elon Musk’s Grok AI ‘compute factory’ will use Dell and Supermicro servers The factory’s location is still shrouded in mystery and shrink wrap. By John E. Dunn Jun 21, 2024 3 mins Generative AI Servers PODCASTS VIDEOS RESOURCES EVENTS NEWSLETTERS Newsletter Promo Module Test Description for newsletter promo module. Please enter a valid email address Subscribe