Obsessed with technology?
Subscribe to the latest tech news as well as exciting promotions from us and our partners!
By subscribing, you indicate that you have read & understood the SPH's Privacy Policy and PDPA Statement.
News
News Categories

Intel unveils two Nervana AI accelerators for deep learning training and inference

By Wong Chung Wee - on 22 Aug 2019, 1:35pm

Intel unveils two Nervana AI accelerators for deep learning training and inference

Intel has revealed details of its pair of AI accelerators that are based on Nervana NNP architecture. One is designed from ground up for deep learning training within a “given power budget”, while the other is intended for deep learning inference at data centre deployments.

Image source: Intel

The Intel Nervana Neural Network Processor for Training (Intel Nervana NNP-T) AI accelerator, codenamed Spring Crest, features 24 Tensor processing cores and four HBM2 memory die stacks that add up to 32GB. The central chip, with a 688mm square die, is based on TSMC 16nm CLN16FF+ process.

The TPCs and HBM2 stacks, as well as other components like a management CPU, HBM memory controllers, PCIe Gen 4.0 controller, sit on a 1,200mm square passive interposer. The central chip is rated to operate up to 1.1GHz frequency, and power consumption is rated between 150- to 250W.

The Intel Nervana NNP-T is meant to handle current and developing deep learning training workloads. According to Intel, this AI accelerator is designed to reuse on-die data as much as possible, and this explains the memory subsystem of the 24 TPCs. This will reduce the expensive process of fetching data from the HBM2 memory die stacks. The accelerator is air-cooled and comes in OCP Accelerator Module (OAM) and PCIe form factors.

The Intel Nervana Neural Network Processor for Inference (Intel Nervana NNP-I), codenamed Spring Hill, is meant for deep learning inference. It’s offered on an entirely different form factor, i.e., it’s intended for a M.2 interface and sports a 10- 50W power envelope. It features a pair of 10nm Intel Ice Lake CPU cores.

The CPU cores work with 12 Inference Compute Engines (ICE) to handle AI inference workloads. The inference accelerator features 24MB of shared cache, and there are memory controllers for accessing LPDDR4 memory system. It supports both PCIe Gen 4.0 and 3.0 connections.

Image source: Intel

There appears to be another form factor for Intel Nervana NNP-I in the form of a PCIe expansion card. For more information, do visit their respective product presentations, i.e., Intel Nervana NNP-T and Intel Nervana NNP-I. The official Intel announcement for its two AI accelerators can be read in full here.

In 2016, Intel purchased AI startup Nervana to boost its AI portfolio, and the Intel Nervana NNP was first announced in October 2017. Together with its new Xeon Scalable Processors, Intel wants to challenge NVIDIA GPU-based AI foothold.

Source: Intel (1), (2), (3)

Loading...