AI Processors Dive Into Deep Learning At The Edge

CEVA’s NeuPro artificial intelligence (AI) processor family is designed for deep learning inference at the edge and targets smart and connected-edge device vendors looking for a streamlined way to quickly take advantage of the significant possibilities that deep neural network technologies offer. This dedicated AI processors offers a considerable step-up in performance, ranging from 2 Tera Ops Per Second (TOPS) for the entry-level processor to 12.5 TOPS for the most advanced configuration.

 

The NeuPro processor line extends the use of AI beyond machine vision to edge-based applications including natural language processing, real-time translation, authentication, workflow management, and many other learning-based applications that make devices smarter and reduce human involvement. The architecture is composed of a combination of hardware-based and software-based engines coupled for a complete, scalable and expandable AI solution. Optimizations for power, performance, and area (PPA) are achieved using a precise mix of hardware, software and configurable performance options for each application tier.

 

The NeuPro family:

  • NP500 is the smallest processor, including 512 MAC units and targeting IoT, wearables and cameras
  • NP1000 includes 1024 MAC units and targets mid-range smartphones, ADAS, industrial applications and AR/VR headsets
  • NP2000 includes 2048 MAC units and targets high-end smartphones, surveillance, robots and drones
  • NP4000 includes 4096 MAC units for high-performance edge processing in enterprise surveillance and autonomous driving

Each processor consists of the NeuPro engine and the NeuPro VPU. The NeuPro engine includes the hardwired implementation of neural network layers among which are convolutional, fully-connected, pooling, and activation. The NeuPro VPU is a cost-efficient programmable vector DSP, which handles the CDNN software and provides software-based support for new advances in AI workloads. NeuPro supports both 8-bit and 16-bit neural networks, with an optimized decision made in real time to deliver the best tradeoff between precision and performance. The MAC units achieve better than 90% utilization when running, ensuring highly optimized neural network performance. The overall processor design reduces DDR bandwidth substantially, improving power consumption levels for any AI application.

 

The NeuPro family, coupled with CDNN, CEVA’s neural network software framework, provides a deep learning solution for developers to generate and port their proprietary neural networks to the processor. CDNN supports the full gamut of layer types and network topologies.

 

In conjunction with the NeuPro processor line, CEVA will also offer the NeuPro hardware engine as a Convolutional Neural Network (CNN) accelerator. When combined with the CEVA-XM4 or CEVA-XM6 vision platforms, this provides an option for customers seeking a single unified platform for imaging, computer vision and neural network workloads. NeuPro will be available for licensing to select customers in the second quarter of 2018 and for general licensing in the third quarter of 2018. More information is available on the NeuPro product page.