BrainChip Enters AI Territory with Spiking Neural Network

BrainChip Enters AI Territory with Spiking Neural Network

BrainChip’s new accelerator is based on spiking neural-network technology that promises high performance with low overhead.

It may take a machine to learn about all of the new machine-learning techniques popping up these days. BrainChip is adopting a neuromorphic computing approach that’s different from the other deep neural network (DNN)/convolutional neural network (CNN) solutions already available. BrainChip’s approach uses a technology called spiking neural networks (SNNs) (Fig. 1).

1. Spiking neural networks use asynchronous spikes from neurons to trigger other neurons through reinforced connections.

SNNs differ from DNNs and CNNs (Fig. 2) in training methods and compute requirements. At this point, they’re complementary with some overlap in application areas, but this is likely to change over time. SNNs have some interesting and useful characteristics, including the ability to be trained instantaneously, often called “one-shot learning.”

2. CNNs use a high-overhead feature-learning process—it generates weights used in a classification process that can identify multiple items. SNNs use a more efficient training mechanism, but it’s designed to identify a single entity.

SNNs can be very accurate and have low compute overhead versus DNNs/CNNs for both training and classification. The threshold logic and connection reinforcement of SNNs both exhibit high efficiency, allowing even software implementations to work on lower-end hardware. They also require less hardware to implement accelerators. Figure 3 highlights the differences between the two approaches.

3. Convolutional neural networks tend to be math-intensive with more complex training requirements, while spiking neural networks have less overhead but target more specific classification results.

BrainChip has a software solution called BrainChip Studio that runs on Intel Xeon and Core i7 processors, but its latest offering is the BrainChip accelerator (Fig. 4). The x8 PCI Express board uses Xilinx’s Kintex UltraScale FPGA. Six SNN cores are squeezed into the FPGA, with each core consuming under 1 W while processing at over 100 frames/s. The SSN approach delivers a 7X improvement compared to GPU-accelerated deep-learning classification neural networks like GoogleNet and AlexNet in terms of frames/s per watt. The board uses less than 15 W.

Initially, the accelerator is used with video-surveillance applications that need to identify a particular object. For example, a particular backpack or person may need to be tracked through video streams from dozens of cameras for security purposes. The initial image resolution can be as low as 20 by 20 pixels for patterns and 24 by 24 pixels for facial recognition. This means that the SNN-based search application can be used in low-light, low-resolution, and noisy environments while providing good results. The one-step training means the desired image can be quickly incorporated into the system.

4. The BrainChip Accelerator implements SNNs using Xilinx’s Kintex UltraScale FPGA.

SNN has other applications such as facial recognition, gaming, and visual inspection. Upscaling can help in finding objects; thus, the hardware platform handles upscaling as well as downscaling of up to 16 video streams. The maximum input video stream resolution is 1080p.

Developers will need to determine what kind of neural network will work best for their application. SNNs offer significant advantages if they can be applied to the task, whether using BrainChip’s software or hardware. Applications like augmented reality (AR), where objects of a specific type need to be located within an image, are ripe for SNN support. DNN/CNNs can identify many objects, but oftentimes this isn’t required for an application.

Hide comments

Comments

  • Allowed HTML tags: <em> <strong> <blockquote> <br> <p>

Plain text

  • No HTML tags allowed.
  • Web page addresses and e-mail addresses turn into links automatically.
  • Lines and paragraphs break automatically.
Publish