Speck - SynSense

Learn about SynSense's neuromorphic hardware: Speck

Speck At A Glance

Release Year: 2022
Status: Released
Chip Type: Digital, with analog sensor
Software: Sinabs, Samna
Applications: Smart sensing
Neurons: 328K
Synapses: >6.16G CNN, ~278K fully-connected
Weight bits: 8bit ints
Activation bits: 16bit (neuron)
Power: ~5 mW

Speck is a fully event-driven neuromorphic vision SoC. Speck is able to support large-scale spiking convolutional neural network (sCNN) with a fully asynchronous chip architecture. Speck is fully configurable with the spiking neuron capacity of 328K. Furthermore, it integrates the state-of-art dynamic vision sensor (DVS) that enables fully event-driven based, real-time, highly integrated solution for varies dynamic visual scene. For classical applications, Speck can provide intelligence upon the scene at only mWs with a latency of 3.36us for a single event processed by a 9 layer network.

The Speck is an ultra-low power event-driven neuromorphic processor combined with an event-driven vision sensor on a single chip. It was developed by SynSense AG, a neuromorphic engineering startup based in Zurich, Switzerland.

Overview

The Speck chip contains 328K spiking integrate-and-fire neurons and is fully configurable for implementing different SCNN architectures. It utilizes in-memory computing techniques to perform sparse, event-driven neural network computations, enabling extremely low power consumption in the sub-milliwatt range.

It comprises 9 layers of convolution combined with pooling, that can be freely connected to each other. The layers can additionally be operated as dense fully connected layers for last layer classification. It has additional input preprocessing and decision readout blocks to make it out-of-the-box integratable in diverse environments and use cases.

The chip has a dynamic vision sensor (DVS). This allows it to receive input spike streams directly without ever leaving the chip, reducing latency. It can support various types of convolutional network layers like convolutional, ReLU, pooling, etc., as well as popular network models like LeNet, ResNet, and Inception.

Development

The Speck was developed by SynSense AG, a neuromorphic AI startup based in Zurich, Switzerland. The chip’s architecture and circuits were designed to optimize performance, power, and area specifically for ultra-low power SCNN inferencing rather than general-purpose computing.

The hardware interfaces with a software framework called SINABS (https://github.com/synsense/sinabs ) developed by SynSense for converting deep learning models from frameworks like Keras and PyTorch into equivalent SCNNs. It also integrates with the Samna (https://pypi.org/project/samna/ ) middleware which handles interfacing the chip to sensors and visualization.

Applications

The ultra-low latency and power consumption of the Speck make it suitable for embedded and edge applications like:

  • Computer vision
  • Robotics
  • Internet-of-Things devices
  • Autonomous vehicles
  • Drones and other mobile platforms

A face recognition application for DVS cameras was demonstrated running on the Speck at extremely low average power (<1mW). Such always-on vision applications are ideally suited for the event-driven capabilities of the chip.

DateTitleAuthorsVenue/Source
April 2022Speck: A Smart event-based Vision Sensor with a low latency 327K Neuron Convolutional Neuronal Network Processing PipelineOle Richter, Yannan Xing, Michele De Marchi, Carsten Nielsen, Merkourios Katsimpris, Roberto Cattaneo, Yudi Ren, Qian Liu, Sadique Sheik, Tugba Demirci, Ning Qiaoarxiv (temporarily unavailable)
June 2019Live Demonstration: Face Recognition on an Ultra-Low Power Event-Driven Convolutional Neural Network ASICQian Liu, Ole Richter, Carsten Nielsen, Sadique Sheik, Giacomo Indiveri, Ning QiaoCVPR 2019
~ Share this Site ~
Efficient Compression for Event-Based Data in Neuromorphic Applications

Efficient Compression for Event-Based Data in Neuromorphic Applications

  • Gregor Lenz, Fabrizio Ottati, Alexandre Marcireau

Discover methods to efficiently encode and store event-based data from high-resolution event cameras, striking a balance between file size and fast retrieval for spiking neural network training.

TrueNorth: A Deep Dive into IBM's Neuromorphic Chip Design

TrueNorth: A Deep Dive into IBM's Neuromorphic Chip Design

  • Fabrizio Ottati

Explore the innovative TrueNorth neuromorphic chip, its event-driven architecture, low power operation, massive parallelism, real-time capabilities, and scalable design.

NorthPole, IBM's latest Neuromorphic AI Hardware

NorthPole, IBM's latest Neuromorphic AI Hardware

  • Fabrizio Ottati

Translating the NorthPole paper from IBM to human language.