BindsNET

Built on top of PyTorch, used for simulating SNNs, geared towards ML and reinforcement learning.

BindsNET

Overview

BindsNET is an open-source computational framework designed to simulate spiking neural networks (SNNs). Built atop the PyTorch deep learning library, it was created in 2018 by Hazan Hananel and Daniel Saunders. Their work is supported by a Defense Advanced Research Project Agency Grant they acquired. BindsNET provides tools and functionality for creating, managing and simulating neural networks of spiking neurons and synapses. It utilizes the GPU/CPU acceleration capabilities of PyTorch, fully leveraging the low-powered nature of SNNs. The framework is also accompanied by extensive documentation, including installation guides, a user manual, and detailed reference materials, making it accessible for researchers and practitioners in the field of computational neuroscience and machine learning.

The framework supports a variety of different types of neuron models and learning algorithms. It offers versatility, allowing for specific connections between neuron models and different types of synaptic strengths and connections. This flexibility is invaluable for practitioners and researchers when designing their own network architectures. BindsNET allows for customization of neuron models, enabling users to modify weights, tensor bias, weight maximum value, and a normalization factor for all the weights, which is crucial for synaptic plasticity, learning, and memory. During network creation, you can specify a simulation time-step constant dt, which determines the granularity of the simulation. The time-step parameter induces a trade-off between simulation speed and numerical precision: a larger value results in faster simulation, but reduced accuracy.

While BindsNET opens up many possibilities for SNN research and applications, it may require familiarity with PyTorch and a solid understanding of SNN principles. However, this requirement does not diminish the versatility, customizability, and practical applications of the BindsNET library.

Can you contribute tutorial guides or case studies?

Get Involved with ONM

TrueNorth: A Deep Dive into IBM's Neuromorphic Chip Design

TrueNorth: A Deep Dive into IBM's Neuromorphic Chip Design

  • Fabrizio Ottati

Explore the innovative TrueNorth neuromorphic chip, its event-driven architecture, low power operation, massive parallelism, real-time capabilities, and scalable design.

Spiking Neurons: A Digital Hardware Implementation

Spiking Neurons: A Digital Hardware Implementation

  • Fabrizio Ottati

Learn how to model Leaky Integrate and Fire (LIF) neurons in digital hardware. Understand spike communication, synapse integration, and more for hardware implementation.

Spiking Neural Network (SNN) Library Benchmarks

Spiking Neural Network (SNN) Library Benchmarks

  • Gregor Lenz, Kade Heckel, Sumit Bam Shrestha, Cameron Barker, Jens Egholm Pedersen

Discover the fastest Spiking Neural Network (SNN) frameworks for deep learning-based optimization. Performance, flexibility, and more analyzed in-depth