C-DNN and C-Transformer: Mixing ANNs and SNNs for the Best of Both Worlds

Join us for a talk by Sangyeob Kim, Postdoctoral researcher at KAIST, on designing efficient accelerators that mix SNNs and ANNs.

Sangyeob and his team have developed a C-DNN processor that effectively processes object recognition workloads, achieving 51.3% higher energy efficiency compared to the previous state-of-the-art processor. Subsequently, they have applied C-DNN not only to image classification but also to other applications, and have developed the C-Transformer, which applies this technique to a Large Language Model (LLM). As a result, they demonstrate that the energy consumed in LLM can be reduced by 30% to 72% using the C-DNN technique, compared to the previous state-of-the-art processor. In this talk, we will introduce the processor developed for C-DNN and C-Transformer, and discuss how neuromorphic computing can be used in actual applications in the future.

Social share preview for C-DNN and C-Transformer: mixing ANNs and SNNs for the best of both worlds

Upcoming Workshops

No workshops are currently scheduled. Check back soon for new events!

Are you an expert in a neuromorphic topic? We invite you to share your knowledge with our community. Hosting a workshop is a great way to engage with peers and share your work.

About the Speakers

Sangyeob Kim

Sangyeob Kim

Post-Doctoral Associate at KAIST, PhD in Electrical Engineering. Researches energy-efficient SoCs, DNN accelerators, and neuromorphic hardware.
Fabrizio Ottati

Fabrizio Ottati

AI/ML Processor Engineer at NXP, PhD from Politecnico di Torino. Focuses on event cameras, digital hardware, and deep learning. Maintains Tonic & Expelliarmus.

Inspired? Share your work.

Share your expertise with the community by speaking at a workshop, student talk, or hacking hour. It’s a great way to get feedback and help others learn.

Related Workshops

Does the Brain do Gradient Descent?

Does the Brain do Gradient Descent?

Explore the brain's potential use of gradient descent in learning processes with Konrad Kording in this engaging recorded session.

The TSP1 Neural Network Accelerator Chip: Advancing Brain-Inspired Computing

The TSP1 Neural Network Accelerator Chip: Advancing Brain-Inspired Computing

Join Chris Eliasmith for an in-depth exploration of the TSP1 chip from Applied Brain Research. Learn about this groundbreaking hardware platform and its implications for brain-inspired computing.

PEPITA - A Forward-Forward Alternative to Backpropagation

PEPITA - A Forward-Forward Alternative to Backpropagation

Explore PEPITA, a forward-forward approach as an alternative to backpropagation, presented by Giorgia Dellaferrera. Learn about its advantages and implementation with PyTorch.