SlideShare a Scribd company logo
1 of 20
Download to read offline
Efficient Neuromorphic Computing
with Dynamic Vision Sensor,
Spiking Neural Network Accelerator
and Hardware-Aware Algorithms
Jae-sun Seo
Arizona State University
• Neurons in biological nervous systems
communicate with ‘spikes’  binary pulses
• Spike represents events in
spiking neural networks (SNNs)
• Spiking activities are sparse
• Major dynamic energy only consumed when
spikes occur
• Spike-based, event-driven computing:
Low energy consumption
Event-Driven Neuromorphic Computing
2
Arizona State University
[1] S. Song, PLoS Biology, 2005
Artificial vs. Spiking Neuron Model
3
Arizona State University
𝑽𝒌
𝑳
(𝒕)
• Neuron behavior: integrate & fire
• Integrate membrane potential (Vk ) over time
with weighted sum of spikes & weights
• If membrane potential crosses threshold,
• Neuron spikes
• Membrane potential resets
𝑉𝑘
𝐿
𝑡 > 𝜃
𝑉𝑘 𝑡 = 𝑉𝑟𝑒𝑠𝑒𝑡
𝑉𝑘
𝐿
𝑡 = 𝑉𝑘
𝐿
𝑡 − 1 + ෍
𝑖=1
𝑚
𝑎𝑘
𝐿−1
𝑡 × 𝑤𝑖,𝑘
𝐿−1
+ 𝑏𝑘
𝐿
𝑎𝑘
𝐿
𝑡 = 1
[2] S. Yin et al., BioCAS 2017
SNN Training with Temporal Information
4
Arizona State University
t = 0
t = 1
t = 2
Residual membrane potential Vmem(t-1)
• Conventional training:
• 2-D tensor for input
• batch + neurons
• SNN training w/ time:
• 3-D tensor for input
• batch + neurons + time
• Difference with RNN:
• Include all time steps with same importance
(membrane potential integration)
ANN vs. SNN Accuracy for ImageNet Dataset
5
Arizona State University
• Similar to ANN’s success on accuracy improvement for ImageNet dataset,
recently there have been large improvements on SNN algorithms for ImageNet
ResNet-152
Spike-Norm
AlexNet
Spikeformer
S-ResNet
ResNeXt-101
BiT-L
BASIC-L
• Natural images (for computer vision tasks) are not in spiking format
 each input image needs to be converted into spikes via rate, burst, or latency encoding
• Such spike encoding implementation adds overhead on latency, energy, and area.
Spike Encoding for Input Images
6
Arizona State University
[3] A. Andreopoulos, et al.,
IBM Journal of Res. & Dev., 2015 Source: https://snntorch.readthedocs.io/
• If front-end sensor has outputs that are spikes, it can directly connect to SNNs
• Dynamic vision sensor, or event-based camera, sends spikes only when events occur
e.g. changes in pixels (instead of sending full image at 30 fps irrespective of events)
 fast event detection possible with low power
Event-based Vision Sensor for SNNs
7
Arizona State University
[4] [5] [6]
• Suitable applications: fast moving object detection, control based on it, etc.
• Conventional CIS will always have blurry images for fast moving objects
Advantages of Event-based Camera
8
Arizona State University
[5] B. Son, ISSCC, 2017
SNN Accuracy for DVS-CIFAR10 Dataset
9
Arizona State University
[7] H. Li, Front. Neuroscience, 2017
• DVS-CIFAR10: dataset of event-stream
recordings (done with DVS camera &
image movement) for classification of
10 different objects
• Various training techniques have been largely improving the SOTA accuracy for
event-based DVS-CIFAR10 dataset
TKS
TEBN
TJCA-SNN
PLIF
NeuNorm
BPSA+BPTA
• For on-device SNN inference, model size and total memory footprint are important
• Many SNN algorithms have been using FP32 precision to achieve high accuracy
• Our proposed SNN works: learnable threshold with low-precision quantization
SNN Accuracy vs. Model Size
10
Arizona State University
[8] E. Perot,
NeurIPS, 2020
Prophesee Gen1 dataset:
- 228,123 cars
- 27,658 pedestrians
- 39 hours, recorded w/
Prophesee Gen1 sensor
(1M pixel)
DVS-CIFAR10 Dataset Prophesee Gen1 Dataset
• For ANNs/DNNs, you don’t use the previous frame’s high-precision weighted sum value in the
current/future frames, so don’t need to store the weighted sums for each neuron
• For SNNs, events occur over time, and you need to accumulate membrane potential (neuron state)
over time for every neuron in the entire SNN
• Every neuron’s membrane potential is different  need to store each of them in entire SNN
• Total SNN memory = (weight_precision × # of weights) + (1-bit × # of neurons) +
(mem_pot_precision × # of neurons) + etc.
• Membrane potential memory: more significant for activation-heavy SNNs (e.g. high input resolution)
• For Prophesee Gen1 dataset, memory of VGG-11+SSD [10] - weight: 35MB, mem. pot.: 40MB
Total Memory Footprint for SNNs
11
Arizona State University
[9] SynSense, TinyML Talks, 2021
• Spike-FlowNet: spike-based optical flow detection
Event-based Computer Vision Algorithms
12
Arizona State University
[12] Prophesee, CVPR Workshop, 2019
MobileNet-V2
7.1M weights
Prophesee, Edge AI & Vision Alliance, 2019
• Prophesee: real-time detection of fast-moving objects
[11] C. Lee, ECCV, 2020
Digital Neuromorphic Chips in the Literature
13
Arizona State University
• A number of industrial/academic neuromorphic chips have been presented to date
• Quickly evolving SNN algorithms need to be accommodated
[13] Y. Sandamirskaya, Science Robotics, 2022
• Algorithm:
• SNN trained with Prophesee’s Gen1 dataset
• Object detection: YOLOv3 variant (backbone: ResNet18)
• After training, 8-bit quantization
• Frame stacking improves mAP
• Hardware:
• Custom SoC simulated: 16nm, 600MHz, w/ DRAM model
Custom Hardware for Event-based SNN
14
Arizona State University
[14] B. Crafton, AICAS, 2021
• Efficient SNN for compact MobileNet architecture with
configurable approximate computing
• Fine-grain pipelined architecture for low-precision SNN algorithms
• Prototype chip implemented with Intel 16 CMOS technology, 2mm x 2mm chip area
ASIC Chip Design for Mobile-SNN
15
Arizona State University
Anupreetham et al.,
ASU
• Event-based cameras produce a sparse stream of events that can be processed more
efficiently and with a lower latency than images, enabling ultra-fast vision-driven UAV control.
• Event-based vision algorithm implemented as SNN on Loihi chip  used in drone controller.
• Seamless integration of event-based perception on Loihi chip leads to
faster control rates and lower latency
Event-based Vision & Control for UAVs w/ Loihi
16
Arizona State University
[15] A. Vitale, ICRA, 2021
• Ideal end-to-end neuromorphic computing system will require & integrate:
• Front-end DVS: high-resolution event sensor w/ sparse spike outputs
• Mature event-based cameras are being commercially available
• Back-end SNN accelerator: low-power custom hardware fully exploiting sparsity
• Commercial/academia chips exist, but support for SOTA SNN algorithms isn’t clear
• Hardware-aware SNN algorithms: high-accuracy and compact algorithms
• SOTA SNN algorithms have shown noticeable accuracy improvement, while many still
use high-precision floating-point precision
• Large event-based dataset available and could be forthcoming
• Fitting neuromorphic applications:
• High-speed motion/object tracking, latency-sensitive tasks
• Smart drones, robotics with fast perception and control
Summary
17
Arizona State University
• Faculty: Priya Panda (Yale)
• Students: Jian Meng, Ahmed Hasssan, Anupreetham (ASU)
• Sponsors:
Acknowledgements
18
Arizona State University
[1] S. Song et al., “Highly Nonrandom Features of Synaptic Connectivity in Local Cortical Circuits,” PLOS
Biology, 2015.
[2] S. Yin et al., “Algorithm and Hardware Design of Discrete-time Spiking Neural Networks based on Back
Propagation with Binary Activations,” IEEE BioCAS, 2017.
[3] A. Andreopoulos et al., “Visual Saliency on Networks of Neurosynaptic Cores,” IBM Journal of Research
and Development, 2015.
[4] P. Lichtsteiner et al., “A 128×128 120 dB 15 μs Latency Asynchronous Temporal Contrast Vision Sensor,“
IEEE JSSC, 2008.
[5] B. Son et al., “A 640×480 Dynamic Vision Sensor with a 9µm Pixel and 300Meps Address-event
Representation,” IEEE ISSCC, 2017.
[6] T. Finateu et al., “A 1280×720 Back-Illuminated Stacked Temporal Contrast Event-Based Vision Sensor
with 4.86µm Pixels, 1.066GEPS Readout, Programmable Event-Rate Controller and Compressive Data-
Formatting Pipeline," IEEE ISSCC, 2020.
[7] H. Li et al., “CIFAR10-DVS: An Event-Stream Dataset for Object Classification,” Frontier of Neuroscience,
2017.
References
19
Arizona State University
[8] E. Perot et al., "Learning to Detect Objects with a 1 Megapixel Event Camera," NeurIPS, 2020.
[9] M. Sorbaro et al., “Always-on visual classification below 1 mWwith spiking convolutional networks on
Dynap™-CNN,” TinyML Talks, 2021.
[10] L. Cordone et al., “Object Detection with Spiking Neural Networks on Automotive Event Data”, IJCNN,
2022.
[11] C. Lee et al., “Spike-FlowNet: Event-based Optical Flow Estimation with Energy-efficient Hybrid Neural
Networks,” ECCV, 2020.
[12] A. Sironi et al., “Learning from Events: on the Future of Machine Learning for Event-based Cameras,”
CVPR Workshop, 2019.
[13] Y. Sandamirskaya et al., “Neuromorphic Computing Hardware and Neural Architectures for Robotics,”
Science Robotics, 2022.
[14] B. Crafton et al., “Hardware-Algorithm Co-Design Enabling Efficient Event-based Object Detection,”
IEEE AICAS, 2021.
[15] A. Vitale et al., “Event-driven Vision and Control for UAVs on a Neuromorphic Chip,” IEEE ICRA, 2021.
References
20
Arizona State University

More Related Content

What's hot

What's hot (20)

ICLR2020読み会 (neural-tangents)
ICLR2020読み会 (neural-tangents)ICLR2020読み会 (neural-tangents)
ICLR2020読み会 (neural-tangents)
 
Loihi many core_neuromorphic_chip
Loihi many core_neuromorphic_chipLoihi many core_neuromorphic_chip
Loihi many core_neuromorphic_chip
 
Pred net使ってみた
Pred net使ってみたPred net使ってみた
Pred net使ってみた
 
(DL輪読)Variational Dropout Sparsifies Deep Neural Networks
(DL輪読)Variational Dropout Sparsifies Deep Neural Networks(DL輪読)Variational Dropout Sparsifies Deep Neural Networks
(DL輪読)Variational Dropout Sparsifies Deep Neural Networks
 
【DL輪読会】A Path Towards Autonomous Machine Intelligence
【DL輪読会】A Path Towards Autonomous Machine Intelligence【DL輪読会】A Path Towards Autonomous Machine Intelligence
【DL輪読会】A Path Towards Autonomous Machine Intelligence
 
[DL輪読会]Meta-Learning Probabilistic Inference for Prediction
[DL輪読会]Meta-Learning Probabilistic Inference for Prediction[DL輪読会]Meta-Learning Probabilistic Inference for Prediction
[DL輪読会]Meta-Learning Probabilistic Inference for Prediction
 
[DL輪読会]Ensemble Distribution Distillation
[DL輪読会]Ensemble Distribution Distillation[DL輪読会]Ensemble Distribution Distillation
[DL輪読会]Ensemble Distribution Distillation
 
【DL輪読会】SimPer: Simple self-supervised learning of periodic targets( ICLR 2023 )
【DL輪読会】SimPer: Simple self-supervised learning of periodic targets( ICLR 2023 )【DL輪読会】SimPer: Simple self-supervised learning of periodic targets( ICLR 2023 )
【DL輪読会】SimPer: Simple self-supervised learning of periodic targets( ICLR 2023 )
 
[DL輪読会]Neuroscience-Inspired Artificial Intelligence
[DL輪読会]Neuroscience-Inspired Artificial Intelligence[DL輪読会]Neuroscience-Inspired Artificial Intelligence
[DL輪読会]Neuroscience-Inspired Artificial Intelligence
 
Quantum_Safe_Crypto_Overview_v3.pdf
Quantum_Safe_Crypto_Overview_v3.pdfQuantum_Safe_Crypto_Overview_v3.pdf
Quantum_Safe_Crypto_Overview_v3.pdf
 
20180115_東大医学部機能生物学セミナー_深層学習の最前線とこれから_岡野原大輔
20180115_東大医学部機能生物学セミナー_深層学習の最前線とこれから_岡野原大輔20180115_東大医学部機能生物学セミナー_深層学習の最前線とこれから_岡野原大輔
20180115_東大医学部機能生物学セミナー_深層学習の最前線とこれから_岡野原大輔
 
Introduction to Deep learning
Introduction to Deep learningIntroduction to Deep learning
Introduction to Deep learning
 
タクシー運行最適化を実現する機械学習システムの社会実装
タクシー運行最適化を実現する機械学習システムの社会実装タクシー運行最適化を実現する機械学習システムの社会実装
タクシー運行最適化を実現する機械学習システムの社会実装
 
ブラックボックスからXAI (説明可能なAI) へ - LIME (Local Interpretable Model-agnostic Explanat...
ブラックボックスからXAI (説明可能なAI) へ - LIME (Local Interpretable Model-agnostic Explanat...ブラックボックスからXAI (説明可能なAI) へ - LIME (Local Interpretable Model-agnostic Explanat...
ブラックボックスからXAI (説明可能なAI) へ - LIME (Local Interpretable Model-agnostic Explanat...
 
最近のディープラーニングのトレンド紹介_20200925
最近のディープラーニングのトレンド紹介_20200925最近のディープラーニングのトレンド紹介_20200925
最近のディープラーニングのトレンド紹介_20200925
 
Handwritten Digit Recognition using Convolutional Neural Networks
Handwritten Digit Recognition using Convolutional Neural  NetworksHandwritten Digit Recognition using Convolutional Neural  Networks
Handwritten Digit Recognition using Convolutional Neural Networks
 
batch-4.pptx
batch-4.pptxbatch-4.pptx
batch-4.pptx
 
"2D and 3D Sensing: Markets, Applications, and Technologies," a Presentation ...
"2D and 3D Sensing: Markets, Applications, and Technologies," a Presentation ..."2D and 3D Sensing: Markets, Applications, and Technologies," a Presentation ...
"2D and 3D Sensing: Markets, Applications, and Technologies," a Presentation ...
 
[DL輪読会]Attention is not Explanation (NAACL2019)
[DL輪読会]Attention is not Explanation (NAACL2019)[DL輪読会]Attention is not Explanation (NAACL2019)
[DL輪読会]Attention is not Explanation (NAACL2019)
 
World model
World modelWorld model
World model
 

Similar to “Efficient Neuromorphic Computing with Dynamic Vision Sensor, Spiking Neural Network Accelerator and Hardware-aware Algorithms,” a Presentation from Arizona State University

Integrating wireless sensor network into cloud services for real time data co...
Integrating wireless sensor network into cloud services for real time data co...Integrating wireless sensor network into cloud services for real time data co...
Integrating wireless sensor network into cloud services for real time data co...
Rajeev Piyare
 
final_project_1_2k21cse07.pptx
final_project_1_2k21cse07.pptxfinal_project_1_2k21cse07.pptx
final_project_1_2k21cse07.pptx
shwetabhagat25
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)
theijes
 

Similar to “Efficient Neuromorphic Computing with Dynamic Vision Sensor, Spiking Neural Network Accelerator and Hardware-aware Algorithms,” a Presentation from Arizona State University (20)

  Brain-Inspired Computation based on Spiking Neural Networks ...
 Brain-Inspired Computation based on Spiking Neural Networks               ... Brain-Inspired Computation based on Spiking Neural Networks               ...
  Brain-Inspired Computation based on Spiking Neural Networks ...
 
Seminar
SeminarSeminar
Seminar
 
Threat Detection in Surveillance Videos
Threat Detection in Surveillance VideosThreat Detection in Surveillance Videos
Threat Detection in Surveillance Videos
 
Integrating wireless sensor network into cloud services for real time data co...
Integrating wireless sensor network into cloud services for real time data co...Integrating wireless sensor network into cloud services for real time data co...
Integrating wireless sensor network into cloud services for real time data co...
 
Temporal Contrast Vision Sensor
Temporal Contrast Vision SensorTemporal Contrast Vision Sensor
Temporal Contrast Vision Sensor
 
Integrating Wireless Sensor Network into Cloud Services for Real-time Data Co...
Integrating Wireless Sensor Network into Cloud Services for Real-time Data Co...Integrating Wireless Sensor Network into Cloud Services for Real-time Data Co...
Integrating Wireless Sensor Network into Cloud Services for Real-time Data Co...
 
SECURE: Semantics Empowered resCUe enviRonmEnt
SECURE: Semantics Empowered resCUe enviRonmEntSECURE: Semantics Empowered resCUe enviRonmEnt
SECURE: Semantics Empowered resCUe enviRonmEnt
 
Object Detetcion using SSD-MobileNet
Object Detetcion using SSD-MobileNetObject Detetcion using SSD-MobileNet
Object Detetcion using SSD-MobileNet
 
MDM-2013, Milan, Italy, 6 June, 2013
MDM-2013, Milan, Italy, 6 June, 2013MDM-2013, Milan, Italy, 6 June, 2013
MDM-2013, Milan, Italy, 6 June, 2013
 
Dd net
Dd netDd net
Dd net
 
Aplications for machine learning in IoT
Aplications for machine learning in IoTAplications for machine learning in IoT
Aplications for machine learning in IoT
 
ACC-2012, Bangalore, India, 28 July, 2012
ACC-2012, Bangalore, India, 28 July, 2012ACC-2012, Bangalore, India, 28 July, 2012
ACC-2012, Bangalore, India, 28 July, 2012
 
BRAINREGION.pptx
BRAINREGION.pptxBRAINREGION.pptx
BRAINREGION.pptx
 
final_project_1_2k21cse07.pptx
final_project_1_2k21cse07.pptxfinal_project_1_2k21cse07.pptx
final_project_1_2k21cse07.pptx
 
Embedded Sensing and Computational Behaviour Science
Embedded Sensing and Computational Behaviour ScienceEmbedded Sensing and Computational Behaviour Science
Embedded Sensing and Computational Behaviour Science
 
Event Processing Using Semantic Web Technologies
Event Processing Using Semantic Web TechnologiesEvent Processing Using Semantic Web Technologies
Event Processing Using Semantic Web Technologies
 
PR-110: An Analysis of Scale Invariance in Object Detection – SNIP
PR-110: An Analysis of Scale Invariance in Object Detection – SNIPPR-110: An Analysis of Scale Invariance in Object Detection – SNIP
PR-110: An Analysis of Scale Invariance in Object Detection – SNIP
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)
 
Semantics in Sensor Networks
Semantics in Sensor NetworksSemantics in Sensor Networks
Semantics in Sensor Networks
 
EEG Based BCI Applications with Deep Learning
EEG Based BCI Applications with Deep LearningEEG Based BCI Applications with Deep Learning
EEG Based BCI Applications with Deep Learning
 

More from Edge AI and Vision Alliance

“Learning Compact DNN Models for Embedded Vision,” a Presentation from the Un...
“Learning Compact DNN Models for Embedded Vision,” a Presentation from the Un...“Learning Compact DNN Models for Embedded Vision,” a Presentation from the Un...
“Learning Compact DNN Models for Embedded Vision,” a Presentation from the Un...
Edge AI and Vision Alliance
 
“Selecting Tools for Developing, Monitoring and Maintaining ML Models,” a Pre...
“Selecting Tools for Developing, Monitoring and Maintaining ML Models,” a Pre...“Selecting Tools for Developing, Monitoring and Maintaining ML Models,” a Pre...
“Selecting Tools for Developing, Monitoring and Maintaining ML Models,” a Pre...
Edge AI and Vision Alliance
 
“Vision-language Representations for Robotics,” a Presentation from the Unive...
“Vision-language Representations for Robotics,” a Presentation from the Unive...“Vision-language Representations for Robotics,” a Presentation from the Unive...
“Vision-language Representations for Robotics,” a Presentation from the Unive...
Edge AI and Vision Alliance
 
“Computer Vision in Sports: Scalable Solutions for Downmarkets,” a Presentati...
“Computer Vision in Sports: Scalable Solutions for Downmarkets,” a Presentati...“Computer Vision in Sports: Scalable Solutions for Downmarkets,” a Presentati...
“Computer Vision in Sports: Scalable Solutions for Downmarkets,” a Presentati...
Edge AI and Vision Alliance
 
“AI Start-ups: The Perils of Fishing for Whales (War Stories from the Entrepr...
“AI Start-ups: The Perils of Fishing for Whales (War Stories from the Entrepr...“AI Start-ups: The Perils of Fishing for Whales (War Stories from the Entrepr...
“AI Start-ups: The Perils of Fishing for Whales (War Stories from the Entrepr...
Edge AI and Vision Alliance
 
“Bias in Computer Vision—It’s Bigger Than Facial Recognition!,” a Presentatio...
“Bias in Computer Vision—It’s Bigger Than Facial Recognition!,” a Presentatio...“Bias in Computer Vision—It’s Bigger Than Facial Recognition!,” a Presentatio...
“Bias in Computer Vision—It’s Bigger Than Facial Recognition!,” a Presentatio...
Edge AI and Vision Alliance
 
“Updating the Edge ML Development Process,” a Presentation from Samsara
“Updating the Edge ML Development Process,” a Presentation from Samsara“Updating the Edge ML Development Process,” a Presentation from Samsara
“Updating the Edge ML Development Process,” a Presentation from Samsara
Edge AI and Vision Alliance
 
“Developing an Embedded Vision AI-powered Fitness System,” a Presentation fro...
“Developing an Embedded Vision AI-powered Fitness System,” a Presentation fro...“Developing an Embedded Vision AI-powered Fitness System,” a Presentation fro...
“Developing an Embedded Vision AI-powered Fitness System,” a Presentation fro...
Edge AI and Vision Alliance
 

More from Edge AI and Vision Alliance (20)

“Learning Compact DNN Models for Embedded Vision,” a Presentation from the Un...
“Learning Compact DNN Models for Embedded Vision,” a Presentation from the Un...“Learning Compact DNN Models for Embedded Vision,” a Presentation from the Un...
“Learning Compact DNN Models for Embedded Vision,” a Presentation from the Un...
 
“Introduction to Computer Vision with CNNs,” a Presentation from Mohammad Hag...
“Introduction to Computer Vision with CNNs,” a Presentation from Mohammad Hag...“Introduction to Computer Vision with CNNs,” a Presentation from Mohammad Hag...
“Introduction to Computer Vision with CNNs,” a Presentation from Mohammad Hag...
 
“Selecting Tools for Developing, Monitoring and Maintaining ML Models,” a Pre...
“Selecting Tools for Developing, Monitoring and Maintaining ML Models,” a Pre...“Selecting Tools for Developing, Monitoring and Maintaining ML Models,” a Pre...
“Selecting Tools for Developing, Monitoring and Maintaining ML Models,” a Pre...
 
“Building Accelerated GStreamer Applications for Video and Audio AI,” a Prese...
“Building Accelerated GStreamer Applications for Video and Audio AI,” a Prese...“Building Accelerated GStreamer Applications for Video and Audio AI,” a Prese...
“Building Accelerated GStreamer Applications for Video and Audio AI,” a Prese...
 
“Understanding, Selecting and Optimizing Object Detectors for Edge Applicatio...
“Understanding, Selecting and Optimizing Object Detectors for Edge Applicatio...“Understanding, Selecting and Optimizing Object Detectors for Edge Applicatio...
“Understanding, Selecting and Optimizing Object Detectors for Edge Applicatio...
 
“Introduction to Modern LiDAR for Machine Perception,” a Presentation from th...
“Introduction to Modern LiDAR for Machine Perception,” a Presentation from th...“Introduction to Modern LiDAR for Machine Perception,” a Presentation from th...
“Introduction to Modern LiDAR for Machine Perception,” a Presentation from th...
 
“Vision-language Representations for Robotics,” a Presentation from the Unive...
“Vision-language Representations for Robotics,” a Presentation from the Unive...“Vision-language Representations for Robotics,” a Presentation from the Unive...
“Vision-language Representations for Robotics,” a Presentation from the Unive...
 
“ADAS and AV Sensors: What’s Winning and Why?,” a Presentation from TechInsights
“ADAS and AV Sensors: What’s Winning and Why?,” a Presentation from TechInsights“ADAS and AV Sensors: What’s Winning and Why?,” a Presentation from TechInsights
“ADAS and AV Sensors: What’s Winning and Why?,” a Presentation from TechInsights
 
“Computer Vision in Sports: Scalable Solutions for Downmarkets,” a Presentati...
“Computer Vision in Sports: Scalable Solutions for Downmarkets,” a Presentati...“Computer Vision in Sports: Scalable Solutions for Downmarkets,” a Presentati...
“Computer Vision in Sports: Scalable Solutions for Downmarkets,” a Presentati...
 
“Detecting Data Drift in Image Classification Neural Networks,” a Presentatio...
“Detecting Data Drift in Image Classification Neural Networks,” a Presentatio...“Detecting Data Drift in Image Classification Neural Networks,” a Presentatio...
“Detecting Data Drift in Image Classification Neural Networks,” a Presentatio...
 
“Deep Neural Network Training: Diagnosing Problems and Implementing Solutions...
“Deep Neural Network Training: Diagnosing Problems and Implementing Solutions...“Deep Neural Network Training: Diagnosing Problems and Implementing Solutions...
“Deep Neural Network Training: Diagnosing Problems and Implementing Solutions...
 
“AI Start-ups: The Perils of Fishing for Whales (War Stories from the Entrepr...
“AI Start-ups: The Perils of Fishing for Whales (War Stories from the Entrepr...“AI Start-ups: The Perils of Fishing for Whales (War Stories from the Entrepr...
“AI Start-ups: The Perils of Fishing for Whales (War Stories from the Entrepr...
 
“A Computer Vision System for Autonomous Satellite Maneuvering,” a Presentati...
“A Computer Vision System for Autonomous Satellite Maneuvering,” a Presentati...“A Computer Vision System for Autonomous Satellite Maneuvering,” a Presentati...
“A Computer Vision System for Autonomous Satellite Maneuvering,” a Presentati...
 
“Bias in Computer Vision—It’s Bigger Than Facial Recognition!,” a Presentatio...
“Bias in Computer Vision—It’s Bigger Than Facial Recognition!,” a Presentatio...“Bias in Computer Vision—It’s Bigger Than Facial Recognition!,” a Presentatio...
“Bias in Computer Vision—It’s Bigger Than Facial Recognition!,” a Presentatio...
 
“Sensor Fusion Techniques for Accurate Perception of Objects in the Environme...
“Sensor Fusion Techniques for Accurate Perception of Objects in the Environme...“Sensor Fusion Techniques for Accurate Perception of Objects in the Environme...
“Sensor Fusion Techniques for Accurate Perception of Objects in the Environme...
 
“Updating the Edge ML Development Process,” a Presentation from Samsara
“Updating the Edge ML Development Process,” a Presentation from Samsara“Updating the Edge ML Development Process,” a Presentation from Samsara
“Updating the Edge ML Development Process,” a Presentation from Samsara
 
“Combating Bias in Production Computer Vision Systems,” a Presentation from R...
“Combating Bias in Production Computer Vision Systems,” a Presentation from R...“Combating Bias in Production Computer Vision Systems,” a Presentation from R...
“Combating Bias in Production Computer Vision Systems,” a Presentation from R...
 
“Developing an Embedded Vision AI-powered Fitness System,” a Presentation fro...
“Developing an Embedded Vision AI-powered Fitness System,” a Presentation fro...“Developing an Embedded Vision AI-powered Fitness System,” a Presentation fro...
“Developing an Embedded Vision AI-powered Fitness System,” a Presentation fro...
 
“Navigating the Evolving Venture Capital Landscape for Edge AI Start-ups,” a ...
“Navigating the Evolving Venture Capital Landscape for Edge AI Start-ups,” a ...“Navigating the Evolving Venture Capital Landscape for Edge AI Start-ups,” a ...
“Navigating the Evolving Venture Capital Landscape for Edge AI Start-ups,” a ...
 
“Advanced Presence Sensing: What It Means for the Smart Home,” a Presentation...
“Advanced Presence Sensing: What It Means for the Smart Home,” a Presentation...“Advanced Presence Sensing: What It Means for the Smart Home,” a Presentation...
“Advanced Presence Sensing: What It Means for the Smart Home,” a Presentation...
 

Recently uploaded

Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Safe Software
 
Architecting Cloud Native Applications
Architecting Cloud Native ApplicationsArchitecting Cloud Native Applications
Architecting Cloud Native Applications
WSO2
 
TrustArc Webinar - Unified Trust Center for Privacy, Security, Compliance, an...
TrustArc Webinar - Unified Trust Center for Privacy, Security, Compliance, an...TrustArc Webinar - Unified Trust Center for Privacy, Security, Compliance, an...
TrustArc Webinar - Unified Trust Center for Privacy, Security, Compliance, an...
TrustArc
 

Recently uploaded (20)

JavaScript Usage Statistics 2024 - The Ultimate Guide
JavaScript Usage Statistics 2024 - The Ultimate GuideJavaScript Usage Statistics 2024 - The Ultimate Guide
JavaScript Usage Statistics 2024 - The Ultimate Guide
 
DBX First Quarter 2024 Investor Presentation
DBX First Quarter 2024 Investor PresentationDBX First Quarter 2024 Investor Presentation
DBX First Quarter 2024 Investor Presentation
 
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
 
ChatGPT and Beyond - Elevating DevOps Productivity
ChatGPT and Beyond - Elevating DevOps ProductivityChatGPT and Beyond - Elevating DevOps Productivity
ChatGPT and Beyond - Elevating DevOps Productivity
 
Polkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin WoodPolkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin Wood
 
Architecting Cloud Native Applications
Architecting Cloud Native ApplicationsArchitecting Cloud Native Applications
Architecting Cloud Native Applications
 
Introduction to use of FHIR Documents in ABDM
Introduction to use of FHIR Documents in ABDMIntroduction to use of FHIR Documents in ABDM
Introduction to use of FHIR Documents in ABDM
 
TrustArc Webinar - Unified Trust Center for Privacy, Security, Compliance, an...
TrustArc Webinar - Unified Trust Center for Privacy, Security, Compliance, an...TrustArc Webinar - Unified Trust Center for Privacy, Security, Compliance, an...
TrustArc Webinar - Unified Trust Center for Privacy, Security, Compliance, an...
 
"I see eyes in my soup": How Delivery Hero implemented the safety system for ...
"I see eyes in my soup": How Delivery Hero implemented the safety system for ..."I see eyes in my soup": How Delivery Hero implemented the safety system for ...
"I see eyes in my soup": How Delivery Hero implemented the safety system for ...
 
The Zero-ETL Approach: Enhancing Data Agility and Insight
The Zero-ETL Approach: Enhancing Data Agility and InsightThe Zero-ETL Approach: Enhancing Data Agility and Insight
The Zero-ETL Approach: Enhancing Data Agility and Insight
 
DEV meet-up UiPath Document Understanding May 7 2024 Amsterdam
DEV meet-up UiPath Document Understanding May 7 2024 AmsterdamDEV meet-up UiPath Document Understanding May 7 2024 Amsterdam
DEV meet-up UiPath Document Understanding May 7 2024 Amsterdam
 
Quantum Leap in Next-Generation Computing
Quantum Leap in Next-Generation ComputingQuantum Leap in Next-Generation Computing
Quantum Leap in Next-Generation Computing
 
Platformless Horizons for Digital Adaptability
Platformless Horizons for Digital AdaptabilityPlatformless Horizons for Digital Adaptability
Platformless Horizons for Digital Adaptability
 
Understanding the FAA Part 107 License ..
Understanding the FAA Part 107 License ..Understanding the FAA Part 107 License ..
Understanding the FAA Part 107 License ..
 
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost SavingRepurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
 
MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024
 
AI in Action: Real World Use Cases by Anitaraj
AI in Action: Real World Use Cases by AnitarajAI in Action: Real World Use Cases by Anitaraj
AI in Action: Real World Use Cases by Anitaraj
 
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdfRising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
 
Choreo: Empowering the Future of Enterprise Software Engineering
Choreo: Empowering the Future of Enterprise Software EngineeringChoreo: Empowering the Future of Enterprise Software Engineering
Choreo: Empowering the Future of Enterprise Software Engineering
 
Vector Search -An Introduction in Oracle Database 23ai.pptx
Vector Search -An Introduction in Oracle Database 23ai.pptxVector Search -An Introduction in Oracle Database 23ai.pptx
Vector Search -An Introduction in Oracle Database 23ai.pptx
 

“Efficient Neuromorphic Computing with Dynamic Vision Sensor, Spiking Neural Network Accelerator and Hardware-aware Algorithms,” a Presentation from Arizona State University

  • 1. Efficient Neuromorphic Computing with Dynamic Vision Sensor, Spiking Neural Network Accelerator and Hardware-Aware Algorithms Jae-sun Seo Arizona State University
  • 2. • Neurons in biological nervous systems communicate with ‘spikes’  binary pulses • Spike represents events in spiking neural networks (SNNs) • Spiking activities are sparse • Major dynamic energy only consumed when spikes occur • Spike-based, event-driven computing: Low energy consumption Event-Driven Neuromorphic Computing 2 Arizona State University [1] S. Song, PLoS Biology, 2005
  • 3. Artificial vs. Spiking Neuron Model 3 Arizona State University 𝑽𝒌 𝑳 (𝒕) • Neuron behavior: integrate & fire • Integrate membrane potential (Vk ) over time with weighted sum of spikes & weights • If membrane potential crosses threshold, • Neuron spikes • Membrane potential resets 𝑉𝑘 𝐿 𝑡 > 𝜃 𝑉𝑘 𝑡 = 𝑉𝑟𝑒𝑠𝑒𝑡 𝑉𝑘 𝐿 𝑡 = 𝑉𝑘 𝐿 𝑡 − 1 + ෍ 𝑖=1 𝑚 𝑎𝑘 𝐿−1 𝑡 × 𝑤𝑖,𝑘 𝐿−1 + 𝑏𝑘 𝐿 𝑎𝑘 𝐿 𝑡 = 1 [2] S. Yin et al., BioCAS 2017
  • 4. SNN Training with Temporal Information 4 Arizona State University t = 0 t = 1 t = 2 Residual membrane potential Vmem(t-1) • Conventional training: • 2-D tensor for input • batch + neurons • SNN training w/ time: • 3-D tensor for input • batch + neurons + time • Difference with RNN: • Include all time steps with same importance (membrane potential integration)
  • 5. ANN vs. SNN Accuracy for ImageNet Dataset 5 Arizona State University • Similar to ANN’s success on accuracy improvement for ImageNet dataset, recently there have been large improvements on SNN algorithms for ImageNet ResNet-152 Spike-Norm AlexNet Spikeformer S-ResNet ResNeXt-101 BiT-L BASIC-L
  • 6. • Natural images (for computer vision tasks) are not in spiking format  each input image needs to be converted into spikes via rate, burst, or latency encoding • Such spike encoding implementation adds overhead on latency, energy, and area. Spike Encoding for Input Images 6 Arizona State University [3] A. Andreopoulos, et al., IBM Journal of Res. & Dev., 2015 Source: https://snntorch.readthedocs.io/
  • 7. • If front-end sensor has outputs that are spikes, it can directly connect to SNNs • Dynamic vision sensor, or event-based camera, sends spikes only when events occur e.g. changes in pixels (instead of sending full image at 30 fps irrespective of events)  fast event detection possible with low power Event-based Vision Sensor for SNNs 7 Arizona State University [4] [5] [6]
  • 8. • Suitable applications: fast moving object detection, control based on it, etc. • Conventional CIS will always have blurry images for fast moving objects Advantages of Event-based Camera 8 Arizona State University [5] B. Son, ISSCC, 2017
  • 9. SNN Accuracy for DVS-CIFAR10 Dataset 9 Arizona State University [7] H. Li, Front. Neuroscience, 2017 • DVS-CIFAR10: dataset of event-stream recordings (done with DVS camera & image movement) for classification of 10 different objects • Various training techniques have been largely improving the SOTA accuracy for event-based DVS-CIFAR10 dataset TKS TEBN TJCA-SNN PLIF NeuNorm BPSA+BPTA
  • 10. • For on-device SNN inference, model size and total memory footprint are important • Many SNN algorithms have been using FP32 precision to achieve high accuracy • Our proposed SNN works: learnable threshold with low-precision quantization SNN Accuracy vs. Model Size 10 Arizona State University [8] E. Perot, NeurIPS, 2020 Prophesee Gen1 dataset: - 228,123 cars - 27,658 pedestrians - 39 hours, recorded w/ Prophesee Gen1 sensor (1M pixel) DVS-CIFAR10 Dataset Prophesee Gen1 Dataset
  • 11. • For ANNs/DNNs, you don’t use the previous frame’s high-precision weighted sum value in the current/future frames, so don’t need to store the weighted sums for each neuron • For SNNs, events occur over time, and you need to accumulate membrane potential (neuron state) over time for every neuron in the entire SNN • Every neuron’s membrane potential is different  need to store each of them in entire SNN • Total SNN memory = (weight_precision × # of weights) + (1-bit × # of neurons) + (mem_pot_precision × # of neurons) + etc. • Membrane potential memory: more significant for activation-heavy SNNs (e.g. high input resolution) • For Prophesee Gen1 dataset, memory of VGG-11+SSD [10] - weight: 35MB, mem. pot.: 40MB Total Memory Footprint for SNNs 11 Arizona State University [9] SynSense, TinyML Talks, 2021
  • 12. • Spike-FlowNet: spike-based optical flow detection Event-based Computer Vision Algorithms 12 Arizona State University [12] Prophesee, CVPR Workshop, 2019 MobileNet-V2 7.1M weights Prophesee, Edge AI & Vision Alliance, 2019 • Prophesee: real-time detection of fast-moving objects [11] C. Lee, ECCV, 2020
  • 13. Digital Neuromorphic Chips in the Literature 13 Arizona State University • A number of industrial/academic neuromorphic chips have been presented to date • Quickly evolving SNN algorithms need to be accommodated [13] Y. Sandamirskaya, Science Robotics, 2022
  • 14. • Algorithm: • SNN trained with Prophesee’s Gen1 dataset • Object detection: YOLOv3 variant (backbone: ResNet18) • After training, 8-bit quantization • Frame stacking improves mAP • Hardware: • Custom SoC simulated: 16nm, 600MHz, w/ DRAM model Custom Hardware for Event-based SNN 14 Arizona State University [14] B. Crafton, AICAS, 2021
  • 15. • Efficient SNN for compact MobileNet architecture with configurable approximate computing • Fine-grain pipelined architecture for low-precision SNN algorithms • Prototype chip implemented with Intel 16 CMOS technology, 2mm x 2mm chip area ASIC Chip Design for Mobile-SNN 15 Arizona State University Anupreetham et al., ASU
  • 16. • Event-based cameras produce a sparse stream of events that can be processed more efficiently and with a lower latency than images, enabling ultra-fast vision-driven UAV control. • Event-based vision algorithm implemented as SNN on Loihi chip  used in drone controller. • Seamless integration of event-based perception on Loihi chip leads to faster control rates and lower latency Event-based Vision & Control for UAVs w/ Loihi 16 Arizona State University [15] A. Vitale, ICRA, 2021
  • 17. • Ideal end-to-end neuromorphic computing system will require & integrate: • Front-end DVS: high-resolution event sensor w/ sparse spike outputs • Mature event-based cameras are being commercially available • Back-end SNN accelerator: low-power custom hardware fully exploiting sparsity • Commercial/academia chips exist, but support for SOTA SNN algorithms isn’t clear • Hardware-aware SNN algorithms: high-accuracy and compact algorithms • SOTA SNN algorithms have shown noticeable accuracy improvement, while many still use high-precision floating-point precision • Large event-based dataset available and could be forthcoming • Fitting neuromorphic applications: • High-speed motion/object tracking, latency-sensitive tasks • Smart drones, robotics with fast perception and control Summary 17 Arizona State University
  • 18. • Faculty: Priya Panda (Yale) • Students: Jian Meng, Ahmed Hasssan, Anupreetham (ASU) • Sponsors: Acknowledgements 18 Arizona State University
  • 19. [1] S. Song et al., “Highly Nonrandom Features of Synaptic Connectivity in Local Cortical Circuits,” PLOS Biology, 2015. [2] S. Yin et al., “Algorithm and Hardware Design of Discrete-time Spiking Neural Networks based on Back Propagation with Binary Activations,” IEEE BioCAS, 2017. [3] A. Andreopoulos et al., “Visual Saliency on Networks of Neurosynaptic Cores,” IBM Journal of Research and Development, 2015. [4] P. Lichtsteiner et al., “A 128×128 120 dB 15 μs Latency Asynchronous Temporal Contrast Vision Sensor,“ IEEE JSSC, 2008. [5] B. Son et al., “A 640×480 Dynamic Vision Sensor with a 9µm Pixel and 300Meps Address-event Representation,” IEEE ISSCC, 2017. [6] T. Finateu et al., “A 1280×720 Back-Illuminated Stacked Temporal Contrast Event-Based Vision Sensor with 4.86µm Pixels, 1.066GEPS Readout, Programmable Event-Rate Controller and Compressive Data- Formatting Pipeline," IEEE ISSCC, 2020. [7] H. Li et al., “CIFAR10-DVS: An Event-Stream Dataset for Object Classification,” Frontier of Neuroscience, 2017. References 19 Arizona State University
  • 20. [8] E. Perot et al., "Learning to Detect Objects with a 1 Megapixel Event Camera," NeurIPS, 2020. [9] M. Sorbaro et al., “Always-on visual classification below 1 mWwith spiking convolutional networks on Dynap™-CNN,” TinyML Talks, 2021. [10] L. Cordone et al., “Object Detection with Spiking Neural Networks on Automotive Event Data”, IJCNN, 2022. [11] C. Lee et al., “Spike-FlowNet: Event-based Optical Flow Estimation with Energy-efficient Hybrid Neural Networks,” ECCV, 2020. [12] A. Sironi et al., “Learning from Events: on the Future of Machine Learning for Event-based Cameras,” CVPR Workshop, 2019. [13] Y. Sandamirskaya et al., “Neuromorphic Computing Hardware and Neural Architectures for Robotics,” Science Robotics, 2022. [14] B. Crafton et al., “Hardware-Algorithm Co-Design Enabling Efficient Event-based Object Detection,” IEEE AICAS, 2021. [15] A. Vitale et al., “Event-driven Vision and Control for UAVs on a Neuromorphic Chip,” IEEE ICRA, 2021. References 20 Arizona State University