The document provides an introduction to spiking neural networks (SNNs) and neuromorphic computing. It discusses the characteristics and advantages of SNNs, including their spatio-temporal nature, asynchronous processing, sparsity, and energy efficiency. It also covers basic neuroscience concepts like neurons, action potentials, synaptic plasticity, and learning rules like STDP. Common SNN models and neural encoding schemes are described. Examples of SNN applications in visual processing and pattern generation are presented. Finally, neuromorphic hardware platforms like Intel's Loihi chip are introduced.
Introduction to Spiking Neural Networks: From a Computational Neuroscience pe...Jason Tsai
Abstract:
Being the third generation of neural network models, the study of spiking neural networks is an interdisciplinary field among brain science, theoretical neuroscience, and artificial neural networks research. Recently it is gaining attention and momentum, especially in neuromorphic device design for real-time machine learning. Some of you might have heard of it, but its underneath principles probably remain unknown for most of you. In this talk, I will briefly illustrate the basic building blocks of this emerging architecture and technology.
Introduction to Spiking Neural Networks: From a Computational Neuroscience pe...Jason Tsai
Abstract:
Being the third generation of neural network models, the study of spiking neural networks is an interdisciplinary field among brain science, theoretical neuroscience, and artificial neural networks research. Recently it is gaining attention and momentum, especially in neuromorphic device design for real-time machine learning. Some of you might have heard of it, but its underneath principles probably remain unknown for most of you. In this talk, I will briefly illustrate the basic building blocks of this emerging architecture and technology.
It’s long ago, approx. 30 years, since AI was not only a topic for Science-Fiction writers, but also a major research field surrounded with huge hopes and investments. But the over-inflated expectations ended in a subsequent crash and followed by a period of absent funding and interest – the so-called AI winter. However, the last 3 years changed everything – again. Deep learning, a machine learning technique inspired by the human brain, successfully crushed one benchmark after another and tech companies, like Google, Facebook and Microsoft, started to invest billions in AI research. “The pace of progress in artificial general intelligence is incredible fast” (Elon Musk – CEO Tesla & SpaceX) leading to an AI that “would be either the best or the worst thing ever to happen to humanity” (Stephen Hawking – Physicist).
What sparked this new Hype? How is Deep Learning different from previous approaches? Are the advancing AI technologies really a threat for humanity? Let’s look behind the curtain and unravel the reality. This talk will explore why Sundar Pichai (CEO Google) recently announced that “machine learning is a core transformative way by which Google is rethinking everything they are doing” and explain why "Deep Learning is probably one of the most exciting things that is happening in the computer industry” (Jen-Hsun Huang – CEO NVIDIA).
Either a new AI “winter is coming” (Ned Stark – House Stark) or this new wave of innovation might turn out as the “last invention humans ever need to make” (Nick Bostrom – AI Philosoph). Or maybe it’s just another great technology helping humans to achieve more.
Survey of Attention mechanism & Use in Computer VisionSwatiNarkhede1
This presentation contains the overview of Attention models. It also has information of the stand alone self attention model used for Computer Vision tasks.
Part 2 of the Deep Learning Fundamentals Series, this session discusses Tuning Training (including hyperparameters, overfitting/underfitting), Training Algorithms (including different learning rates, backpropagation), Optimization (including stochastic gradient descent, momentum, Nesterov Accelerated Gradient, RMSprop, Adaptive algorithms - Adam, Adadelta, etc.), and a primer on Convolutional Neural Networks. The demos included in these slides are running on Keras with TensorFlow backend on Databricks.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/10/deploying-pytorch-models-for-real-time-inference-on-the-edge-a-presentation-from-nomitri/
Moritz August, CDO at Nomitri GmbH, presents the “Deploying PyTorch Models for Real-time Inference On the Edge” tutorial at the May 2021 Embedded Vision Summit.
In this presentation, August provides an overview of workflows for deploying compressed deep learning models, starting with PyTorch and creating native C++ application code running in real-time on embedded hardware platforms. He illustrates these workflows on smartphones with real-world examples targeting ARM-based CPU, GPUs, and NPUs as well as embedded chips and modules like the NXP i.MX8+ and NVIDIA Jetson Nano.
August examines TorchScript, architecture-side optimizations, quantization and common pitfalls. Additionally, he shows how the PyTorch deployment workflow can be extended to conversion to ONNX and quantization of ONNX models using an ONNX Runtime. On the application side, he demonstrates how deployed models can be integrated efficiently into a C++ library that runs natively on mobile and embedded devices and highlights known limitations.
Variational Autoencoders For Image GenerationJason Anderson
Meetup: https://www.meetup.com/Cognitive-Computing-Enthusiasts/events/260580395/
Video: https://www.youtube.com/watch?v=fnULFOyNZn8
Blog: http://www.compthree.com/blog/autoencoder/
Code: https://github.com/compthree/variational-autoencoder
An autoencoder is a machine learning algorithm that represents unlabeled high-dimensional data as points in a low-dimensional space. A variational autoencoder (VAE) is an autoencoder that represents unlabeled high-dimensional data as low-dimensional probability distributions. In addition to data compression, the randomness of the VAE algorithm gives it a second powerful feature: the ability to generate new data similar to its training data. For example, a VAE trained on images of faces can generate a compelling image of a new "fake" face. It can also map new features onto input data, such as glasses or a mustache onto the image of a face that initially lacks these features. In this talk, we will survey VAE model designs that use deep learning, and we will implement a basic VAE in TensorFlow. We will also demonstrate the encoding and generative capabilities of VAEs and discuss their industry applications.
Artificial Intelligence: Artificial Neural NetworksThe Integral Worm
This presentation covers artificial neural networks for artificial intelligence. Topics covered are as follows: artificial neural networks, basic representation, hidden units, exclusive OR problem, backpropagation, advantages of artificial neural networks, properties of artificial neural networks, and disadvantages of artificial neural networks.
파이콘 한국 2020) 파이썬으로 구현하는 신경세포 기반의 인공 뇌 시뮬레이터Seonghyun Kim
* 파이콘 한국 2020의 발표자료입니다.
현대 인공 신경망의 뿌리가 되었던 뇌 과학!
이 발표에서는 인공 신경망에 대한 뇌 과학적 접근과,
뇌 세포의 발화를 모사하는 파이썬 기반의 뉴로모픽 신경망 모델에 대한 사례를 공유할 예정입니다.
뉴로모픽 신경망은 단순히 기존의 딥러닝에서 셀 구조만을 변경한 것이 아닙니다.
실제로 실험을 수행하기 어려운 생물학적 한계점을 뇌 시뮬레이션을 통해서 극복할 수 있으며,
나아가 뇌의 정보처리 메커니즘을 밝히고, 다양한 뇌 질환 치료제의 타겟을 연구하는데 아주 중요한 역할을 할 수 있습니다.
이번 발표를 통해, 기계학습을 연구하고 있는 많은 연구자 분들에게 새로운 아이디어에 대한 영감이 될 수 있기를 희망합니다.
Brain-Inspired Computation based on Spiking Neural Networks ...Jorge Pires
On this live, prof. Kasabov gives us a gentle overview of Spiking Neural Networks, and their current applications
Full live here, with discussion: https://www.youtube.com/watch?v=niAannUB3pc&t=232s
Have fun 😎😂😁😀
It’s long ago, approx. 30 years, since AI was not only a topic for Science-Fiction writers, but also a major research field surrounded with huge hopes and investments. But the over-inflated expectations ended in a subsequent crash and followed by a period of absent funding and interest – the so-called AI winter. However, the last 3 years changed everything – again. Deep learning, a machine learning technique inspired by the human brain, successfully crushed one benchmark after another and tech companies, like Google, Facebook and Microsoft, started to invest billions in AI research. “The pace of progress in artificial general intelligence is incredible fast” (Elon Musk – CEO Tesla & SpaceX) leading to an AI that “would be either the best or the worst thing ever to happen to humanity” (Stephen Hawking – Physicist).
What sparked this new Hype? How is Deep Learning different from previous approaches? Are the advancing AI technologies really a threat for humanity? Let’s look behind the curtain and unravel the reality. This talk will explore why Sundar Pichai (CEO Google) recently announced that “machine learning is a core transformative way by which Google is rethinking everything they are doing” and explain why "Deep Learning is probably one of the most exciting things that is happening in the computer industry” (Jen-Hsun Huang – CEO NVIDIA).
Either a new AI “winter is coming” (Ned Stark – House Stark) or this new wave of innovation might turn out as the “last invention humans ever need to make” (Nick Bostrom – AI Philosoph). Or maybe it’s just another great technology helping humans to achieve more.
Survey of Attention mechanism & Use in Computer VisionSwatiNarkhede1
This presentation contains the overview of Attention models. It also has information of the stand alone self attention model used for Computer Vision tasks.
Part 2 of the Deep Learning Fundamentals Series, this session discusses Tuning Training (including hyperparameters, overfitting/underfitting), Training Algorithms (including different learning rates, backpropagation), Optimization (including stochastic gradient descent, momentum, Nesterov Accelerated Gradient, RMSprop, Adaptive algorithms - Adam, Adadelta, etc.), and a primer on Convolutional Neural Networks. The demos included in these slides are running on Keras with TensorFlow backend on Databricks.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/10/deploying-pytorch-models-for-real-time-inference-on-the-edge-a-presentation-from-nomitri/
Moritz August, CDO at Nomitri GmbH, presents the “Deploying PyTorch Models for Real-time Inference On the Edge” tutorial at the May 2021 Embedded Vision Summit.
In this presentation, August provides an overview of workflows for deploying compressed deep learning models, starting with PyTorch and creating native C++ application code running in real-time on embedded hardware platforms. He illustrates these workflows on smartphones with real-world examples targeting ARM-based CPU, GPUs, and NPUs as well as embedded chips and modules like the NXP i.MX8+ and NVIDIA Jetson Nano.
August examines TorchScript, architecture-side optimizations, quantization and common pitfalls. Additionally, he shows how the PyTorch deployment workflow can be extended to conversion to ONNX and quantization of ONNX models using an ONNX Runtime. On the application side, he demonstrates how deployed models can be integrated efficiently into a C++ library that runs natively on mobile and embedded devices and highlights known limitations.
Variational Autoencoders For Image GenerationJason Anderson
Meetup: https://www.meetup.com/Cognitive-Computing-Enthusiasts/events/260580395/
Video: https://www.youtube.com/watch?v=fnULFOyNZn8
Blog: http://www.compthree.com/blog/autoencoder/
Code: https://github.com/compthree/variational-autoencoder
An autoencoder is a machine learning algorithm that represents unlabeled high-dimensional data as points in a low-dimensional space. A variational autoencoder (VAE) is an autoencoder that represents unlabeled high-dimensional data as low-dimensional probability distributions. In addition to data compression, the randomness of the VAE algorithm gives it a second powerful feature: the ability to generate new data similar to its training data. For example, a VAE trained on images of faces can generate a compelling image of a new "fake" face. It can also map new features onto input data, such as glasses or a mustache onto the image of a face that initially lacks these features. In this talk, we will survey VAE model designs that use deep learning, and we will implement a basic VAE in TensorFlow. We will also demonstrate the encoding and generative capabilities of VAEs and discuss their industry applications.
Artificial Intelligence: Artificial Neural NetworksThe Integral Worm
This presentation covers artificial neural networks for artificial intelligence. Topics covered are as follows: artificial neural networks, basic representation, hidden units, exclusive OR problem, backpropagation, advantages of artificial neural networks, properties of artificial neural networks, and disadvantages of artificial neural networks.
파이콘 한국 2020) 파이썬으로 구현하는 신경세포 기반의 인공 뇌 시뮬레이터Seonghyun Kim
* 파이콘 한국 2020의 발표자료입니다.
현대 인공 신경망의 뿌리가 되었던 뇌 과학!
이 발표에서는 인공 신경망에 대한 뇌 과학적 접근과,
뇌 세포의 발화를 모사하는 파이썬 기반의 뉴로모픽 신경망 모델에 대한 사례를 공유할 예정입니다.
뉴로모픽 신경망은 단순히 기존의 딥러닝에서 셀 구조만을 변경한 것이 아닙니다.
실제로 실험을 수행하기 어려운 생물학적 한계점을 뇌 시뮬레이션을 통해서 극복할 수 있으며,
나아가 뇌의 정보처리 메커니즘을 밝히고, 다양한 뇌 질환 치료제의 타겟을 연구하는데 아주 중요한 역할을 할 수 있습니다.
이번 발표를 통해, 기계학습을 연구하고 있는 많은 연구자 분들에게 새로운 아이디어에 대한 영감이 될 수 있기를 희망합니다.
Brain-Inspired Computation based on Spiking Neural Networks ...Jorge Pires
On this live, prof. Kasabov gives us a gentle overview of Spiking Neural Networks, and their current applications
Full live here, with discussion: https://www.youtube.com/watch?v=niAannUB3pc&t=232s
Have fun 😎😂😁😀
PowerPoint slides from a 2015 Guest Lecture in PSYCH-268A: Computational Neuroscience, Prof. Jeff Krichmar, University of California, Irvine (UCI).
Corresponding publication:
Beyeler*, M., Carlson*, K. D. , Chou*, T-S., Dutt, N., Krichmar, J. L. (2015). CARLsim 3: A user-friendly and highly optimized library for the creation of neurobiologically detailed spiking neural networks. Proceedings of IEEE International Joint Conference on Neural Networks (IJCNN), Killarney, Ireland. (*equal contribution)
Robust Feature Learning with Deep Neural Networks
http://snu-primo.hosted.exlibrisgroup.com/primo_library/libweb/action/display.do?tabs=viewOnlineTab&doc=82SNU_INST21557911060002591
Artificial Neural Network and its Applicationsshritosh kumar
Abstract
This report is an introduction to Artificial Neural
Networks. The various types of neural networks are
explained and demonstrated, applications of neural
networks like ANNs in medicine are described, and a
detailed historical background is provided. The
connection between the artificial and the real thing is
also investigated and explained. Finally, the
mathematical models involved are presented and
demonstrated.
Lecture for Reinforcement Learning study group held on August 19th, 2017.
Reference book: http://incompleteideas.net/book/the-book.html
Video: https://youtu.be/xv5ZsOSf6ZQ
Initiated by Taiwan AI Group (https://www.facebook.com/groups/Taiwan.AI.Group/permalink/1796526840669749/)
Apical-amplification, apical-isolation, apical-drive. two-compartment spiking model. ThetaPlanes piecewise linear approximation of mutlicompartment neuron activity. Sleep passed the evolutionary siege in all studied animal species, notwithstanding its apparent unproductivity (lower reactivity to external dangers, no feeding, no mating). In humans, the time spent in sleep is higher in younger individuals, precisely when learning is faster. Another element to be considered is that, thanks to an evolutionary history that spanned hundreds of millions of years and selected among countless individuals, the inter-areal and local connectome captures the priors necessary to optimize the flow and combination of internal hypotheses and sensorial evidence.
At the cellular level, optimal combination of contextual information and local computation is provided by the apical amplification principle, active during wakefulness. Deep-sleep (NREM) and REM sleep are characterized in mammals by pyramidal neurons changing to a different management of apical signals, namely apical-isolation and apical-drive.
The cognitive and energetic functions of sleep and its relations with awake performance have beeninvestigated by INFN in spiking models, engaged in learning and sleep cycles, that will be presented in this seminar. Also, preliminar information about a next generation of neural models supporting apical mechanisms will be presented.
A short talk for a forum held by Taiwan Association for Human Rights: https://www.tahr.org.tw/event/2670
Video: https://youtu.be/-hYQRHqyR9g (28:10 - 50:35)
Lecture for Neural Networks study group held on February 8, 2020.
Reference book: http://hagan.okstate.edu/nnd.html
Video: https://youtu.be/TyyoPU13ME0
Python demo codes: https://bit.ly/3893GHB
Initiated by Taiwan AI Group (https://www.facebook.com/groups/Taiwan.AI.Group/permalink/2017771298545301/)
Lecture for Neural Networks study group held on January 11, 2020.
Reference book: http://hagan.okstate.edu/nnd.html
Video: https://youtu.be/H4NKgliTFUw
Initiated by Taiwan AI Group (https://www.facebook.com/groups/Taiwan.AI.Group/permalink/2017771298545301/)
Deep Learning: Chapter 11 Practical MethodologyJason Tsai
Lecture for Deep Learning 101 study group to be held on June 9th, 2017.
Reference book: https://www.deeplearningbook.org/
Past video archives: https://goo.gl/hxermB
Initiated by Taiwan AI Group (https://www.facebook.com/groups/Taiwan.AI.Group/)
Deep Learning: Introduction & Chapter 5 Machine Learning BasicsJason Tsai
Given lecture for Deep Learning 101 study group with Frank Wu on Dec. 9th, 2016.
Reference: https://www.deeplearningbook.org/
Initiated by Taiwan AI Group (https://www.facebook.com/groups/Taiwan.AI.Group/)
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
Key Trends Shaping the Future of Infrastructure.pdf
Introduction to Spiking Neural Networks: From a Computational Neuroscience perspective
1. Jason Tsai (蔡志順)
Oct. 11, 2018 @台灣人工智慧學校 新竹分校
INTRODUCTION TO SPIKING NEURAL
NETWORKS
*Picture adopted from
https://bit.ly/2Rh7cYy
2. *Copyright Notice:
All figures in this presentation are taken from
the quoted sources as mentioned in the
respective slides and their copyright belongs
to the owners. This presentation itself adopts
Creative Commons license.
3. Neural Networks 3D Simulation
(Video demo)
*Video from https://youtu.be/3JQ3hYko51Y
4. Questions
What are the advantages of spiking
neural networks and neuromorphic
computing?
What are current challenges of spiking
neural networks (SNNs)?
9. Neuron’s Spike: Action Potential
*Figure adopted from https://en.wikipedia.org/wiki/Action_potential & The front cover of
“Spikes: Exploring the Neural Code (1999)”
10. The Effect of Presynaptic Spikes on
Postsynaptic Neuron
*Figure adopted from Wulfram Gerstner & Werner M. Kistler. Spiking Neuron Models:
Single Neurons, Populations, Plasticity. Cambridge University Press. 2002. Page 5.
11. The Firing of a Leaky Integrate-and-
Fire Model Neuron
*Figure adopted from https://doi.org/10.1371/journal.pone.0001377
12. Hebb’s Learning Postulate
"When an axon of cell A is near enough to excite a cell B and
repeatedly or persistently takes part in firing it, some growth
process or metabolic change takes place in one or both cells such
that A's efficiency, as one of the cells firing B, is increased.“*
* Refer to Donald O. Hebb, The Organization of Behavior: A Neuropsychological Theory. 1949 & 2002. Page 62.
Causality
Repetition
13. Long-Term Potentiation (LTP) / Long-
Term Depression (LTD)
LTP is a long-lasting, activity-dependent increase in synaptic
strength that is a leading candidate as a cellular mechanism
contributing to memory formation in mammals in a very
broadly applicable sense.*
* Refer to J. David Sweatt. Mechanisms of Memory, Second Edition. Academic Press. 2010. Page 112.
14. Synaptic Plasticity
*Figure adopted from Wulfram Gerstner & Werner M. Kistler. Spiking Neuron Models:
Single Neurons, Populations, Plasticity. Cambridge University Press. 2002. Page 353.
15. Back-propagating Action Potential (bAP)
*Further reading: https://en.wikipedia.org/wiki/Neural_backpropagation
Induction of tLTP requires activation of the presynaptic
input milliseconds before the bAP in the postsynaptic
dendrite.
16. *Figure adopted from https://doi.org/10.3389/fnsyn.2011.00004
Spike-Timing-Dependent Plasticity
(STDP)
17. Experiment Evidence of STDP
From Wikipedia:
“Henry Markram, when he was in Bert Sakmann's lab and published their
work in 1997, used dual patch clamping techniques to repetitively
activate pre-synaptic neurons 10 milliseconds before activating the post-
synaptic target neurons, and found the strength of the synapse
increased. When the activation order was reversed so that the pre-
synaptic neuron was activated 10 milliseconds after its post-synaptic
target neuron, the strength of the pre-to-post synaptic connection
decreased.
Further work, by Guoqiang Bi, Li Zhang, and Huizhong Tao in Mu-Ming
Poo's lab in 1998, continued the mapping of the entire time course
relating pre- and post-synaptic activity and synaptic change, to show that
in their preparation synapses that are activated within 5-20 ms before a
postsynaptic spike are strengthened, and those that are activated within a
similar time window after the spike are weakened.”
*Further reading: https://en.wikipedia.org/wiki/Spike-timing-dependent_plasticity
18. Lateral Inhibition
Lateral inhibition is a Central Nervous System process whereby
application of a stimulus to the center of the receptive field excites a
neuron, but a stimulus applied near the edge inhibits it.
*Figure adopted from https://bit.ly/2yaat37
21. Dopamine: Essential for Reward
Processing in Mammalian Brain
*Figure adopted from http://www.jneurosci.org/content/29/2/444
Dopamine neurons form huge synaptic contacts to target!
23. Two Hot Approaches
Supervised: Stochastic Gradient Descent
based Backpropagation learning rule
(Treat the membrane potentials of spiking neurons as
differentiable signals, where discontinuities at spike
times are considered as noise.*)
Unsupervised: STDP (Spike-Timing-
Dependent Plasticity) based learning rule
*Refer to Jun Haeng Lee, et al., Training Deep Spiking Neural Networks Using Backpropagation.
Frontiers in Neuroscience, 08 November 2016. https://doi.org/10.3389/fnins.2016.00508
24. *Refer to Yu, Q., Tang, H., Hu, J., Tan, K.C., Neuromorphic Cognitive Systems: A Learning and Memory
Centered Approach. Springer International Publishing. 2017. Page 9.
STDP Learning Rule
25. STDP Learning Rule (1-to-1)
*Figure adopted from http://dx.doi.org/10.7551/978-0-262-33027-5-ch037
26. STDP Learning Rule (2-to-1)
N0 is stimulated until N1 fires, then e0 is stopped for 30 ms.
N2 is stimulated by e2 during those 30 ms.
*Figure adopted from http://dx.doi.org/10.7551/978-0-262-33027-5-ch037
27. STDP Finds Spike Patterns
*Figure adopted from https://doi.org/10.1371/journal.pone.0001377
30. 1st Generation of Neuron Models
(McCulloch–Pitts Neuron Model)
*Figure adopted from http://wwwold.ece.utep.edu/research/webfuzzy/docs/kk-thesis/kk-thesis-html/node12.html
31. 2nd Generation of Neuron Models
*Figure adopted from http://cs231n.github.io/neural-networks-1/
32. 3rd Generation of Neuron Models
(Spiking Neuron Models)
*Figure adopted from http://kzyjc.cnjournals.com/html/2018/5/20180512.htm
33. Spiking Neuron Models
Miscellaneous models:
Hodgkin-Huxley model
Izhikevish model
Leaky Integrate-and-Fire (LIF) model
Spike Response model (SRM)
……
*Further reading: https://en.wikipedia.org/wiki/Biological_neuron_model
& http://www.scholarpedia.org/article/Spike-response_model
34. Hodgkin-Huxley Model
*Figure adopted from Wulfram Gerstner & Werner M. Kistler. Spiking Neuron Models: Single Neurons,
Populations, Plasticity. Cambridge University Press. 2002. Page 34.
38. Leaky Integrate-and-Fire Model
*Figure adopted from Wulfram Gerstner, Werner M. Kistler, Richard Naud and Liam Paninski “Neuronal Dynamics:
From Single Neurons to Networks and Models of Cognition” Cambridge University Press. 2014. Page 11.
44. Sparse Coding
*Figure adopted from http://brainworkshow.sparsey.com/measuring-similarity-in-localist-vs-distributed-representations/
45. Sparse Coding with Inhibitory Neurons
Population sparseness: Few neurons are
active at any given time
Lifetime sparseness: Individual neurons
are responsive to few specific stimuli
*Figure adopted from https://doi.org/10.1523/JNEUROSCI.4188-12.2013
48. Refer to “Milad Mozafari, et al., First-spike-based
visual categorization using reward-modulated STDP
(2018)” https://doi.org/10.1109/TNNLS.2018.2826721
54. Refer to “Luziwei Leng, et al., Spiking neurons with
short-term synaptic plasticity form superior
generative networks (2018)”
https://doi.org/10.1038/s41598-018-28999-2
62. ANN-to-SNN Conversion
Train ANNs using standard supervised training
techniques like backpropagation to leverage
the superior performance of trained ANNs and
subsequently convert to event-driven SNNs for
inference operation on neuromorphic platform.
Rate-encoded spikes are approximately
proportional to the magnitude of the original
ANN inputs.
66. Further Reading
Wulfram Gerstner & Werner M. Kistler, “Spiking Neuron Models:
Single Neurons, Populations, Plasticity”. Cambridge University
Press (2002)
Wulfram Gerstner, Werner M. Kistler, Richard Naud and Liam
Paninski, “Neuronal Dynamics: From Single Neurons to Networks
and Models of Cognition”. Cambridge University Press (2014)
Eugene M. Izhikevich, “The Dynamical Systems in Neuroscience:
Geometry of Excitability and Bursting”. The MIT Press (2007)
Nikola K. Kasabov, “Time-Space, Spiking Neural Networks and
Brain-Inspired Artificial Intelligence”. Springer International
Publishing (2018)
Amirhossein Tavanaei, et al., “Deep Learning in Spiking Neural
Networks”. arXiv:1804.08150 (2018)