The document discusses various pooling operations used in image processing and convolutional neural networks (CNNs). It provides an overview of common pooling methods like max pooling, average pooling, and spatial pyramid pooling. It also discusses more advanced and trainable pooling techniques like stochastic pooling, mixed/gated pooling, fractional pooling, local importance pooling, and global feature guided local pooling. The document analyzes the tradeoffs of different pooling methods and how they can balance preserving details versus achieving invariance to changes in position or lighting. It references several influential papers that analyzed properties of pooling operations.
In machine learning, a convolutional neural network is a class of deep, feed-forward artificial neural networks that have successfully been applied fpr analyzing visual imagery.
Deep learning (also known as deep structured learning or hierarchical learning) is the application of artificial neural networks (ANNs) to learning tasks that contain more than one hidden layer. Deep learning is part of a broader family of machine learning methods based on learning data representations, as opposed to task-specific algorithms. Learning can be supervised, partially supervised or unsupervised.
In machine learning, a convolutional neural network is a class of deep, feed-forward artificial neural networks that have successfully been applied fpr analyzing visual imagery.
Deep learning (also known as deep structured learning or hierarchical learning) is the application of artificial neural networks (ANNs) to learning tasks that contain more than one hidden layer. Deep learning is part of a broader family of machine learning methods based on learning data representations, as opposed to task-specific algorithms. Learning can be supervised, partially supervised or unsupervised.
What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...Simplilearn
This Deep Learning Presentation will help you in understanding what is Deep learning, why do we need Deep learning, applications of Deep Learning along with a detailed explanation on Neural Networks and how these Neural Networks work. Deep learning is inspired by the integral function of the human brain specific to artificial neural networks. These networks, which represent the decision-making process of the brain, use complex algorithms that process data in a non-linear way, learning in an unsupervised manner to make choices based on the input. This Deep Learning tutorial is ideal for professionals with beginners to intermediate levels of experience. Now, let us dive deep into this topic and understand what Deep learning actually is.
Below topics are explained in this Deep Learning Presentation:
1. What is Deep Learning?
2. Why do we need Deep Learning?
3. Applications of Deep Learning
4. What is Neural Network?
5. Activation Functions
6. Working of Neural Network
Simplilearn’s Deep Learning course will transform you into an expert in deep learning techniques using TensorFlow, the open-source software library designed to conduct machine learning & deep neural network research. With our deep learning course, you’ll master deep learning and TensorFlow concepts, learn to implement algorithms, build artificial neural networks and traverse layers of data abstraction to understand the power of data and prepare you for your new role as deep learning scientist.
Why Deep Learning?
It is one of the most popular software platforms used for deep learning and contains powerful tools to help you build and implement artificial neural networks.
Advancements in deep learning are being seen in smartphone applications, creating efficiencies in the power grid, driving advancements in healthcare, improving agricultural yields, and helping us find solutions to climate change. With this Tensorflow course, you’ll build expertise in deep learning models, learn to operate TensorFlow to manage neural networks and interpret the results.
You can gain in-depth knowledge of Deep Learning by taking our Deep Learning certification training course. With Simplilearn’s Deep Learning course, you will prepare for a career as a Deep Learning engineer as you master concepts and techniques including supervised and unsupervised learning, mathematical and heuristic aspects, and hands-on modeling to develop algorithms.
There is booming demand for skilled deep learning engineers across a wide range of industries, making this deep learning course with TensorFlow training well-suited for professionals at the intermediate to advanced level of experience. We recommend this deep learning online course particularly for the following professionals:
1. Software engineers
2. Data scientists
3. Data analysts
4. Statisticians with an interest in deep learning
Hot Topics in Machine Learning For Research and thesisWriteMyThesis
Machine Learning and its subsequent fields have undergone tremendous growth in the past few years. It has a number of potential applications and is being used in different fields. A lot of research work is going on in this field. For more information, check out the PPT details...
Talk @ ACM SF Bayarea Chapter on Deep Learning for medical imaging space.
The talk covers use cases, special challenges and solutions for Deep Learning for Medical Image Analysis using Tensorflow+Keras. You will learn about:
- Use cases for Deep Learning in Medical Image Analysis
- Different DNN architectures used for Medical Image Analysis
- Special purpose compute / accelerators for Deep Learning (in the Cloud / On-prem)
- How to parallelize your models for faster training of models and serving for inferenceing.
- Optimization techniques to get the best performance from your cluster (like Kubernetes/ Apache Mesos / Spark)
- How to build an efficient Data Pipeline for Medical Image Analysis using Deep Learning
- Resources to jump start your journey - like public data sets, common models used in Medical Image Analysis
Organizations are collecting massive amounts of data from disparate sources. However, they continuously face the challenge of identifying patterns, detecting anomalies, and projecting future trends based on large data sets. Machine learning for anomaly detection provides a promising alternative for the detection and classification of anomalies.
Find out how you can implement machine learning to increase speed and effectiveness in identifying and reporting anomalies.
In this webinar, we will discuss :
How machine learning can help in identifying anomalies
Steps to approach an anomaly detection problem
Various techniques available for anomaly detection
Best algorithms that fit in different situations
Implementing an anomaly detection use case on the StreamAnalytix platform
To view the webinar - https://bit.ly/2IV2ahC
Introduction to Graph Neural Networks: Basics and Applications - Katsuhiko Is...Preferred Networks
This presentation explains basic ideas of graph neural networks (GNNs) and their common applications. Primary target audiences are students, engineers and researchers who are new to GNNs but interested in using GNNs for their projects. This is a modified version of the course material for a special lecture on Data Science at Nara Institute of Science and Technology (NAIST), given by Preferred Networks researcher Katsuhiko Ishiguro, PhD.
https://telecombcn-dl.github.io/2017-dlcv/
Deep learning technologies are at the core of the current revolution in artificial intelligence for multimedia data analysis. The convergence of large-scale annotated datasets and affordable GPU hardware has allowed the training of neural networks for data analysis tasks which were previously addressed with hand-crafted features. Architectures such as convolutional neural networks, recurrent neural networks and Q-nets for reinforcement learning have shaped a brand new scenario in signal processing. This course will cover the basic principles and applications of deep learning to computer vision problems, such as image classification, object detection or image captioning.
In this presentation we discuss the convolution operation, the architecture of a convolution neural network, different layers such as pooling etc. This presentation draws heavily from A Karpathy's Stanford Course CS 231n
Tensor Field Network (and other ConvNet Generalisations)Peng Cheng
In this session we will discuss the relationship between data augmentation, invariant/equivariant features and the abstract concept of convolution layers, specifically, how this abstraction can be extended to devise concrete neural network architectures that are robust to diverse data and augmentation types (all of which are published after 2016). We will focus on the latest of the series, designed to handle spatial graph data augmentable by 3d translations & rotations. In the end, we will showcase it's applications in molecule analysis and autonomous flight. We will focus on the latest of the series, designed to handle spatial graph data augmentable by 3d translations & rotations. In the end, we will showcase it's applications in molecule analysis and autonomous flight.
What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...Simplilearn
This Deep Learning Presentation will help you in understanding what is Deep learning, why do we need Deep learning, applications of Deep Learning along with a detailed explanation on Neural Networks and how these Neural Networks work. Deep learning is inspired by the integral function of the human brain specific to artificial neural networks. These networks, which represent the decision-making process of the brain, use complex algorithms that process data in a non-linear way, learning in an unsupervised manner to make choices based on the input. This Deep Learning tutorial is ideal for professionals with beginners to intermediate levels of experience. Now, let us dive deep into this topic and understand what Deep learning actually is.
Below topics are explained in this Deep Learning Presentation:
1. What is Deep Learning?
2. Why do we need Deep Learning?
3. Applications of Deep Learning
4. What is Neural Network?
5. Activation Functions
6. Working of Neural Network
Simplilearn’s Deep Learning course will transform you into an expert in deep learning techniques using TensorFlow, the open-source software library designed to conduct machine learning & deep neural network research. With our deep learning course, you’ll master deep learning and TensorFlow concepts, learn to implement algorithms, build artificial neural networks and traverse layers of data abstraction to understand the power of data and prepare you for your new role as deep learning scientist.
Why Deep Learning?
It is one of the most popular software platforms used for deep learning and contains powerful tools to help you build and implement artificial neural networks.
Advancements in deep learning are being seen in smartphone applications, creating efficiencies in the power grid, driving advancements in healthcare, improving agricultural yields, and helping us find solutions to climate change. With this Tensorflow course, you’ll build expertise in deep learning models, learn to operate TensorFlow to manage neural networks and interpret the results.
You can gain in-depth knowledge of Deep Learning by taking our Deep Learning certification training course. With Simplilearn’s Deep Learning course, you will prepare for a career as a Deep Learning engineer as you master concepts and techniques including supervised and unsupervised learning, mathematical and heuristic aspects, and hands-on modeling to develop algorithms.
There is booming demand for skilled deep learning engineers across a wide range of industries, making this deep learning course with TensorFlow training well-suited for professionals at the intermediate to advanced level of experience. We recommend this deep learning online course particularly for the following professionals:
1. Software engineers
2. Data scientists
3. Data analysts
4. Statisticians with an interest in deep learning
Hot Topics in Machine Learning For Research and thesisWriteMyThesis
Machine Learning and its subsequent fields have undergone tremendous growth in the past few years. It has a number of potential applications and is being used in different fields. A lot of research work is going on in this field. For more information, check out the PPT details...
Talk @ ACM SF Bayarea Chapter on Deep Learning for medical imaging space.
The talk covers use cases, special challenges and solutions for Deep Learning for Medical Image Analysis using Tensorflow+Keras. You will learn about:
- Use cases for Deep Learning in Medical Image Analysis
- Different DNN architectures used for Medical Image Analysis
- Special purpose compute / accelerators for Deep Learning (in the Cloud / On-prem)
- How to parallelize your models for faster training of models and serving for inferenceing.
- Optimization techniques to get the best performance from your cluster (like Kubernetes/ Apache Mesos / Spark)
- How to build an efficient Data Pipeline for Medical Image Analysis using Deep Learning
- Resources to jump start your journey - like public data sets, common models used in Medical Image Analysis
Organizations are collecting massive amounts of data from disparate sources. However, they continuously face the challenge of identifying patterns, detecting anomalies, and projecting future trends based on large data sets. Machine learning for anomaly detection provides a promising alternative for the detection and classification of anomalies.
Find out how you can implement machine learning to increase speed and effectiveness in identifying and reporting anomalies.
In this webinar, we will discuss :
How machine learning can help in identifying anomalies
Steps to approach an anomaly detection problem
Various techniques available for anomaly detection
Best algorithms that fit in different situations
Implementing an anomaly detection use case on the StreamAnalytix platform
To view the webinar - https://bit.ly/2IV2ahC
Introduction to Graph Neural Networks: Basics and Applications - Katsuhiko Is...Preferred Networks
This presentation explains basic ideas of graph neural networks (GNNs) and their common applications. Primary target audiences are students, engineers and researchers who are new to GNNs but interested in using GNNs for their projects. This is a modified version of the course material for a special lecture on Data Science at Nara Institute of Science and Technology (NAIST), given by Preferred Networks researcher Katsuhiko Ishiguro, PhD.
https://telecombcn-dl.github.io/2017-dlcv/
Deep learning technologies are at the core of the current revolution in artificial intelligence for multimedia data analysis. The convergence of large-scale annotated datasets and affordable GPU hardware has allowed the training of neural networks for data analysis tasks which were previously addressed with hand-crafted features. Architectures such as convolutional neural networks, recurrent neural networks and Q-nets for reinforcement learning have shaped a brand new scenario in signal processing. This course will cover the basic principles and applications of deep learning to computer vision problems, such as image classification, object detection or image captioning.
In this presentation we discuss the convolution operation, the architecture of a convolution neural network, different layers such as pooling etc. This presentation draws heavily from A Karpathy's Stanford Course CS 231n
Tensor Field Network (and other ConvNet Generalisations)Peng Cheng
In this session we will discuss the relationship between data augmentation, invariant/equivariant features and the abstract concept of convolution layers, specifically, how this abstraction can be extended to devise concrete neural network architectures that are robust to diverse data and augmentation types (all of which are published after 2016). We will focus on the latest of the series, designed to handle spatial graph data augmentable by 3d translations & rotations. In the end, we will showcase it's applications in molecule analysis and autonomous flight. We will focus on the latest of the series, designed to handle spatial graph data augmentable by 3d translations & rotations. In the end, we will showcase it's applications in molecule analysis and autonomous flight.
Unexpected Challenges in Large Scale Machine Learning by Charles ParkerBigMine
Talk by Charles Parker (BigML) at BigMine12 at KDD12.
In machine learning, scale adds complexity. The most obvious consequence of scale is that data takes longer to process. At certain points, however, scale makes trivial operations costly, thus forcing us to re-evaluate algorithms in light of the complexity of those operations. Here, we will discuss one important way a general large scale machine learning setting may differ from the standard supervised classification setting and show some the results of some preliminary experiments highlighting this difference. The results suggest that there is potential for significant improvement beyond obvious solutions.
The slides for the first ever SnappyData webinar. Covers SnappyData core concepts, programming models, benchmarks and more.
SnappyData is open sourced here: https://github.com/SnappyDataInc/snappydata
We also have a deep technical paper here: http://www.snappydata.io/snappy-industrial
We can be easily contacted on Slack, Gitter and more: http://www.snappydata.io/about#contactus
Spark is rapidly catching fire with the machine learning and data science community for a number of reasons. Predominantly, it is making it possible to extend and enhance machine learning algorithms to a level we’ve never seen before. In this talk, we’ll give examples of two areas Alpine Data Labs has contributed to the Spark project:
Bio:
DB Tsai is a Machine Learning Engineer working at Alpine Data Labs. His current focus is on Big Data, Data Mining, and Machine Learning. He uses Hadoop, Spark, Mahout, and several Machine Learning algorithms to build powerful, scalable, and robust cloud-driven applications. His favorite programming languages are Java, Scala, and Python. DB is a Ph.D. candidate in Applied Physics at Stanford University (currently taking leave of absence). He holds a Master’s degree in Electrical Engineering from Stanford University, as well as a Master's degree in Physics from National Taiwan University.
This report is executed by StorPool Storage and compares the block storage
offerings of well-known public clouds
(AWS, Digital Ocean, OVH, DreamHost)
with a number of StorPool-based public
cloud offerings.
Big Data Real Time Analytics - A Facebook Case StudyNati Shalom
Building Your Own Facebook Real Time Analytics System with Cassandra and GigaSpaces.
Facebook's real time analytics system is a good reference for those looking to build their real time analytics system for big data.
The first part covers the lessons from Facebook's experience and the reason they chose HBase over Cassandra.
In the second part of the session, we learn how we can build our own Real Time Analytics system, achieve better performance, gain real business insights, and business analytics on our big data, and make the deployment and scaling significantly simpler using the new version of Cassandra and GigaSpaces Cloudify.
RNNs for Recommendations and PersonalizationNick Pentreath
In the last few years, RNNs have achieved significant success in modeling time series and sequence data, in particular within the speech, language, and text domains. Recently, these techniques have been begun to be applied to session-based recommendation tasks, with very promising results. Nick Pentreath explores the latest research advances in this domain, as well as practical applications.
Analysis of the Pending Interest Table behavior in the context of a distributed denial of service attack.
Slides presented at:
3rd ACM SIGCOMM Workshop on Information-Centric Networking (ICN 2013) - Hong Kong, China
The paper is available at:
http://conferences.sigcomm.org/sigcomm/2013/papers/icn/p67.pdf
Magnet Shuffle Service: Push-based Shuffle at LinkedInDatabricks
The number of daily Apache Spark applications at LinkedIn has increased by 3X in the past year. The shuffle process alone, which is one of the most costly operators in batch computation, is processing PBs of data and billions of blocks daily in our clusters. With such a rapid increase of Apache Spark workloads, we quickly realized that the shuffle process can become a severe bottleneck for both infrastructure scalability and workloads efficiency. In our production clusters, we have observed both reliability issues due to shuffle fetch connection failures and efficiency issues due to the random reads of small shuffle blocks on HDDs.
To tackle those challenges and optimize shuffle performance in Apache Spark, we have developed Magnet shuffle service, a push-based shuffle mechanism that works natively with Apache Spark. Our paper on Magnet has been accepted by VLDB 2020. In this talk, we will introduce how push-based shuffle can drastically increase shuffle efficiency when compared with the existing pull-based shuffle. In addition, by combining push-based shuffle and pull-based shuffle, we show how Magnet shuffle service helps to harden shuffle infrastructure at LinkedIn scale by both reducing shuffle related failures and removing scaling bottlenecks. Furthermore, we will share our experiences of productionizing Magnet at LinkedIn to process close to 10 PB of daily shuffle data.
Scalable Machine Learning: The Role of Stratified Data Shardinginside-BigData.com
In this deck from the 2019 Stanford HPC Conference, Srinivasan Parthasarathy from Ohio State University presents: Scalable Machine Learning: The Role of Stratified Data Sharding.
"With the increasing popularity of structured data stores, social networks and Web 2.0 and 3.0 applications, complex data formats, such as trees and graphs, are becoming ubiquitous. Managing and learning from such large and complex data stores, on modern computational eco-systems, to realize actionable information efficiently, is daunting. In this talk I will begin with discussing some of these challenges. Subsequently I will discuss a critical element at the heart of this challenge relates to the sharding, placement, storage and access of such tera- and peta- scale data. In this work we develop a novel distributed framework to ease the burden on the programmer and propose an agile and intelligent placement service layer as a flexible yet unified means to address this challenge. Central to our framework is the notion of stratification which seeks to initially group structurally (or semantically) similar entities into strata. Subsequently strata are partitioned within this eco-system according to the needs of the application to maximize locality, balance load, minimize data skew or even take into account energy consumption. Results on several real-world applications validate the efficacy and efficiency of our approach. (Notes: Joint work with Y. Wang (Airbnb) and A. Chakrabarti (MSR))."
Srinivasan Parthasarathy, Professor of Computer Science & Engineering, The Ohio State University
Srinivasan Parthasarathy is a Professor of Computer Science and Engineering and the director of the data mining research laboratory at Ohio State. His research interests span databases, data mining and high performance computing. He is among a handful of researchers nationwide to have won both the Department of Energy and National Science Foundation Career awards. He and his students have won multiple best paper awards or "best of" nominations from leading forums in the field including: SIAM Data Mining, ACM SIGKDD, VLDB, ISMB, WWW, ICDM, and ACM Bioinformatics. He chairs the SIAM data mining conference steering committee and serves on the action board of ACM TKDD and ACM DMKD --leading journals in the field. Since 2012 he also helped lead the creation of OSU's first-of-a-kind nationwide (USA) undergraduate major in data analytics and serves as one of its founding directors.
Watch the video: https://youtu.be/hOJI8e0p-UI
Learn more: http://web.cse.ohio-state.edu/~parthasarathy.2/
and
http://hpcadvisorycouncil.com/events/2019/stanford-workshop/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Show drafts
volume_up
Empowering the Data Analytics Ecosystem: A Laser Focus on Value
The data analytics ecosystem thrives when every component functions at its peak, unlocking the true potential of data. Here's a laser focus on key areas for an empowered ecosystem:
1. Democratize Access, Not Data:
Granular Access Controls: Provide users with self-service tools tailored to their specific needs, preventing data overload and misuse.
Data Catalogs: Implement robust data catalogs for easy discovery and understanding of available data sources.
2. Foster Collaboration with Clear Roles:
Data Mesh Architecture: Break down data silos by creating a distributed data ownership model with clear ownership and responsibilities.
Collaborative Workspaces: Utilize interactive platforms where data scientists, analysts, and domain experts can work seamlessly together.
3. Leverage Advanced Analytics Strategically:
AI-powered Automation: Automate repetitive tasks like data cleaning and feature engineering, freeing up data talent for higher-level analysis.
Right-Tool Selection: Strategically choose the most effective advanced analytics techniques (e.g., AI, ML) based on specific business problems.
4. Prioritize Data Quality with Automation:
Automated Data Validation: Implement automated data quality checks to identify and rectify errors at the source, minimizing downstream issues.
Data Lineage Tracking: Track the flow of data throughout the ecosystem, ensuring transparency and facilitating root cause analysis for errors.
5. Cultivate a Data-Driven Mindset:
Metrics-Driven Performance Management: Align KPIs and performance metrics with data-driven insights to ensure actionable decision making.
Data Storytelling Workshops: Equip stakeholders with the skills to translate complex data findings into compelling narratives that drive action.
Benefits of a Precise Ecosystem:
Sharpened Focus: Precise access and clear roles ensure everyone works with the most relevant data, maximizing efficiency.
Actionable Insights: Strategic analytics and automated quality checks lead to more reliable and actionable data insights.
Continuous Improvement: Data-driven performance management fosters a culture of learning and continuous improvement.
Sustainable Growth: Empowered by data, organizations can make informed decisions to drive sustainable growth and innovation.
By focusing on these precise actions, organizations can create an empowered data analytics ecosystem that delivers real value by driving data-driven decisions and maximizing the return on their data investment.
Explore our comprehensive data analysis project presentation on predicting product ad campaign performance. Learn how data-driven insights can optimize your marketing strategies and enhance campaign effectiveness. Perfect for professionals and students looking to understand the power of data analysis in advertising. for more details visit: https://bostoninstituteofanalytics.org/data-science-and-artificial-intelligence/
Chatty Kathy - UNC Bootcamp Final Project Presentation - Final Version - 5.23...John Andrews
SlideShare Description for "Chatty Kathy - UNC Bootcamp Final Project Presentation"
Title: Chatty Kathy: Enhancing Physical Activity Among Older Adults
Description:
Discover how Chatty Kathy, an innovative project developed at the UNC Bootcamp, aims to tackle the challenge of low physical activity among older adults. Our AI-driven solution uses peer interaction to boost and sustain exercise levels, significantly improving health outcomes. This presentation covers our problem statement, the rationale behind Chatty Kathy, synthetic data and persona creation, model performance metrics, a visual demonstration of the project, and potential future developments. Join us for an insightful Q&A session to explore the potential of this groundbreaking project.
Project Team: Jay Requarth, Jana Avery, John Andrews, Dr. Dick Davis II, Nee Buntoum, Nam Yeongjin & Mat Nicholas
4. What is Pooling/Subsampling for?
Obj : Transform the "joint feature representation” into a new, more usable
Const
- discarding irrelevant detail
=> Invariance to changes in position/lighting, Robust to clutter, Compactness
of representation...
A Theoretical Analysis of Feature Pooling in Visual Recognition (2010)
Pooling Operations
5. This is the point
Const
- discarding irrelevant detail
Pros
Increasing Receptive Field
Decrease Model Size
Remove Noise
Translation Invariance
Cons
Lose Details
Lack (or NOT) of Equivariance
Pooling Operations
6. for TLDR;
StaticLearnable
Universal
Particular
Max Pooling
Avg Pooling
Fractional Pooling
LP Pooling
Wavelet Pooling
Softmax Pooling
Stochastic Pooling
Blur Pooling
Orderable Pooling
Global Average Pooling
Strided Convolution
Mixed, Gated, Tree Pooling
Local Importance Pooling
Detail-Preserving Pooling
Sensitive
Insensitive
Pooling Operations
Global Feature Guided Local Pooling
Spatial Pyramid Pooling
7. for TLDR;
Sensitive
Insensitive
Pooling Operations
Response Sensitivity
Someone, Orderable/Ranking/Attention
Max Pooling
Global Max Pooling
Softmax Pooling
Orderable Pooling
Mixed, Gated, Tree Pooling
Local Importance Pooling
Detail-Preserving Pooling
Spatial Pyramid Pooling
Fractional Pooling
Stochastic Pooling
Strided Convolution
Avg Pooling
Wavelet Pooling
LP Pooling
Global Feature Guided Local Pooling
Blur Pooling
8. 시작하기 전 TMI
CNN - Cat neuron response experiment (1962)
Average Pooling - Lenet (1998) : Subsampling
Max Pooling - Monkey neuron response exp (1999) -> DBN (2006)
-> Hand gesture recognition with cnn (2010)
Pooling Operations
9. Main Pooling Papers
Average Pooling
Lenet’s Subsampling [2]
Max Pooling
Similar to Neuron’s (at least monkey) response than avg [1]
LP Pooling
Average Pool - Max Pool. With hyperparameter, we can get avg/weighted/max pooling
Spatial Pyramid Pooling
Overcome CNN’s fixed output (GAP), Diversiform Receptive Field
Strided Convolution
Trainable Pooling
Pooling Operations
10. Average Pooling vs Max Pooling
Pooling Operations
Feature
Use all information with sum
> Pros
Backpropagation to all response
> Cons
Not robust to noise
Feature
Use highest value
> Pros
Robust to Noise
> Cons
Details are removed
vs
Suitable for Sparse Information
- Image Classification
Suitable for Dense Information
- NLP [below 1]
Avg / Response Insensitive Max / Response Sensitive
Start with 90s
11. Average Pooling / Max Pooling Balancing
- Transformable 2012
Pooling Operations
Convolutional Neural Networks Applied to House Numbers Digit Classification (2012)
LP Pool
by hyperparameter ‘p’, LP Pool is controllable from Avg(p=1) to Max(p=inf)
- Exp on Lenet Base
*SS: Single Stage, MS : Multi stage (Depth)
12. Average Pooling / Max Pooling Balancing
- Transformable 2017
Pooling Operations
DCASE 2017 SUBMISSION: MULTIPLE INSTANCE LEARNING FOR SOUND EVENT DETECTION (2017)
adaptive pooling operators for weakly labeled sound event detection (2018)
Softmax Pool
( Auto Pool )
by hyperparameter ‘a’, controllable from Avg(a=0), Softmax(a=1) to
Max(a=inf)
* RAP : Restricted Auto Pool (add regularizer for moving near 0), CAP : Contrainted Auto Pool ( value const for moving near 0 )
* Strong - time-varying label / Weak - No time labeled
- CNN based voice-spectrogram cls
13. Average Pooling / Max Pooling Balancing
- Hybrid 2015
Pooling Operations
Generalizing Pooling Functions in Convolutional Neural Networks:Mixed, Gated, and Tree (2015)
Mix/Gate Pool
- Maybe Exp on (pool 2) Alexnet Base
* Baseline : Max pool
+ Better Result & More Robust on transition
‘a’ parameter mix avg/max response
14. Average Pooling / Max Pooling Balancing
- Transformable + Hybrid 2019
Pooling Operations
Global Feature Guided Local Pooling (2019)
Global Feature Guided Local Pooling
- Softmax Pooling인데, GAP-FC로 parameter 데이터 마다 조정!
- More Depth, Pooling activated AVG/MAX
*trainable
`lambda` : 0, ‘rho’ :0 -> Avg
`lambda` : inf, ‘rho’ :0 -> Max
`lambda` : -inf, ‘rho’ :0 -> Min
Response Position
- Pooling Activation is Different by Cls (Imagenet)
15. (Max Pooling) Only Response Sensitive!
- Use Nth Max Responses 2016 - 2019
Pooling Operations
Rank-based pooling for deep convolutional neural networks (2016)
Ordinal Pooling (2019)
Ordinal Pooling
Weighting by Response Value
Notable thing
- At First Pooling, Closed to Avg. At Last, Close to Max
* baseline :Modified Lenet Avg/GAP - MNIST
16. Detail Please
- Detail reinforce 2018
Pooling Operations
Rapid, Detail-Preserving Image Downscaling (2016)
Detail-Preserving Pooling in Deep Networks (2018)
18. Detail Please
- Detail reinforce 2019
Pooling Operations
LIP: Local Importance-based Pooling (2019)
Local Importance Pooling
Calculate Pixel Importance
(Gate)
Reinforce
Response by Gate
LIP-AVG-Strided
CAM - 코알라
19. Hmm… Multiple Pooling
- Diversity Bags
Pooling Operations
Beyond Bags of Features: Spatial Pyramid Matching for Recognizing Natural Scene Categories (2006)
Spatial Pyramid Pooling in Deep ConvolutionalNetworks for Visual Recognition (2014)
Spatial Pyramid Pooling
GAP
20. Training, Training, Training
- NONO Pooling! 2014
Pooling Operations
Striving for simplicity, the all convolutional net (2014)
Strided Convolution
* BUT! FishNet reports
Strided Conv is worse
than Max Pool
* Stride Convs can’t
approximate MaxPool
(Ordinal)
21. Jackpot plz
- Stochastic 2013
Pooling Operations
Stochastic Pooling for Regularization of Deep Convolutional Neural Networks (2013)
Stochastic Pooling
- Train time : Stochastically Pool
- Test time : Weight Pool
22. Why only Integer division?
- Rebellion of root 2015
Pooling Operations
Fractional Max Pooling 2015
Fractional Max-Pooling
- Random Kernel Size for Max pooling -
for achieving root size downsample
* In = 25, Out =18
(root 2) = 1.414 / 25/18 = 1.388
23. Pooling has a WEAKNESS
- Avoiding Kryptonium 2019
Pooling Operations
Making Convolutional Networks Shift-Invariant Again 2019
https://richzhang.github.io/antialiased-cnns/
* Data Aug was removed
Blur Pool
24. Pooling has a WEAKNESS
- Failed KING 2017
Pooling Operations
CapsuleNet 2017
* Notable, “Evaluation of Pooling Operations in Convolutional Architectures for Object Recognition (ICANN 2010)
- Exp shows overlapping drops accuracy
* By Contrast, “AlexNet (2012)” Report overlapping is helpful for prevent overfitting
25. Family Tree
Max Pooling Avg Pooling
Spatial Pyramid Pool
- 좀 더 다양하게 볼래
LP/Softmax Pool
- 왔다 갔다할 수 있음!
족보 없음
Stochastic Pooling
- 아무나 고를래!
Wavelet Pooling
- Edge는 무조건 중요!
Ordinal Pool
- Flex 여러개 봄
Mixed Pool
- 동시에 둘다 볼래
Strided Convolution
- 요즘은 러닝 시대
Rank Pool
- 여러개 볼래
Blur Pool
- Blurring하면 더 좋아짐!
Fractional Pooling
- 커널 크기 랜덤할래
Global Pool
- 다 묻고 하나만 가!
Detail Preserve Pool
- Detail 어디감
Local Importance Pool
- More Trainable
Robust Attention Pool
- GMP도 Trainable 하게
Global Feature Guided Local Pool
- Input에 따라 avg/max 조절할래
( Trainable LP )
26.
27. Evaluation of Pooling Operations in Convolutional Architectures for Object
Recognition (2010)
- Overlapping pool drops accuracy
A Theoretical Analysis of Feature Pooling in Visual Recognition (2010)
- Proof if Response is sparse, max pooling is better than avg pooling
Pooling is neither necessary nor sufficient for appropriate deformation stability
in CNNS (2018)
Ask the locals: multi-way local pooling for image recognition (2011)
Signal recovery from Pooling Representations (2014)
stats385
Emergence of Invariance and Disentanglement in Deep Representations
(2018)
Quantifying Translation-Invariance in Convolutional Neural Networks (2017)
A Mathematical Theory of Deep Convolutional Neural Networks for Feature
Extraction (2017)
Learning to Linearize Under Uncertainty (2015)
Pooling Analysis