Outlier detection using machine learning, deep learning as well as statistical analysis.
The slide includes time series analysis. Also included is the hands on exercises with code and data, for a 3-day course.
NBITS is a best data science training institute in Hyderabad. It can provide data science course by real time experts. It can conduct real time projects and also provides job assistance in python and other courses like block chain, Mean stack, python, Hadoop, Sales force, sap.
NBITS is a best data science training institute in Hyderabad. It can provide data science course by real time experts. It can conduct real time projects and also provides job assistance in python and other courses like block chain, Mean stack, python, Hadoop, Sales force, sap.
How Machine Learning Helps Organizations to Work More Efficiently?Tuan Yang
Data is increasing day by day and so is the cost of data storage and handling. However, by understanding the concepts of machine learning one can easily handle the excessive data and can process it in an affordable manner.
The process includes making models by using several kinds of algorithms. If the model is created precisely for certain task, then the organizations have a very wide chance of making use of profitable opportunities and avoiding the risks lurking behind the scenes.
Learn more about:
» Understanding Machine Learning Objectives.
» Data dimensions in Machine Learning.
» Fundamentals of Algorithms and Mapping from Input/Output.
» Parametric and Non-parametric Machine Learning Algorithms.
» Supervised, Unsupervised and Semi-Supervised Learning.
» Estimating Over-fitting and Under-fitting.
» Use Cases.
An introduction to variable and feature selectionMarco Meoni
Presentation of a great paper from Isabelle Guyon (Clopinet) and André Elisseeff (Max Planck Institute) back in 2003, which outlines the main techniques for feature selection and model validation in machine learning systems
In Part II of the Anomaly Detection Series, we discuss the challenges in analyzing Temporal datasets and discuss methods for outlier analysis. We focus on single time series and discuss point outlier and sub-sequence methods.
Supervised learning is a machine learning paradigm where the algorithm is trained on a labeled dataset, learning patterns and relationships between input features and corresponding output labels to make accurate predictions on new, unseen data. It involves a teacher-supervisor relationship, where the algorithm strives to minimize the error between its predictions and the actual outcomes during training.
Data Analysis: Statistical Methods: Regression modelling, Multivariate Analysis - Classification: SVM & Kernel Methods - Rule Mining - Cluster Analysis, Types of Data in Cluster Analysis, Partitioning Methods, Hierarchical Methods, Density Based Methods, Grid Based Methods, Model Based Clustering Methods, Clustering High Dimensional Data - Predictive Analytics – Data analysis using R.
bioinformatics using statistical learning, machine learning and deep learning.
Day 2 and 3 materials from 12 days course, focusing on statistical analysis.
Meta analysis for medical data handling is include.
How Machine Learning Helps Organizations to Work More Efficiently?Tuan Yang
Data is increasing day by day and so is the cost of data storage and handling. However, by understanding the concepts of machine learning one can easily handle the excessive data and can process it in an affordable manner.
The process includes making models by using several kinds of algorithms. If the model is created precisely for certain task, then the organizations have a very wide chance of making use of profitable opportunities and avoiding the risks lurking behind the scenes.
Learn more about:
» Understanding Machine Learning Objectives.
» Data dimensions in Machine Learning.
» Fundamentals of Algorithms and Mapping from Input/Output.
» Parametric and Non-parametric Machine Learning Algorithms.
» Supervised, Unsupervised and Semi-Supervised Learning.
» Estimating Over-fitting and Under-fitting.
» Use Cases.
An introduction to variable and feature selectionMarco Meoni
Presentation of a great paper from Isabelle Guyon (Clopinet) and André Elisseeff (Max Planck Institute) back in 2003, which outlines the main techniques for feature selection and model validation in machine learning systems
In Part II of the Anomaly Detection Series, we discuss the challenges in analyzing Temporal datasets and discuss methods for outlier analysis. We focus on single time series and discuss point outlier and sub-sequence methods.
Supervised learning is a machine learning paradigm where the algorithm is trained on a labeled dataset, learning patterns and relationships between input features and corresponding output labels to make accurate predictions on new, unseen data. It involves a teacher-supervisor relationship, where the algorithm strives to minimize the error between its predictions and the actual outcomes during training.
Data Analysis: Statistical Methods: Regression modelling, Multivariate Analysis - Classification: SVM & Kernel Methods - Rule Mining - Cluster Analysis, Types of Data in Cluster Analysis, Partitioning Methods, Hierarchical Methods, Density Based Methods, Grid Based Methods, Model Based Clustering Methods, Clustering High Dimensional Data - Predictive Analytics – Data analysis using R.
bioinformatics using statistical learning, machine learning and deep learning.
Day 2 and 3 materials from 12 days course, focusing on statistical analysis.
Meta analysis for medical data handling is include.
Arduino, Raspberry Pi, Beagleblack and so on, all are signaling new tide of open source hardware.
In other words, open source is widening from software into hardware.
It will also affect the IOT, Internet of Things, as the major IOT frameworks are also open source based.
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on the notifications, alerts, and approval requests using Slack for Bonterra Impact Management. The solutions covered in this webinar can also be deployed for Microsoft Teams.
Interested in deploying notification automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
Connector Corner: Automate dynamic content and events by pushing a buttonDianaGray10
Here is something new! In our next Connector Corner webinar, we will demonstrate how you can use a single workflow to:
Create a campaign using Mailchimp with merge tags/fields
Send an interactive Slack channel message (using buttons)
Have the message received by managers and peers along with a test email for review
But there’s more:
In a second workflow supporting the same use case, you’ll see:
Your campaign sent to target colleagues for approval
If the “Approve” button is clicked, a Jira/Zendesk ticket is created for the marketing design team
But—if the “Reject” button is pushed, colleagues will be alerted via Slack message
Join us to learn more about this new, human-in-the-loop capability, brought to you by Integration Service connectors.
And...
Speakers:
Akshay Agnihotri, Product Manager
Charlie Greenberg, Host
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
"Impact of front-end architecture on development cost", Viktor TurskyiFwdays
I have heard many times that architecture is not important for the front-end. Also, many times I have seen how developers implement features on the front-end just following the standard rules for a framework and think that this is enough to successfully launch the project, and then the project fails. How to prevent this and what approach to choose? I have launched dozens of complex projects and during the talk we will analyze which approaches have worked for me and which have not.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
7. • 추정
• 정규분포
• t-분포
• 가설검정
• p-Value를 이용한 가설검정
• p-value = 관측된 유의수준 (level of significance)
• defines the smallest value of 𝛼 for which the H0 can be rejected.
• “α 가 p보다 커야만 H0를 reject 가능”
12. • 주요 이슈
• Subjectivity, Interestingness and noise
• Subjective judgement, as to what constitutes a “sufficient” deviation
• In real applications, the data may be embedded in a significant amount of noise
• Noise 문제
• Representation의 문제
• Normality vs. Anomaly
• 특성공학
• 이상치 분석과 데이터 모델
13. • 알고리즘 분류 (1)
• Outlier scores: quantify level of “outlierness”
• Binary labels: “outlier? or not?”
• the threshold is based on the statistical distribution of the scores.
• 알고리즘 분류 (2)
14.
15. • qualitative techniques,
• 전문가 의견
• time series analysis
• 관심사항: patterns and pattern changes
• 과거 데이터가 중요한 역할
• causal models.
• 인과관계
• 과거 데이터가 중요한 역할
• Regression
• 독립변수와 종속변수
16. 단변량 Time Series 모델
• AR Model
• 자기상관 (Autocorrelation)
• 정상성 (Stationarity)과 ADF Test
• Differencing a Time Series
• Autocorrelation에서의 Lags
• Partial Autocorrelation
• AR 모델 정의
• Yule-Walker Equation 이용한 AR 추정
• MA Model
• MA 모델 정의
• MA Model의 Fitting
• Stationarity
• AR vs. MA 모델 선택
• Model Retraining을 통한 다단계 예측
• 최적의 MA Order를 찾기 위한 Grid
Search
17. • ARMA 모델
• 모델 정의
• ARMA(1,1) Model의 Fitting
• Automated Hyperparameter Tuning
• Grid Search
• 성능향상을 위한 Tuning
• ARIMA 모델
• 모델 정의
• SARIMA 모델
• 모델 정의
• regular AR part φp
• seasonal AR part
• regular MA part θq
• seasonal MA part
• regular integration part; order d
• seasonal integration part; order D
• Coefficient of seasonality s
18. • Multivariate Time Series Models
• SARIMAX 모델
• SARIMA 모델에 외생변수 (X)를 추가
• VAR 모델
• Since VAR model proposes one model for multiple target variables, it regroups those variables
as a vector
• VAR 계수의 추정
• VARMAX 모델
• VAR model에 MV 항을 추가하고 외생변수 허용
• V for vector indicating that it’s a multivariate model
• AR for autoregression
• MA for moving average
• X for the use of exogenous variables (in addition to the endogenous variables)
20. • RNN/LSTM
• Predicting a Sequence Rather Than a Value
• SimpleRNN
• GRU, LSTM
• Prophet 모델 (Facebook)
• an automated procedure for building forecasting models developed by Facebook.
• Input possibilities are
• Seasonality of any regular order
• Holidays
• Additional regressors
• hyperparameters
• Fourier order of the seasonality: A higher order means more flexibility.
• changepoint_prior_scale plays on the trend: The higher the value, the more flexible the trend.
• holidays_prior_scale: The lower it is, the less important the holidays are for the model.
• prior scale for the seasonality
• DeepAR 모델
23. • Distance Distribution-based Techniques
• to model entire data set to be normally distributed about its mean in the form of a multivariate
Gaussian distribution.
• Let ҧ
𝜇 be d-dimensional (row) data set, and Σ be its d x d covariance matrix.
• Then, the probability distribution 𝑓( ത
𝑋) for a d-dimensional (row vector) data point X is:
• |Σ| = determinant of covariance matrix.
• 지수부: (half) squared Mahalanobis distance of the data point X to the centroid μ of the data.
= outlier score
24.
25. • Extreme-Value Analysis
• 극값을 판별해 내는 것
• = Probabilistic Tail Inequalities
• Markov Inequality
• Chebychev Inequality
• …
• determine statistical tails of the underlying distribution.
• Univariate
• Box Plots
• 다변량 데이터에서의 극값분석
• Depth-Based Methods – Convex hull 분석
• Deviation-Based Methods
• Angle-based
• Extreme-value analysis is usually required as a final step on these modeled deviations
26. • 개요
• 확률모형에서는 “likelihood fit of a data point to a generative model is the outlier score”.
• 예
• GMM
• EM
• 장단점
• 장점
• 다양한 경우에 적용 가능 (any data type or mixed data type), as long as an appropriate
generative model is available for each mixture component.
• 단점
• 분포를 특정하기 어려운 경우.
• As the number of model parameters increases, over-fitting becomes more common.
27. • 일반형
• a convex non-linear programming – OLS
• Model the data along lower-dimensional subspaces using linear correlations
• Hyperplane과 데이터와의 거리 → outlier scores.
• PCA
• 행렬분해
• Spectral Models
• Some variations of matrix decomposition (ex: PCA) used in certain types of data such as
graphs and networks, are called spectral models.
• They are used commonly for clustering graph data, and are often used in order to identify
anomalous changes in temporal sequences of graphs.
29. • 개념
• outliers increase the minimum code length (i.e., minimum length of the summary) required to
describe a data set as they represent deviations from natural attempts to summarize data.
• 예(1)
• 예(2) multidimensional data sets
• 확률모델: a data set in terms of generative model parameters, such as a mixture of
Gaussian distributions or a mixture of exponential power distributions.
• 군집화 / 밀도기반 요약 : describes a data set in terms of cluster descriptions,
histograms, or other summarized representations, along with maximum error tolerances.
• PCA / spectral 모델: describes the data in terms of lower dimensional subspaces of
projection of multi-dimensional data or a latent representation of a network.
• FP mining : describes the data in terms of an underlying code book of frequent patterns.
30. • High-dimension
• Subspace outlier detection
• Assumption: “outliers are often hidden in the unusual local behavior of low-dimensional
subspaces, and this deviant behavior is masked by full-dimensional analysis”.
• High-dimensional space에서 데이터는 sparse 및 almost equidistant.
• → outlier scores become less distinguishable.
• Outliers are best emphasized in a lower-dimensional local subspace of relevant attributes.
31.
32. • Max Voting
• 주로 classification 에 적용.
• 다수 모델로 각각의 데이터를 예측 – 이를 ‘vote’로 처리.
• 예: 영화에 대한 평점
• 기법
• Averaging과 Weighted Averaging
• Stacking
• Blending
• Bagging 및 Boosting
33. • Outlier 분석 ensemble 의 2 종류 :
• sequential ensembles
• a given algorithm or set of algorithms are applied sequentially, so that future applications of the
algorithms are influenced by previous applications, in terms of either modifications of the base
data for analysis or in terms of the specific choices of the algorithms.
• 최종 산출물: either a weighted combination of, or the final result of the last application.
(예) 분류모델에서 boosting methods may be considered examples of sequential ensembles.
• independent ensembles
• different algorithms, or different instantiations of the same algorithm are applied to either the
complete data or portions of the data. The choices made about the data and algorithms applied
are independent of the results obtained from these different algorithmic executions.
• 최종산출물: executions are combined together in order to obtain more robust outliers.
34.
35. • 범주형 데이터, 텍스트 및 Mixed Attributes
• categorical attributes that take on discrete unordered values.
• Mixed attribute data contain both numerical and categorical attributes.
• Regression-based models can be used in a limited way over discrete attribute values,
• 대책
• convert the discrete data to binary data by creating one attribute for each categorical value.
Such methods can be more easily extended to text
• 모델 적용
• LSA (latent semantic analysis)
• Clustering
• proximity-based methods
• probabilistic models
• frequent pattern mining
• 데이터 내에서의 Dependency 문제
• 시계열 데이터
• Discrete Sequence 데이터
• 그래프, 네트워크 형, …
40. • Feature (특성)
• A feature is a numeric representation of raw data.
• Simple Numbers
• Scalars, vectors, spaces
• Counts
• Binarization. Quantization or binning
• Feature Scaling (Normalization)
• Min-max scaling
• Standardization (variance scaling)
• Feature Selection
Bucketing Crossing Hashing Embedding
41. • Log 변환
• 텍스트 데이터
• Flat Vectors
• Bag-of-words, Bag-of-N-Grams
• Filtering
• Stopwords, Frequency-based filtering, Stemming
• Semantic기법
• Parsing, tokenization, Phrase Detection, TF-IDF
• 범주형 변수 - Encoding
• One-hot encoding, Dummy coding
• 차원축소와 행렬분해
• PCA, SVD
• 모델 적용
• LSA (latent semantic analysis), Clustering, 확률모형
• Data Value에서의 Dependency 문제
42.
43. • kNN
• KNN graph (k-nearest neighbor graph)?
• a graph in which 2 vertices p and q are connected by an edge, if the distance between p and q is
among the k-th smallest distances from p to other objects from P.
• has a vertex for each point, and a directed edge from p to q whenever q is a nearest neighbor of p, a
point whose distance from p is minimum among all the given points other than p itself.
• (변형 1) 1-NNG
• Directions of the edges are ignored and NNG is defined instead as an undirected graph.
However, the nearest neighbor relation is not a symmetric one.
• (변형 2) FNG (farthest neighbor graph)
44. • Outlier Detection using In-degree Number (ODIN)
• 각 data point의 in-degree를 계산
• in-degree = the number of nearest neighbour sets to this point belongs.
• In-degree값이 크면 ; more confidence of this point belonging to some dense region in the
space.
• In-degree값이 작으면 ; it’s not part of many nearest neighbour sets
• 즉, is kind of isolated in the space.
• the reverse of KNN.
45. • SVM
• 개념
• Linear SVM vs. Non-Linear SVM
• One-Class Classification
• 1. Outlier Detection
• 2. AD in Acoustic Signals
• 3. Novelty Detection and many others.
• One-class SVM (1)
• to ensure the widest street
• maximize 2/|w| == to minimizing 1/2*(|w|^2).
• + Lagrange multiplier →
• w is a vector of random weights.
• alpha = Lagrange multiplier,
• y = either +1 or -1 i.e., class of the sample,
• x = samples from data.
46.
47. 비지도학습 일반론
• 차원축소
• Linear Projection
• PCA, SVD
• Random projection
• Manifold Learning
• Isomap
• T-SNE
• Dictionary learning
• ICA, Latent Dirichlet Allocation
• 군집화
• K-Means
• Hierarchical Clustering
• DBSCAN
• 혼합모형/EM
• 딥러닝 기반 비지도학습
• Feature Extraction
• Autoencoders
• Unsupervised Pretraining
• 생성모델과 네트워크 모델
• RBM
• Deep Belief Networks
• GAN
• Sequential Data 적용
• Hidden Markov model
• 강화학습과 비지도학습
• Semi-supervised Learning
48. • 비지도학습
• 목적: interesting pattern과 숨겨진 데이터 속성을 찾는 것
• = 자율학습(unsupervised learning) (vs. 지도학습(supervised))
• Can we visualize data?
• Can we find meaningful subgroups of observations or variables?
• Challenges
• EDA - goal is not as clearly defined
• 객관적 성능측정이 쉽지 않다 - don’t know the “right answer”
• High-dimensional data
• 대표적 적용 예
• As a stand-alone tool to get insight into data distribution
• As a preprocessing step for other algorithms
49.
50. • 군집모형
• 군집 (cluster) = a subset of data which are similar.
51. • K-Means
• K-Means 기반 이상탐지
• we can define outliers by ourselves.
• define what is a ‘far’ distance
• define how many data points should be outliers.
• outlier/anomaly
• a data point far from the centroid of its cluster
52. • LOF (Local Outlier Factors)
• 개요
• identify an outlier considering the density of the neighborhood.
• 특히 데이터의 밀도 (density of the data)가 일정치 않을 때 효과가 큼
• = ratio of the average LRD of K neighbors of A to the LRD of A.
• LRD of each point is used to compare with the average LRD of its K neighbors.
• If the point is not an outlier (inlier), the ratio of average LRD of neighbors is approximately equal to the
LRD of a point. In that case, LOF is nearly equal to 1.
• If the point is an outlier, LRD of a point < average LRD of neighbors. Then LOF value will be high.
• If LOF> 1, is considered as an outlier, but not always true.
• 관련 개념
• Reachability distance (RD)
• Local reachability density (LRD)
• Local Outlier Factor (LOF)
53. • LOF ≈ 1 similar density as neighbors
• LOF < 1 higher density than neighbors (normal point)
• LOF > 1 lower density than neighbors (anomaly)
54. • K-distance와 K-Neighbors
• K-distance
• = distance between the point, and it’s Kth nearest neighbor.
• K-neighbors, Nₖ(A), includes a set of points that lie in or on the circle of radius K-distance.
• Reachability Distance (RD)
• = maximum of K-distance of Xj and the distance between Xi and Xj.
• Local RD (LRD)
K-distance of A with K=2
55.
56. • Mixture Model
• model the data in terms of a mixture of several components, where each component has
a simple parametric form (예: Gaussian).
• assuming class mixture component is known and estimating class membership given
parameters.
• Mixtures of {Sequences, Curves, …}
• 생성모형
• select a component ck for individual i
• generate data according to p(Di | ck)
• p(Di | ck) can be very general
• GMM (Gaussian Mixture Model)
• Multivariate Gaussian models
60. • Neuron과 Artificial Nodes
• 개별 신경망의 특징을 결정하는 요소:
• 활성함수
• step, sigmoid, tanh, relu
• Network topology (or architecture)
• 모델이 가진 뉴론의 수 + 연결된 layer의 수
• Training 알고리즘
• Gradient descent, Newton method, Conjugate gradient, …
• 학습 – BP through Gradient Descent
• Computation Graph
64. • Python기반 딥러닝 프레임워크
• TensorFlow와 Keras
• TensorFlow
• Keras 이용
• R interface
• PyTorch
• 기타 주요 라이브러리
65. • 개념
• Anomaly Detection (AD) OR novelty detection
• Normality Representation
• ☞ 기술통계
• Measures of Frequency
• Measures of Central Tendency
• Measures of Dispersion
• Anomaly representation
• Outlier detection 알고리즘에서의 2가지 출력 양식
• Outlier scores: quantify level of “outlier-ness” outlier tendency.
• Binary labels: “Whether a data point is an outlier or not”
• 주요 이슈
• Subjectivity, Interestingness and noise
• the data may be embedded in a significant amount of noise
66.
67. • RNN
• 개념
• RNN은 지금 들어온 입력데이터와 과거에 입력 받았던 데이터를 동시에 고려
• 장단점
• (장점) see how previous layer is stimulated
→ NN interprets sequences much better.
• (단점) more parameters to be calculated
A recurrent neuron (왼쪽) unrolled through time (오른쪽)
68. • Long short-term memory models
• 목적: 기존의 RNN의 문제점 해결:
• Vanishing gradients와 Exploding gradients
• inability to remember or forget certain aspects of input sequences
• 특징: previous time step으로부터 previous output뿐 아니라 state 정보도 함께 전달받음.
• 동작원리
• Output control: How much an output neuron is stimulated by the previous output and current state
• Memory control: How much of previous state will be forgotten
• Input control: How much of the previous output and new state (memory) will be considered to
determine the new current state
• These are trainable and optimized
69.
70. • 정의
• shallow, 2-layer NNs constituting DBN (deep-belief networks)
• Restriction = no intra-layer communication.
71. • Reconstruction
• activations of hidden layer no.1 become input in a backward pass.
• Forward pass – RBM to predict node activations: p(a | x; w).
• Backward pass - RBM attempts to estimate p(x | a; w).
• Reconstruction을 통해 입력데이터의 PDF를 추측 (= generative learning)
• 추정된 PDF와 실제 PDF의 거리계산 - Kullback Leibler Divergence.
• Kullback-Leibler (KL) divergence measures divergence of two probability distributions, p and q.
이를 통해 p(x, a)
72. • Autoencoder
• PCA와 유사하지만 보다 flexible.
• Target output is its input (x) in a different form (x’).
• dimensionality of input = dimensionality of the output
• essentially what we want is x’ = x. x’ = Decode (Encode(x))
• Autoencoder를 이용한 이상탐지
• If a point in feature space lies far away from the majority of the points (meaning it holds
different properties), the autoencoder learns the distribution - an anomaly.
• 즉, model more or less correctly re-generates the images leading to low loss values.
• We use these reconstruction loss values as the anomaly scores
• The higher the scores, the higher the chances of input being an anomaly.
73. • LSTM-based Encoder-Decoder for Anomaly Detection
• 정상데이터 (MV TSA)를 Unsupervised 방법으로 학습하고 이상치를 탐지하는 모델
• 특징
• LSTM-Encoder와 LSTM-Decoder로 구성
• Encoder는 다변량 데이터를 압축하여 feature로 변환.
• Decoder는 Encoder에서 받은 feature를 이용 Encoder에서 받은 다변량 데이터를 재구성
• Reconstruction Error 계산
• MSE Loss를 이용하여 학습 But 추론 과정에서 Error 계산 방법은 Absolute Error를 활용.
74. • Self-Organizing Maps
• 자기조직화 지도
• 특징
• Competitive learning by BP
• 일종의 DR 기법
• 동작원리
• Components
• Initialization
• Competition
• Cooperation
• Adaptation
https://arxiv.org/pdf/1312.5753.pdf
source; wikipedia
75.
76.
77.
78. • 사이버보안 이상징후 판단
• Malware 분석
• Network traffic 분석
• 센서 네트워크
79. • 모터 결함 진단 (Fault Diagnosis)
• = 고장진단 (현 상태가 고장인지 여부) + 고장 예측
• 예: 유도 전동기 (Induction motor)
• DBN (Deep Belief Network)
• A generative graphical model
• Stacking RBMs
• 진동 (주파수) 측정 데이터를 이용하여 학습
• → Contrastive divergence using
• Gibbs sampling
82. • Anomaly detection problem Complexities
• Unknownness
• Anomalies are associated with many unknowns, e.g., instances with unknown abrupt behaviors,
data structures, and distributions. They remain unknown until actually occur.
• Heterogeneous anomaly classes.
• Anomalies are irregular, and thus, one class of anomalies may demonstrate completely
different abnormal characteristics from another class of anomalies.
• Rarity and class imbalance
• → unavailability of large-scale labeled data in most applications = class imbalance
• Diverse types of anomaly.
• 3 different types of anomaly have been explored.
• Point anomalies
• Conditional anomalies
• Group anomalies
• 3 different types of anomaly have been explored.
• Point anomalies
• Conditional anomalies = contextual anomalies
• Group anomalies = collective anomalies
83. • Deep Anomaly Detection가 해결 시도하는 문제
• CH1: Low anomaly detection recall rate.
• CH2: Anomaly detection in high-dimensional and/or not-independent data.
• CH3: Data-efficient learning of normality/abnormality.
• CH4: Noise-resilient anomaly detection.
• CH5: Detection of complex anomalies.
• CH6: Anomaly explanation.