This document summarizes Ruiping Wang's tutorial on distance metric learning for visual recognition. It provides an overview of previous works on set-based visual recognition, including set modeling approaches such as linear subspace, affine/convex hull, and nonlinear manifold models. It also reviews previous metric learning methods for set matching, including learning in Euclidean space and on Riemannian manifolds. Specific approaches discussed include mutual subspace method (MSM), discriminant canonical correlations (DCC), Grassmann discriminant analysis (GDA), and projection metric learning (PML).
Image Enhancement: Introduction to Spatial Filters, Low Pass Filter and High Pass Filters. Here Discussed Image Smoothing and Image Sharping, Gaussian Filters
Computational Discovery of Two-Dimensional Materials, Evaluation of Force-Fie...KAMAL CHOUDHARY
JARVIS (Joint Automated Repository for Various Integrated Simulations) is a repository designed to automate materials discovery using classical force-field, density functional theory, machine learning calculations and experiments.
The Force-field section of JARVIS (JARVIS-FF) consists of thousands of automated LAMMPS based force-field calculations on DFT geometries. Some of the properties included in JARVIS-FF are energetics, elastic constants, surface energies, defect formations energies and phonon frequencies of materials.
The Density functional theory section of JARVIS (JARVIS-DFT) consists of thousands of VASP based calculations for 3D-bulk, single layer (2D), nanowire (1D) and molecular (0D) systems. Most of the calculations are carried out with optB88vDW functional. JARVIS-DFT includes materials data such as: energetics, diffraction pattern, radial distribution function, band-structure, density of states, carrier effective mass, temperature and carrier concentration dependent thermoelectric properties, elastic constants and gamma-point phonons.
The Machine-learning section of JARVIS (JARVIS-ML) consists of machine learning prediction tools, trained on JARVIS-DFT data. Some of the ML-predictions focus on energetics, heat of formation, GGA/METAGGA bandgaps, bulk and shear modulus. The ML webpage is visible to NIST employees only right now, but will be available publicly soon.
Introduction to Statistical Machine Learningmahutte
This course provides a broad introduction to the methods and practice
of statistical machine learning, which is concerned with the development
of algorithms and techniques that learn from observed data by
constructing stochastic models that can be used for making predictions
and decisions. Topics covered include Bayesian inference and maximum
likelihood modeling; regression, classi¯cation, density estimation,
clustering, principal component analysis; parametric, semi-parametric,
and non-parametric models; basis functions, neural networks, kernel
methods, and graphical models; deterministic and stochastic
optimization; over¯tting, regularization, and validation.
Image Enhancement: Introduction to Spatial Filters, Low Pass Filter and High Pass Filters. Here Discussed Image Smoothing and Image Sharping, Gaussian Filters
Computational Discovery of Two-Dimensional Materials, Evaluation of Force-Fie...KAMAL CHOUDHARY
JARVIS (Joint Automated Repository for Various Integrated Simulations) is a repository designed to automate materials discovery using classical force-field, density functional theory, machine learning calculations and experiments.
The Force-field section of JARVIS (JARVIS-FF) consists of thousands of automated LAMMPS based force-field calculations on DFT geometries. Some of the properties included in JARVIS-FF are energetics, elastic constants, surface energies, defect formations energies and phonon frequencies of materials.
The Density functional theory section of JARVIS (JARVIS-DFT) consists of thousands of VASP based calculations for 3D-bulk, single layer (2D), nanowire (1D) and molecular (0D) systems. Most of the calculations are carried out with optB88vDW functional. JARVIS-DFT includes materials data such as: energetics, diffraction pattern, radial distribution function, band-structure, density of states, carrier effective mass, temperature and carrier concentration dependent thermoelectric properties, elastic constants and gamma-point phonons.
The Machine-learning section of JARVIS (JARVIS-ML) consists of machine learning prediction tools, trained on JARVIS-DFT data. Some of the ML-predictions focus on energetics, heat of formation, GGA/METAGGA bandgaps, bulk and shear modulus. The ML webpage is visible to NIST employees only right now, but will be available publicly soon.
Introduction to Statistical Machine Learningmahutte
This course provides a broad introduction to the methods and practice
of statistical machine learning, which is concerned with the development
of algorithms and techniques that learn from observed data by
constructing stochastic models that can be used for making predictions
and decisions. Topics covered include Bayesian inference and maximum
likelihood modeling; regression, classi¯cation, density estimation,
clustering, principal component analysis; parametric, semi-parametric,
and non-parametric models; basis functions, neural networks, kernel
methods, and graphical models; deterministic and stochastic
optimization; over¯tting, regularization, and validation.
Introduction to Topological Data AnalysisMason Porter
Here are slides for my 3/14/21 talk on an introduction to topological data analysis.
This is the first talk in our Short Course on topological data analysis at the 2021 American Physical Society (APS) March Meeting: https://march.aps.org/program/dsoft/gsnp-short-course-introduction-to-topological-data-analysis/
Presentation in Vietnam Japan AI Community in 2019-05-26.
The presentation summarizes what I've learned about Regularization in Deep Learning.
Disclaimer: The presentation is given in a community event, so it wasn't thoroughly reviewed or revised.
CS404 Pattern Recognition - Locality Preserving ProjectionsJishnu P
These slides are just providing an overview of locality preserving projections (LPP) which is a dimensionality reduction (DR) technique. DR techniques are very useful as they transform the data into a much more compact form while preserving the original form of the data intact (ideally). This helps in better utilization of memory and faster data processing. This was the study topic of my team during the Pattern Recognition (CS404) technical elective course at IIIT-Vadodara (B.Tech Sem-7).
발표자: 박태성 (UC Berkeley 박사과정)
발표일: 2017.6.
Taesung Park is a Ph.D. student at UC Berkeley in AI and computer vision, advised by Prof. Alexei Efros.
His research interest lies between computer vision and computational photography, such as generating realistic images or enhancing photo qualities. He received B.S. in mathematics and M.S. in computer science from Stanford University.
개요:
Image-to-image translation is a class of vision and graphics problems where the goal is to learn the mapping between an input image and an output image using a training set of aligned image pairs.
However, for many tasks, paired training data will not be available.
We present an approach for learning to translate an image from a source domain X to a target domain Y in the absence of paired examples.
Our goal is to learn a mapping G: X → Y such that the distribution of images from G(X) is indistinguishable from the distribution Y using an adversarial loss.
Because this mapping is highly under-constrained, we couple it with an inverse mapping F: Y → X and introduce a cycle consistency loss to push F(G(X)) ≈ X (and vice versa).
Qualitative results are presented on several tasks where paired training data does not exist, including collection style transfer, object transfiguration, season transfer, photo enhancement, etc.
Quantitative comparisons against several prior methods demonstrate the superiority of our approach.
Scikit-Learn is a powerful machine learning library implemented in Python with numeric and scientific computing powerhouses Numpy, Scipy, and matplotlib for extremely fast analysis of small to medium sized data sets. It is open source, commercially usable and contains many modern machine learning algorithms for classification, regression, clustering, feature extraction, and optimization. For this reason Scikit-Learn is often the first tool in a Data Scientists toolkit for machine learning of incoming data sets.
The purpose of this one day course is to serve as an introduction to Machine Learning with Scikit-Learn. We will explore several clustering, classification, and regression algorithms for a variety of machine learning tasks and learn how to implement these tasks with our data using Scikit-Learn and Python. In particular, we will structure our machine learning models as though we were producing a data product, an actionable model that can be used in larger programs or algorithms; rather than as simply a research or investigation methodology.
Cluster analysis is a major tool in a number of applications in many fields of Business, Engineering & etc.(The odoridis and Koutroubas, 1999):
Data reduction.
Hypothesis generation.
Hypothesis testing.
Prediction based on groups.
Divide the examined window into cells (e.g. 16x16 pixels for each cell).
2- For each pixel in a cell, compare the pixel to each of its 8 neighbors (on its left-top, leftmiddle,
left-bottom, right-top, etc.). Follow the pixels along a circle, i.e. clockwise or counterclockwise.
3- Where the center pixel's value is greater than the neighbor's value, write "1". Otherwise,
write "0". This gives an 8-digit binary number (which is usually converted to decimal for
convenience).
4- Compute the histogram, over the cell, of the frequency of each "number" occurring (i.e.,
each combination of which pixels are smaller and which are greater than the center).
Introduction to Topological Data AnalysisMason Porter
Here are slides for my 3/14/21 talk on an introduction to topological data analysis.
This is the first talk in our Short Course on topological data analysis at the 2021 American Physical Society (APS) March Meeting: https://march.aps.org/program/dsoft/gsnp-short-course-introduction-to-topological-data-analysis/
Presentation in Vietnam Japan AI Community in 2019-05-26.
The presentation summarizes what I've learned about Regularization in Deep Learning.
Disclaimer: The presentation is given in a community event, so it wasn't thoroughly reviewed or revised.
CS404 Pattern Recognition - Locality Preserving ProjectionsJishnu P
These slides are just providing an overview of locality preserving projections (LPP) which is a dimensionality reduction (DR) technique. DR techniques are very useful as they transform the data into a much more compact form while preserving the original form of the data intact (ideally). This helps in better utilization of memory and faster data processing. This was the study topic of my team during the Pattern Recognition (CS404) technical elective course at IIIT-Vadodara (B.Tech Sem-7).
발표자: 박태성 (UC Berkeley 박사과정)
발표일: 2017.6.
Taesung Park is a Ph.D. student at UC Berkeley in AI and computer vision, advised by Prof. Alexei Efros.
His research interest lies between computer vision and computational photography, such as generating realistic images or enhancing photo qualities. He received B.S. in mathematics and M.S. in computer science from Stanford University.
개요:
Image-to-image translation is a class of vision and graphics problems where the goal is to learn the mapping between an input image and an output image using a training set of aligned image pairs.
However, for many tasks, paired training data will not be available.
We present an approach for learning to translate an image from a source domain X to a target domain Y in the absence of paired examples.
Our goal is to learn a mapping G: X → Y such that the distribution of images from G(X) is indistinguishable from the distribution Y using an adversarial loss.
Because this mapping is highly under-constrained, we couple it with an inverse mapping F: Y → X and introduce a cycle consistency loss to push F(G(X)) ≈ X (and vice versa).
Qualitative results are presented on several tasks where paired training data does not exist, including collection style transfer, object transfiguration, season transfer, photo enhancement, etc.
Quantitative comparisons against several prior methods demonstrate the superiority of our approach.
Scikit-Learn is a powerful machine learning library implemented in Python with numeric and scientific computing powerhouses Numpy, Scipy, and matplotlib for extremely fast analysis of small to medium sized data sets. It is open source, commercially usable and contains many modern machine learning algorithms for classification, regression, clustering, feature extraction, and optimization. For this reason Scikit-Learn is often the first tool in a Data Scientists toolkit for machine learning of incoming data sets.
The purpose of this one day course is to serve as an introduction to Machine Learning with Scikit-Learn. We will explore several clustering, classification, and regression algorithms for a variety of machine learning tasks and learn how to implement these tasks with our data using Scikit-Learn and Python. In particular, we will structure our machine learning models as though we were producing a data product, an actionable model that can be used in larger programs or algorithms; rather than as simply a research or investigation methodology.
Cluster analysis is a major tool in a number of applications in many fields of Business, Engineering & etc.(The odoridis and Koutroubas, 1999):
Data reduction.
Hypothesis generation.
Hypothesis testing.
Prediction based on groups.
Divide the examined window into cells (e.g. 16x16 pixels for each cell).
2- For each pixel in a cell, compare the pixel to each of its 8 neighbors (on its left-top, leftmiddle,
left-bottom, right-top, etc.). Follow the pixels along a circle, i.e. clockwise or counterclockwise.
3- Where the center pixel's value is greater than the neighbor's value, write "1". Otherwise,
write "0". This gives an 8-digit binary number (which is usually converted to decimal for
convenience).
4- Compute the histogram, over the cell, of the frequency of each "number" occurring (i.e.,
each combination of which pixels are smaller and which are greater than the center).
Slides by Amaia Salvador at the UPC Computer Vision Reading Group.
Source document on GDocs with clickable links:
https://docs.google.com/presentation/d/1jDTyKTNfZBfMl8OHANZJaYxsXTqGCHMVeMeBe5o1EL0/edit?usp=sharing
Based on the original work:
Ren, Shaoqing, Kaiming He, Ross Girshick, and Jian Sun. "Faster R-CNN: Towards real-time object detection with region proposal networks." In Advances in Neural Information Processing Systems, pp. 91-99. 2015.
Change Detection of Water-Body in Synthetic Aperture Radar ImagesCSCJournals
Change detection is the art of quantifying the changes in the Synthetic Aperture Radar (SAR) images that have happened over a period of time. Remote sensing has been the parental technique to perform change detection analysis. This paper empirically investigates the impact of applying the combination of texture features for different classification techniques to separate water body from non-water body. At first, the images are classified using unsupervised Principle Component Analysis (PCA) based K-means clustering for dimension reduction. Then the texture features like Energy, Entropy, Contrast , Inverse Differential Moment , Directional Moment and the Median are extracted using Gray Level Co-occurrence Matrix (GLCM) and these features are utilized in Linear Vector Quantization (LVQ) and Support Vector Machine (SVM) classifiers. This paper aims to apply a combination of the texture features in order to significantly improve the accuracy of detection. The utility of detection analysis, influences management and policy decision making for long-term construction projects by predicting the preventable losses.
Automatic Skin Lesion Segmentation and Melanoma Detection: Transfer Learning ...Zabir Al Nazi Nabil
Industrial pollution resulting in ozone layer depletion has influenced
increased UV radiation in recent years which is a major environmental risk factor for invasive skin cancer Melanoma and other keratinocyte cancers. The incidence of deaths from Melanoma has risen worldwide in past two decades.
Deep learning has been employed successfully for dermatologic diagnosis. In
this work, we present a deep learning based scheme to automatically segment
skin lesions and detect melanoma from dermoscopy images. U-Net was used
for segmenting out the lesion from surrounding skin. The limitation of utilizing
deep neural networks with limited medical data was solved with data augmentation and transfer learning. In our experiments, U-Net was used with spatial
dropout to solve the problem of overfitting and different augmentation effects
were applied on the training images to increase data samples. The model was
evaluated on two different datasets. It achieved a mean dice score of 0.87 and a
mean jaccard index of 0.80 on ISIC 2018 dataset. The trained model was assessed on PH² dataset where it achieved a mean dice score of 0.93 and a mean
jaccard index of 0.87 with transfer learning. For classification of malignant
melanoma, a DCNN-SVM model was used where we compared state of the art
deep nets as feature extractors to find the applicability of transfer learning in
dermatologic diagnosis domain. Our best model achieved a mean accuracy of
92% on PH² dataset. The findings of this study is expected to be useful in cancer diagnosis research.
Published at IJCCI 2018. Source code available at https://github.com/zabir-nabil/lesion-segmentation-melanoma-tl
Linearity of Feature Extraction Techniques for Medical Images by using Scale ...ijtsrd
In Machine Learning, Pattern Recognition and in the field of image processing, Feature Extraction starts from an initial set of the measured data. Builds derived values are intended to be informative and non redundant, facilating the subsequent learning and in some cases leading to the better human interpretations. Feature Extraction is a dimensionally reduction process, where an initial set of raw variables has been reduced to more manageable groups. Many data analysis software packages provide for feature extraction and for dimension reduction. Determining a subset of the initial features is also known as feature extraction. Common Numerical programming environments are MATLAB, SciLab, NumPy, etc. Ramar S | Keerthiswaran V | Karthik Raj S S "Linearity of Feature Extraction Techniques for Medical Images by using Scale Invariant Feature Transform" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-4 | Issue-3 , April 2020, URL: https://www.ijtsrd.com/papers/ijtsrd30358.pdf Paper Url :https://www.ijtsrd.com/engineering/bio-mechanicaland-biomedical-engineering/30358/linearity-of-feature-extraction-techniques-for-medical-images-by-using-scale-invariant-feature-transform/ramar-s
National Flags Recognition Based on Principal Component Analysisijtsrd
Recognizing an unknown flag in a scene is challenging due to the diversity of the data and to the complexity of the identification process. And flags are associated with geographical regions, countries and nations. But flag identification of different countries is a challenging and difficult task. Recognition of an unknown flag image in a scene is challenging due to the diversity of the data and to the complexity of the identification process. The aim of the study is to propose a feature extraction based recognition system for Myanmar's national flag. Image features are acquired from the region and state of flags which are identified by using principal component analysis PCA . PCA is a statistical approach used for reducing the number of features in National flags recognition system. Soe Moe Myint | Moe Moe Myint | Aye Aye Cho "National Flags Recognition Based on Principal Component Analysis" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-3 | Issue-5 , August 2019, URL: https://www.ijtsrd.com/papers/ijtsrd26775.pdfPaper URL: https://www.ijtsrd.com/other-scientific-research-area/other/26775/national-flags-recognition-based-on-principal-component-analysis/soe-moe-myint
Combination of texture feature extraction and forward selection for one-clas...IJECEIAES
This study aims to validate self-portraits using one-class support vector machine (OCSVM). To validate accurately, we build a model by combining texture feature extraction methods, Haralick and local binary pattern (LBP). We also reduce irrelevant features using forward selection (FS). OCSVM was selected because it can solve the problem caused by the inadequate variation of the negative class population. In OCSVM, we only need to feed the algorithm using the true class data, and the data with pattern that does not match will be classified as false. However, combining the two feature extractions produces many features, leading to the curse of dimensionality. The FS method is used to overcome this problem by selecting the best features. From the experiments carried out, the Haralick+LBP+FS+OCSVM model outperformed other models with an accuracy of 95.25% on validation data and 91.75% on test data.
FINGERPRINT CLASSIFICATION BASED ON ORIENTATION FIELDijesajournal
ABSTRACT
This paper introduces an effective method of fingerprint classification based on discriminative feature gathering from orientation field. A nonlinear support vector machines (SVMs) is adopted for the classification. The orientation field is estimated through a pixel-Wise gradient descent method and the percentage of directional block classes is estimated. These percentages are classified into four-dimensional vector considered as a good feature that can be combined with an accurate singular point to classify the fingerprint into one of five classes. This method shows high classification accuracy relative to other spatial domain classifiers.
The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...Waqas Tariq
Selection of inputs is one of the most substantial components of classification algorithms for data mining and pattern recognition problems since even the best classifier will perform badly if the inputs are not selected very well. Big data and computational complexity are main cause of bad performance and low accuracy for classical classifiers. In other words, the complexity of classifier method is inversely proportional with its classification efficiency. For this purpose, two hybrid classifiers have been developed by using both type-1 and type-2 fuzzy c-means clustering with cascaded a classifier. In this proposed classifier, a large number of data points are reduced by using fuzzy c-means clustering before applied to a classifier algorithm as inputs. The aim of this study is to investigate the effect of fuzzy clustering on well-known and useful classifiers such as artificial neural networks (ANN) and support vector machines (SVM). Then the role of positive effects of these proposed algorithms were investigated on applied different data sets.
The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...CSCJournals
Selection of inputs is one of the most substantial components of classification algorithms for data mining and pattern recognition problems since even the best classifier will perform badly if the inputs are not selected very well. Big data and computational complexity are main cause of bad performance and low accuracy for classical classifiers. In other words, the complexity of classifier method is inversely proportional with its classification efficiency. For this purpose, two hybrid classifiers have been developed by using both type-1 and type-2 fuzzy c-means clustering with cascaded a classifier. In this proposed classifier, a large number of data points are reduced by using fuzzy c-means clustering before applied to a classifier algorithm as inputs. The aim of this study is to investigate the effect of fuzzy clustering on well-known and useful classifiers such as artificial neural networks (ANN) and support vector machines (SVM). Then the role of positive effects of these proposed algorithms were investigated on applied different data sets.
Dual Tree Complex Wavelet Transform, Probabilistic Neural Network and Fuzzy C...IJAEMSJORNAL
The venture suggests an Adhoc technique of MRI brain image classification and image segmentation tactic. It is a programmed structure for phase classification using learning mechanism and to sense the Brain Tumor through spatial fuzzy clustering methods for bio medical applications. Automated classification and recognition of tumors in diverse MRI images is enthused for the high precision when dealing with human life. Our proposal employs a segmentation technique, Spatial Fuzzy Clustering Algorithm, for segmenting MRI images to diagnose the Brain Tumor in its earlier phase for scrutinizing the anatomical makeup. The Artificial Neural Network (ANN) will be exploited to categorize the pretentious tumor part in the brain. Dual Tree-CWT decomposition scheme is utilized for texture scrutiny of an image. Probabilistic Neural Network (PNN)-Radial Basis Function (RBF) will be engaged to execute an automated Brain Tumor classification. The preprocessing steps were operated in two phases: feature mining by means of classification via PNN-RBF network. The functioning of the classifier was assessed with the training performance and classification accuracies.
Special Plenary Lecture at the International Conference on VIBRATION ENGINEERING AND TECHNOLOGY OF MACHINERY (VETOMAC), Lisbon, Portugal, September 10 - 13, 2018
http://www.conf.pt/index.php/v-speakers
Propagation of uncertainties in complex engineering dynamical systems is receiving increasing attention. When uncertainties are taken into account, the equations of motion of discretised dynamical systems can be expressed by coupled ordinary differential equations with stochastic coefficients. The computational cost for the solution of such a system mainly depends on the number of degrees of freedom and number of random variables. Among various numerical methods developed for such systems, the polynomial chaos based Galerkin projection approach shows significant promise because it is more accurate compared to the classical perturbation based methods and computationally more efficient compared to the Monte Carlo simulation based methods. However, the computational cost increases significantly with the number of random variables and the results tend to become less accurate for a longer length of time. In this talk novel approaches will be discussed to address these issues. Reduced-order Galerkin projection schemes in the frequency domain will be discussed to address the problem of a large number of random variables. Practical examples will be given to illustrate the application of the proposed Galerkin projection techniques.
DataEngConf: Feature Extraction: Modern Questions and Challenges at GoogleHakka Labs
By Dmitry Storcheus (Engineer, Google Research)
Feature extraction, as usually understood, seeks an optimal transformation from raw data into features that can be used as an input for a learning algorithm. In recent times this problem has been attacked using a growing number of diverse techniques that originated in separate research communities: from PCA and LDA to manifold and metric learning. The goal of this talk is to contrast and compare feature extraction techniques coming from different machine learning areas as well as discuss the modern challenges and open problems in feature extraction. Moreover, this talk will suggest novel solutions to some of the challenges discussed, particularly to coupled feature extraction.
A NOVEL PROBABILISTIC BASED IMAGE SEGMENTATION MODEL FOR REALTIME HUMAN ACTIV...sipij
Automatic human activity detection is one of the difficult tasks in image segmentation application due to
variations in size, type, shape and location of objects. In the traditional probabilistic graphical
segmentation models, intra and inter region segments may affect the overall segmentation accuracy. Also,
both directed and undirected graphical models such as Markov model, conditional random field have
limitations towards the human activity prediction and heterogeneous relationships. In this paper, we have
studied and proposed a natural solution for automatic human activity segmentation using the enhanced
probabilistic chain graphical model. This system has three main phases, namely activity pre-processing,
iterative threshold based image enhancement and chain graph segmentation algorithm. Experimental
results show that proposed system efficiently detects the human activities at different levels of the action
datasets.
Possibility fuzzy c means clustering for expression invariant face recognitionIJCI JOURNAL
Face being the most natural method of identification for humans is one of the most significant biometric
modalities and various methods to achieve efficient face recognition have been proposed. However the
changes in face owing to different expressions, pose, makeup, illumination, age bring about marked
variations in the facial image. These changes will inevitably occur and they can be controlled only till a
certain degree beyond which they are bound to happen and will affect the face thereby adversely impacting
the performance of any face recognition system. This paper proposes a strategy to improve the
classification methodology in face recognition by using Possibility Fuzzy C-Means Clustering (PFCM).
This clustering technique was used for face recognition due to its properties like outlier insensitivity which
make it a suitable candidate for use in designing such robust applications.PFCM is a hybridization of
Possibilistic C-Means (PCM) and Fuzzy C-Means (FCM) clustering algorithms. PFCM is a robust
clustering technique and is especially significant for its noise insensitivity. It has also resolved the
coincident clusters problem which is faced by other clustering techniques. Therefore the technique can also
be used to increase the overall robustness of a face recognition system and thereby increase its invariance
and make it a reliably usable biometric modality.
COMPARATIVE PERFORMANCE ANALYSIS OF RNSC AND MCL ALGORITHMS ON POWER-LAW DIST...acijjournal
Cluster analysis of graph related problems is an important issue now-a-day. Different types of graph
clustering techniques are appeared in the field but most of them are vulnerable in terms of effectiveness
and fragmentation of output in case of real-world applications in diverse systems. In this paper, we will
provide a comparative behavioural analysis of RNSC (Restricted Neighbourhood Search Clustering) and
MCL (Markov Clustering) algorithms on Power-Law Distribution graphs. RNSC is a graph clustering
technique using stochastic local search. RNSC algorithm tries to achieve optimal cost clustering by
assigning some cost functions to the set of clusterings of a graph. This algorithm was implemented by A.
D. King only for undirected and unweighted random graphs. Another popular graph clustering
algorithm MCL is based on stochastic flow simulation model for weighted graphs. There are plentiful
applications of power-law or scale-free graphs in nature and society. Scale-free topology is stochastic i.e.
nodes are connected in a random manner. Complex network topologies like World Wide Web, the web of
human sexual contacts, or the chemical network of a cell etc., are basically following power-law
distribution to represent different real-life systems. This paper uses real large-scale power-law
distribution graphs to conduct the performance analysis of RNSC behaviour compared with Markov
clustering (MCL) algorithm. Extensive experimental results on several synthetic and real power-law
distribution datasets reveal the effectiveness of our approach to comparative performance measure of
these algorithms on the basis of cost of clustering, cluster size, modularity index of clustering results and
normalized mutual information (NMI).
Similar to Distance Metric Learning tutorial at CVPR 2015 (20)
Forklift Classes Overview by Intella PartsIntella Parts
Discover the different forklift classes and their specific applications. Learn how to choose the right forklift for your needs to ensure safety, efficiency, and compliance in your operations.
For more technical information, visit our website https://intellaparts.com
Cosmetic shop management system project report.pdfKamal Acharya
Buying new cosmetic products is difficult. It can even be scary for those who have sensitive skin and are prone to skin trouble. The information needed to alleviate this problem is on the back of each product, but it's thought to interpret those ingredient lists unless you have a background in chemistry.
Instead of buying and hoping for the best, we can use data science to help us predict which products may be good fits for us. It includes various function programs to do the above mentioned tasks.
Data file handling has been effectively used in the program.
The automated cosmetic shop management system should deal with the automation of general workflow and administration process of the shop. The main processes of the system focus on customer's request where the system is able to search the most appropriate products and deliver it to the customers. It should help the employees to quickly identify the list of cosmetic product that have reached the minimum quantity and also keep a track of expired date for each cosmetic product. It should help the employees to find the rack number in which the product is placed.It is also Faster and more efficient way.
HEAP SORT ILLUSTRATED WITH HEAPIFY, BUILD HEAP FOR DYNAMIC ARRAYS.
Heap sort is a comparison-based sorting technique based on Binary Heap data structure. It is similar to the selection sort where we first find the minimum element and place the minimum element at the beginning. Repeat the same process for the remaining elements.
An Approach to Detecting Writing Styles Based on Clustering Techniquesambekarshweta25
An Approach to Detecting Writing Styles Based on Clustering Techniques
Authors:
-Devkinandan Jagtap
-Shweta Ambekar
-Harshit Singh
-Nakul Sharma (Assistant Professor)
Institution:
VIIT Pune, India
Abstract:
This paper proposes a system to differentiate between human-generated and AI-generated texts using stylometric analysis. The system analyzes text files and classifies writing styles by employing various clustering algorithms, such as k-means, k-means++, hierarchical, and DBSCAN. The effectiveness of these algorithms is measured using silhouette scores. The system successfully identifies distinct writing styles within documents, demonstrating its potential for plagiarism detection.
Introduction:
Stylometry, the study of linguistic and structural features in texts, is used for tasks like plagiarism detection, genre separation, and author verification. This paper leverages stylometric analysis to identify different writing styles and improve plagiarism detection methods.
Methodology:
The system includes data collection, preprocessing, feature extraction, dimensional reduction, machine learning models for clustering, and performance comparison using silhouette scores. Feature extraction focuses on lexical features, vocabulary richness, and readability scores. The study uses a small dataset of texts from various authors and employs algorithms like k-means, k-means++, hierarchical clustering, and DBSCAN for clustering.
Results:
Experiments show that the system effectively identifies writing styles, with silhouette scores indicating reasonable to strong clustering when k=2. As the number of clusters increases, the silhouette scores decrease, indicating a drop in accuracy. K-means and k-means++ perform similarly, while hierarchical clustering is less optimized.
Conclusion and Future Work:
The system works well for distinguishing writing styles with two clusters but becomes less accurate as the number of clusters increases. Future research could focus on adding more parameters and optimizing the methodology to improve accuracy with higher cluster values. This system can enhance existing plagiarism detection tools, especially in academic settings.
Student information management system project report ii.pdfKamal Acharya
Our project explains about the student management. This project mainly explains the various actions related to student details. This project shows some ease in adding, editing and deleting the student details. It also provides a less time consuming process for viewing, adding, editing and deleting the marks of the students.
Literature Review Basics and Understanding Reference Management.pptxDr Ramhari Poudyal
Three-day training on academic research focuses on analytical tools at United Technical College, supported by the University Grant Commission, Nepal. 24-26 May 2024
Final project report on grocery store management system..pdfKamal Acharya
In today’s fast-changing business environment, it’s extremely important to be able to respond to client needs in the most effective and timely manner. If your customers wish to see your business online and have instant access to your products or services.
Online Grocery Store is an e-commerce website, which retails various grocery products. This project allows viewing various products available enables registered users to purchase desired products instantly using Paytm, UPI payment processor (Instant Pay) and also can place order by using Cash on Delivery (Pay Later) option. This project provides an easy access to Administrators and Managers to view orders placed using Pay Later and Instant Pay options.
In order to develop an e-commerce website, a number of Technologies must be studied and understood. These include multi-tiered architecture, server and client-side scripting techniques, implementation technologies, programming language (such as PHP, HTML, CSS, JavaScript) and MySQL relational databases. This is a project with the objective to develop a basic website where a consumer is provided with a shopping cart website and also to know about the technologies used to develop such a website.
This document will discuss each of the underlying technologies to create and implement an e- commerce website.
Technical Drawings introduction to drawing of prisms
Distance Metric Learning tutorial at CVPR 2015
1. Distance Metric Learning for
Set-based Visual Recognition
Ruiping Wang
Institute of Computing Technology (ICT),
Chinese Academy of Sciences (CAS)
June 7, 2015
CVPR2015 Tutorial on “Distance Metric Learning for Visual Recognition” Part-4
3. InstituteofComputingTechnology,ChineseAcademyofSciences
Identification
Typical applications
Photo matching (1:N)
Watch list screening (1:N+1)
Performance metric
FR(@FAR)
Verification
Typical applications
Access control (1:1)
E-passport (1:1)
Performance metric
ROC: FRR+FAR
3
Face Recognition with Single Image
Who is this celebrity? Are they the same guy?
10. InstituteofComputingTechnology,ChineseAcademyofSciences
1010
Face Recognition with Videos
Video shot retrieval
1 32 4 5 76 9 Next Page8
S01E01: 10’48’’ S01E06: 05’22’’ S01E02: 04’20’’ S01E02: 00’21’’
S01E03: 08’06’’ S01E05: 09’23’’ S01E01: 22’20’’ S01E04: 03’42’’
Smart TV-Series Character
Shots Retrieval System
“the Big Bang Theory”
13. InstituteofComputingTechnology,ChineseAcademyofSciences
13
Overview of previous works
From the
view of set
modeling
Statistics
[Shakhnarovich, ECCV’02]
[Arandjelović, CVPR’05]
[Wang, CVPR’12]
[Harandi, ECCV’14]
[Wang, CVPR’15]
G1
G2
K-L
Divergence
S1
S2
Linear subspace
[Yamaguchi, FG’98]
[Kim, PAMI’07]
[Hamm, ICML’08]
[Huang, CVPR’15]
Principal
angle
H1
H2
Affine/Convex hull
[Cevikalp, CVPR’10]
[Hu, CVPR’11]
[Zhu, ICCV’13]
NN
matching
Nonlinear manifold
[Kim, BMVC’05]
[Wang, CVPR’08]
[Wang, CVPR’09]
[Chen, CVPR’13]
[Lu, CVPR’15]
Principal
angle(+)
M1
M2
14. InstituteofComputingTechnology,ChineseAcademyofSciences
14
Overview of previous works
Set modeling
Linear subspaceNonlinear manifold
Affine/Convex Hull (affine subspace)
Parametric PDFs Statistics
Set matching—basic distance
Principal angles-based measure
Nearest neighbor (NN) matching approach
K-L divergenceSPD Riemannian metric…
Set matching—metric learning
Learning in Euclidean space
Learning on Riemannian manifold
15. InstituteofComputingTechnology,ChineseAcademyofSciences
15
Set model I: linear subspace
Properties
PCA on the set of image samples to get subspace
Loose characterization of the set distribution region
Principal angles-based measure discards the
varying importance of different variance directions
Methods
MSM [FG’98]
DCC [PAMI’07]
GDA [ICML’08]
GGDA [CVPR’11]
PML [CVPR’15]
…
S1
S2
No distribution boundary
No direction difference
16. InstituteofComputingTechnology,ChineseAcademyofSciences
16
Set model I: linear subspace
MSM (Mutual Subspace Method) [FG’98]
Pioneering work on image set classification
First exploit principal angles as subspace distance
Metric learning: N/A
[1] O. Yamaguchi, K. Fukui, and K. Maeda. Face Recognition Using Temporal Image
Sequence. IEEE FG 1998.
subspace method Mutual subspace method
17. InstituteofComputingTechnology,ChineseAcademyofSciences
17
Set model I: linear subspace
DCC (Discriminant Canonical Correlations) [PAMI’07]
Metric learning: in Euclidean space
[1] T. Kim, J. Kittler, and R. Cipolla. Discriminative Learning and Recognition of Image
Set Classes Using Canonical Correlations. IEEE T-PAMI, 2007.
Set 1: 𝑿 𝟏 Set 2: 𝑿 𝟐
𝑷 𝟏
𝑷 𝟐
Linear subspace by:
orthonormal basis matrix
𝑿𝑖 𝑿𝑖
𝑇
≃ 𝑷𝑖 𝚲𝑖 𝑷𝑖
𝑇
20. InstituteofComputingTechnology,ChineseAcademyofSciences
20
Set model I: linear subspace
GDA [ICML’08] / GGDA [CVPR’11]
Treat subspaces as points on Grassmann manifold
Metric learning: on Riemannian manifold
[1] J. Hamm and D. D. Lee. Grassmann Discriminant Analysis: a Unifying View on Subspace-Based
Learning. ICML 2008.
[2] M. Harandi, C. Sanderson, S. Shirazi, B. Lovell. Graph Embedding Discriminant Analysis on
Grassmannian Manifolds for Improved Image Set Matching. IEEE CVPR 2011.
Subspace Subspace
Grassmann manifold
implicit kernel
mapping
26. InstituteofComputingTechnology,ChineseAcademyofSciences
26
PML
Discriminative learning
Discriminant function
Minimize/Maximize the projection distances of any within-
class/between-class subspace pairs
𝐽 = min 𝑡𝑟 𝑷 𝑇 𝑨𝑖𝑗 𝑨𝑖𝑗 𝑷𝑙 𝑖=𝑙 𝑗
− 𝜆 𝑡𝑟 𝑷 𝑇 𝑨𝑖𝑗 𝑨𝑖𝑗 𝑷𝑙 𝑖≠𝑙 𝑗
Optimization algorithm
Iterative solution for one of 𝒀′
and 𝑷 by fixing the other
Normalization of 𝒀 by QR-decomposition
Computation of 𝑷 by Riemannian Conjugate Gradient (RCG)
algorithm on the manifold of PSD matrices
within-class between-class
27. InstituteofComputingTechnology,ChineseAcademyofSciences
27
Set model II: nonlinear manifold
Properties
Capture nonlinear complex appearance variation
Need dense sampling and large amount of data
Less appealing computational efficiency
Methods
MMD [CVPR’08]
MDA [CVPR’09]
BoMPA [BMVC’05]
SANS [CVPR’13]
MMDML [CVPR’15]
…
Complex distribution
Large amount of data
28. InstituteofComputingTechnology,ChineseAcademyofSciences
28
Set model II: nonlinear manifold
MMD (Manifold-Manifold Distance) [CVPR’08]
Set modeling with nonlinear appearance manifold
Image set classification distance computation
between two manifolds
Metric learning: N/A
D(M1,M2)
[1] R. Wang, S. Shan, X. Chen, W. Gao. Manifold-Manifold Distance with Application to Face
Recognition based on Image Set. IEEE CVPR 2008. (Best Student Poster Award Runner-up)
[2] R. Wang, S. Shan, X. Chen, Q. Dai, W. Gao. Manifold-Manifold Distance and Its Application to
Face Recognition with Image Sets. IEEE Trans. Image Processing, 2012.
30. InstituteofComputingTechnology,ChineseAcademyofSciences
30
MMD
Formulation & Solution
SSD between pairs
of local models
M1
M2
local linear models
iC jC
1C 2C 3C
1C 2C 3C
Three modules
Local model construction
Local model distance
Global integration of local
distances
1 2 1 1
( , ) ( , )
m n
ij i ji j
d f d C C
M M
1 1
. . 1, 0
m n
ij iji j
s t f f
32. InstituteofComputingTechnology,ChineseAcademyofSciences
32
MDA (Manifold Discriminant Analysis ) [CVPR’09]
Goal: maximize “manifold margin” under Graph
Embedding framework
Metric learning: in Euclidean space
Euclidean distance between pair of image samples
[1] R. Wang, X. Chen. Manifold Discriminant Analysis. IEEE CVPR 2009.
M1
M2
within-class
compactness
between-class
separability
M1
M2
Intrinsic graph Penalty graph
Set model II: nonlinear manifold
33. InstituteofComputingTechnology,ChineseAcademyofSciences
33
Optimization framework
Min. within-class scatter
Max. between-class scatter
Objective function
Global linear transformation
Solved by generalized eigen-decomposition
2
, 2 ( )T T T T
w m n m n
m,n
S w X D W X v x v x v v
2
, 2 ( )T T T T
b m n m n
m,n
S w X D W X v x v x v v
Intrinsic
graph
Penalty
graph
MDA
( )
Maximize ( ) =
( )
T T
b
T T
w
S X D W X
J
S X D W X
v v
v
v v
34. InstituteofComputingTechnology,ChineseAcademyofSciences
34
BoMPA (Boosted Manifold Principal Angles) [BMVC’05]
Goal: optimal fusion of different principal angles
Exploit Adaboost to learn weights for each angle
[1] T. Kim O. Arandjelovic, R. Cipolla. Learning over Sets using Boosted Manifold Principal Angles
(BoMPA). BMVC 2005.
PAsimilar (common illumination) BoMPAdissimilar
Subject A Subject B
Set model II: nonlinear manifold
35. InstituteofComputingTechnology,ChineseAcademyofSciences
35
SANS (Sparse Approximated Nearest Subspaces) [CVPR’13]
Goal: adaptively construct the nearest subspace pair
Metric learning: N/A
[1] S. Chen, C. Sanderson, M.T. Harandi, B.C. Lovell. Improved Image Set Classification via Joint
Sparse Approximated Nearest Subspaces. IEEE CVPR 2013.
Set model II: nonlinear manifold
SSD (Subspace-Subspace Distance):
Joint sparse representation (JSR) is
applied to approximate the nearest
subspace over a Grassmann manifold.
36. InstituteofComputingTechnology,ChineseAcademyofSciences
36
MMDML (Multi-Manifold Deep Metric Learning) [CVPR’15]
Goal: maximize “manifold margin” under Deep Learning
framework
Metric learning: in Euclidean space
Set model II: nonlinear manifold
[1] J. Lu, G. Wang, W. Deng, P. Moulin, and J. Zhou. Multi-Manifold Deep Metric Learning for Image
Set Classification. IEEE CVPR 2015.
Class-specific DNN
Model nonlinearity
Objective function
37. InstituteofComputingTechnology,ChineseAcademyofSciences
37
Set model III: affine subspace
Properties
Linear reconstruction using: mean + subspace basis
Synthesized virtual NN-pair matching
Less characterization of global data structure
Computationally expensive by NN-based matching
Methods
AHISD/CHISD [CVPR’10]
SANP [CVPR’11]
PSDML/SSDML [ICCV’13]
…
H1
H2
Sensitive to noise samples
High computation cost
38. InstituteofComputingTechnology,ChineseAcademyofSciences
38
Set model III: affine subspace
AHISD/CHISD [CVPR’10], SANP [CVPR’11]
NN-based matching using sample Euclidean distance
Metric learning: N/A
Subspace spanned by all the available samples
𝑫 = {𝑑1, … , 𝑑 𝑛} in the set
Affine hull
𝐻 𝑫 = 𝑫𝜶 = 𝑑𝑖 𝛼𝑖| 𝛼𝑖 = 1
Convex hull
𝐻 𝑫 = {𝑫𝜶 = 𝑑𝑖 𝛼𝑖| 𝛼𝑖 = 1, 0 ≤ 𝛼𝑖 ≤ 1}
[1] H. Cevikalp, B. Triggs. Face Recognition Based on Image Sets. IEEE CVPR 2010.
[2] Y. Hu, A.S. Mian, R. Owens. Sparse Approximated Nearest Points for Image Set
Classification. IEEE CVPR 2011.
39. InstituteofComputingTechnology,ChineseAcademyofSciences
39
Set model III: affine subspace
PSDML/SSDML [ICCV’13]
Metric learning: in Euclidean space
[1] P. Zhu, L. Zhang, W. Zuo, and D. Zhang. From Point to Set: Extend the Learning of
Distance Metrics. IEEE ICCV 2013.
point-to-set distance
metric learning (PSDML)
set-to-set distance metric
learning (SSDML)
44. InstituteofComputingTechnology,ChineseAcademyofSciences
44
Set model IV: statistics (COV+)
Properties
The natural raw statistics of a sample set
Flexible model of multiple-order statistical information
Methods
CDL [CVPR’12]
LMKML [ICCV’13]
DARG [CVPR’15]
SPD-ML [ECCV’14]
LEML [ICML’15]
LERM [CVPR’14]
HER [CVPR’15 ]
…
G1
G2
Natural raw statistics
More flexible model
[Shakhnarovich, ECCV’02]
[Arandjelović, CVPR’05]
45. InstituteofComputingTechnology,ChineseAcademyofSciences
45
CDL (Covariance Discriminative Learning) [CVPR’12]
Set modeling by Covariance Matrix (COV)
The 2nd order statistics characterizing set data variations
Robust to noisy set data, scalable to varying set size
Metric learning: on the SPD manifold
[1] R. Wang, H. Guo, L.S. Davis, Q. Dai. Covariance Discriminative Learning: A Natural
and Efficient Approach to Image Set Classification. IEEE CVPR 2012.
TI
log
Si
Ci Cj
Sj
I
Model data variations
No assum. of data distr.
Set model IV: statistics (COV+)
46. InstituteofComputingTechnology,ChineseAcademyofSciences
46
CDL
Set modeling by Covariance Matrix
1 2= [ , ,..., ]N d NX x x x
COV: d*d symmetric positive
definite (SPD) matrix*
1
1
( )( )
1
N
T
i i
iN
C x x x x
Image set: N samples with d-
dimension image feature
*: use regularization to tackle singularity
problem
47. InstituteofComputingTechnology,ChineseAcademyofSciences
47
CDL
Set matching on COV manifold
Riemannian metrics on the SPD manifold
Affine-invariant distance (AID) [1]
or
Log-Euclidean distance (LED) [2]
1 2 1 2( , ) log ( ) log ( ) F
d I IC C C C
2 2
1 2 1 21
( , ) ln ( , )
d
ii
d
C C C C
22 -1/2 -1/2
1 2 1 2 1( , ) log ( ) F
d IC C C C C
High
computational
burden
More efficient,
more appealing
[1] W. Förstner and B. Moonen. A Metric for Covariance Matrices. Technical Report 1999.
[2] V. Arsigny, P. Fillard, X. Pennec and N. Ayache. Geometric Means In A Novel Vector Space
Structure On Symmetric Positive-Definite Matrices. SIAM J. MATRIX ANAL. APPL. Vol. 29, No. 1, pp.
328-347, 2007.
48. InstituteofComputingTechnology,ChineseAcademyofSciences
48
CDL
Set matching on COV manifold (cont.)
Explicit Riemannian kernel feature mapping with LED
1 2 1 2( , ) [log ( ) log ( )]logk trace I IC C C C
Mercer’s
theorem
Tangent space at
Identity matrix I
Riemannian manifold
of non-singular COV
M
TIM
X1 Y1
0
X2
X3
Y2
Y3
x1
x2
x3
y1
y2
y3
logI logI
: log ( ), ( )d d
log R
IC C
49. InstituteofComputingTechnology,ChineseAcademyofSciences
49
CDL
Discriminative learning on COV manifold
Partial Least Squares (PLS ) regression
Goal: Maximize the covariance between
observations and class labels
Space X Space Y
Linear Projection
Latent Space T
feature vectors,
e.g. log-COV
matrices
class label
vectors, e.g. 1/0
indicators
X = T * P’
Y = T * B * C’ = X * Bpls
T is the common latent
representation
50. InstituteofComputingTechnology,ChineseAcademyofSciences
COVSPD manifold
Model
Metric
Kernel
SubspaceGrassmannian
Model
Metric
Kernel
50
CDL vs. GDA
u1
u2
1
1
( )( )
1
N
T
i i
iN
C x x x x
*
T
C U U
1 2[ , , , ]m D mU u u u
1/ 2
1 2 1 1 2 2( , ) 2 T T
proj F
d
U U U U U U
: log ( ), ( )d d
log R
IC C : , ( , )T d d
proj m D R
U UU
1 2 1 2( , ) log ( ) log ( ) F
d I IC C C C
log ( ) log ( ) T
I IC U U
51. InstituteofComputingTechnology,ChineseAcademyofSciences
51
LMKML (Localized Multi-Kernel Metric Learning) [ICCV’13]
Exploring multiple order statistics
Data-adaptive weights for different types of features
Ignoring the geometric structure of 2nd/3rd-order statistics
Metric learning: in Euclidean space
[1] J. Lu, G. Wang, and P. Moulin. Image Set Classification Using Holistic Multiple Order Statistics
Features and Localized Multi-Kernel Metric Learning. IEEE ICCV 2013.
Set model IV: statistics (COV+)
Complementary information
(mean vs. covariance)
1st / 2nd / 3rd-order statistics
Objective function
52. InstituteofComputingTechnology,ChineseAcademyofSciences
52
DARG (Discriminant Analysis on Riemannian manifold of
Gaussian distributions) [CVPR’15]
Set modeling by mixture of Gaussian distribution (GMM)
Naturally encode the 1st order and 2nd order statistics
Metric learning: on Riemannian manifold
[1] W. Wang, R. Wang, Z. Huang, S. Shan, X. Chen. Discriminant Analysis on Riemannian Manifold
of Gaussian Distributions for Face Recognition with Image Sets. IEEE CVPR 2015.
𝓜:Riemannian
manifold of Gaussian
distributions
Non-discriminative
Time-consuming
×
Set model IV: statistics (COV+)
[Shakhnarovich, ECCV’02]
[Arandjelović, CVPR’05]
54. InstituteofComputingTechnology,ChineseAcademyofSciences
54
DARG
Kernels on the Gaussian distribution manifold
kernel based on Lie Group
Distance based on Lie Group (LGD)
𝐿𝐺𝐷 𝑃𝑖, 𝑃𝑗 = log 𝑃𝑖 − log 𝑃𝑗 𝐹
,
𝑔~𝑁(𝑥|𝜇, Σ) ⟼ 𝑃 = Σ −
1
𝑑+1
Σ + 𝜇𝜇 𝑇
𝜇
𝜇 𝑇 1
Kernel function
𝐾LGD 𝑔𝑖, 𝑔𝑗 = exp −
𝐿𝐺𝐷2
𝑃𝑖, 𝑃𝑗
2𝑡2
SPD matrix according
to information geometry
55. InstituteofComputingTechnology,ChineseAcademyofSciences
55
DARG
Kernels on the Gaussian distribution manifold
kernel based on Lie Group
kernel based on MD and LED
Mahalanobis Distance (MD) between mean
𝑀𝐷 𝜇𝑖, 𝜇 𝑗 = 𝜇𝑖 − 𝜇 𝑗
𝑇
Σ𝑖
−1
+ Σ𝑗
−1
𝜇𝑖 − 𝜇 𝑗
LED between covariance matrix
𝐿𝐸𝐷 Σ𝑖, Σ𝑗 = log Σ𝑖 − log Σ𝑗 𝐹
Kernel function
𝐾 𝑀𝐷+𝐿𝐸𝐷 𝑔𝑖, 𝑔𝑗 = 𝛾1 𝐾 𝑀𝐷 𝜇𝑖, 𝜇 𝑗 + 𝛾2 𝐾𝐿𝐸𝐷 Σ𝑖, Σ𝑗
𝐾 𝑀𝐷 𝜇𝑖, 𝜇 𝑗 = exp −
MD2
𝜇𝑖, 𝜇 𝑗
2𝑡2
𝐾𝐿𝐸𝐷 Σ𝑖, Σ𝑗 = exp −
𝐿𝐸𝐷2
Σ𝑖, Σ𝑗
2𝑡2
𝛾1, 𝛾2 are the combination coefficients
57. InstituteofComputingTechnology,ChineseAcademyofSciences
57
Set model IV: statistics (COV+)
Properties
The natural raw statistics of a sample set
Flexible model of multiple-order statistical information
Methods
CDL [CVPR’12]
LMKML [ICCV’13]
DARG [CVPR’15]
SPD-ML [ECCV’14]
LEML [ICML’15]
LERM [CVPR’14]
HER [CVPR’15 ]
…
G1
G2
Natural raw statistics
More flexible model
[Shakhnarovich, ECCV’02]
[Arandjelović, CVPR’05]
58. InstituteofComputingTechnology,ChineseAcademyofSciences
58
SPD-ML (SPD Manifold Learning) [ECCV’14]
Pioneering work on explicit manifold-to-manifold
dimensionality reduction
Metric learning: on Riemannian manifold
Set model IV: statistics (COV+)
[1] M. Harandi, M. Salzmann, R. Hartley. From Manifold to Manifold: Geometry-Aware
Dimensionality Reduction for SPD Matrices. ECCV 2014.
Structure distortion
No explicit mapping
64. InstituteofComputingTechnology,ChineseAcademyofSciences
64
Set model IV: statistics (COV+)
Properties
The natural raw statistics of a sample set
Flexible model of multiple-order statistical information
Methods
CDL [CVPR’12]
LMKML [ICCV’13]
DARG [CVPR’15]
SPD-ML [ECCV’14]
LEML [ICML’15]
LERM [CVPR’14]
HER [CVPR’15 ]
…
G1
G2
Natural raw statistics
More flexible model
[Shakhnarovich, ECCV’02]
[Arandjelović, CVPR’05]
65. InstituteofComputingTechnology,ChineseAcademyofSciences
65
LERM (Learning Euclidean-to-Riemannian Metric) [CVPR’14]
Application scenario: still-to-video face recognition
Metric learning: cross Euclidean space and Riemannian manifold
[1] Z. Huang, R. Wang, S. Shan, X. Chen. Learning Euclidean-to-Riemannian Metric for Point-to-Set
Classification. IEEE CVPR 2014.
ℝ 𝐷
Watch list… Surveillance video
Point Set
Point-to-Set
Classification
Watch list screening
Set model IV: statistics (COV+)
71. InstituteofComputingTechnology,ChineseAcademyofSciences
71
HER (Hashing across Euclidean and Riemannian) [CVPR’15]
Application scenario: video-based face retrieval
Metric learning: hamming distance learning across heter. spaces
[1] Y. Li, R. Wang, Z. Huang, S. Shan, X. Chen. Face Video Retrieval with Image Query via Hashing
across Euclidean Space and Riemannian Manifold. IEEE CVPR 2015.
Set model IV: statistics (COV+)
72. InstituteofComputingTechnology,ChineseAcademyofSciences
72
HER
Heterogeneous Hash Learning
The Proposed Heterogeneous
Hash Learning Method
Previous Multiple Modalities
Hash Learning Methods
Euclidean Space BEuclidean Space A
Hamming Space
001
011 111
110
100000
101
Riemannian ManifoldEuclidean Space
Hamming Space
001
011 111
110
100000
101
73. InstituteofComputingTechnology,ChineseAcademyofSciences
73
HER
Two-Stage Architecture
Stage-1: kernel mapping
Euclidean Kernel (Image)
𝑲 𝑒
𝑖𝑗
= exp(−
𝒙𝑖 − 𝒙𝑗 2
2
2𝜎𝑒
2 )
Riemannian Kernel (Video)
𝑲 𝑟
𝑖𝑗
= exp(−
log(𝔂𝑖) − log(𝔂 𝑗)
𝐹
2
2𝜎𝑟
2 )*
Reproducing Kernel
Hilbert Space (Euc.)
Reproducing Kernel
Hilbert Space (Rie.)
011
Target Hamming Space
001
111
110
100000
101
Subject A
Subject B
Subject C
𝜑 𝜂
[*] Y. Li, R. Wang, Z. Cui, S. Shan, X. Chen. Compact Video Code and Its Application to Robust Face
Retrieval in TV-Series. BMVC 2014. (**: CVC method for video-video face retrieval)
Euclidean Space
(Image-Vector-𝒙𝑖)
Riemannian Manifold
(Video-Covariance Matrix-𝔂𝑖)
74. InstituteofComputingTechnology,ChineseAcademyofSciences
74
HER
Two-Stage Architecture
Stage-2: binary encoding
𝜑 𝜂
Euclidean Space
(Image-Vector-𝒙𝑖)
Riemannian Manifold
(Video-Covariance Matrix-𝔂𝑖)
Reproducing Kernel
Hilbert Space (Euc.)
Reproducing Kernel
Hilbert Space (Rie.)
Hyperplane
Discriminability
Stability
011
Target Hamming Space
001
111
110
100000
101
Subject A
Subject B
Subject C
Hash Functions
𝝎 𝑒
Hash Functions
𝝎 𝑟
82. InstituteofComputingTechnology,ChineseAcademyofSciences
82
Evaluations
Performance on PaSC Challenge (IEEE FG’15)
HERML-DeLF
DCNN learned image feature
Hybrid Euclidean and Riemannian Metric Learning*
[*] Z. Huang, R. Wang, S. Shan, X. Chen. Hybrid Euclidean-and-Riemannian Metric Learning for Image
Set Classification. ACCV 2014. (**: the key reference describing the method used for the challenge)
83. InstituteofComputingTechnology,ChineseAcademyofSciences
83
Evaluations
Performance on EmotiW Challenge (ACM ICMI’14)*
Combination of multiple statistics for video modeling
Learning on the Riemannian manifold
[*] M. Liu, R. Wang, S. Li, S. Shan, Z. Huang, X. Chen. Combining Multiple Kernel Methods on
Riemannian Manifold for Emotion Recognition in the Wild. ACM ICMI 2014.
85. InstituteofComputingTechnology,ChineseAcademyofSciences
85
Route map of our methods
MMD
CVPR’08/TIP’12
MDA
CVPR’09
CDL
CVPR’12
DARG
CVPR’15
LERM
CVPR’14
CVC
BMVC’14
LEML/PML
ICML’15/CVPR’15
HER
CVPR’15
Complex distribution
Large amount of data
Appearance manifold
Natural raw statistics
No assum. of data dist.
Covariance matrix
binary heter.
86. InstituteofComputingTechnology,ChineseAcademyofSciences
86
Summary
What we learn from current studies
Set modeling
Linear(/affine) subspace Manifold Statistics
Set matching
Non-discriminative Discriminative
Metric learning
Euclidean space Riemannian manifold
Future directions
More flexible set modeling for different scenarios
Multi-model combination
Learning method should be more efficient
Set-based vs. sample-based?
87. InstituteofComputingTechnology,ChineseAcademyofSciences
87
Additional references (not listed above)
[Arandjelović, CVPR’05] O. Arandjelović, G. Shakhnarovich, J. Fisher, R.
Cipolla, and T. Darrell. Face Recognition with Image Sets Using Manifold
Density Divergence. IEEE CVPR 2005.
[Chien, PAMI’02] J. Chien and C. Wu. Discriminant waveletfaces and nearest
feature classifiers for face recognition. IEEE T-PAMI 2002.
[Rastegari, ECCV’12] M. Rastegari, A. Farhadi, and D. Forsyth. Attribute
discovery via predictable discriminative binary codes. ECCV 2012.
[Shakhnarovich, ECCV’02] G. Shakhnarovich, J. W. Fisher, and T. Darrell.
Face Recognition from Long-term Observations. ECCV 2002.
[Vemulapalli, CVPR’13] R. Vemulapalli, J. K. Pillai, and R. Chellappa. Kernel
learning for extrinsic classification of manifold features. IEEE CVPR 2013.
[Vincent, NIPS’01] P. Vincent and Y. Bengio. K-local hyperplane and convex
distance nearest neighbor algorithms. NIPS 2001.
88. InstituteofComputingTechnology,ChineseAcademyofSciences
8888
Thanks, Q & A
Xilin ChenShiguang ShanRuiping Wang
Lab of Visual Information Processing and Learning (VIPL) @ICT@CAS
Codes of our methods available at: http://vipl.ict.ac.cn/resources/codes
Zhiwu Huang Yan Li Wen Wang Mengyi Liu