This document summarizes a research paper on a multi-center convolutional network for unconstrained face alignment. The proposed network partitions facial landmarks into clusters and uses multiple, center-specific prediction layers to estimate landmark locations for each cluster. This allows the network to focus on predicting landmarks within local regions. Experimental results on two challenging datasets show the multi-center network achieves state-of-the-art accuracy for face alignment while running in real-time on a CPU.
Medoid based model for face recognition using eigen and fisher facesijscmcj
Biometric technologies have gained a remarkable impetus in high security applications. Various biometric modalities are widely being used these days. The need for unobtrusive biometric recognition can be fulfilled through Face recognition which is the most natural and non intrusive authentication system. However the vulnerability to changes owing to variations in face due to various factors like pose,
illumination, ageing, emotions, expressions etc make it necessary to have robust face recognition systems.
Various statistical models have been developed so far with varying degree of accuracy and efficiency. This
paper discusses a new approach to utilize Eigen face and Fisher face methodology by using medoid instead
of mean as a statistic in calculating the Eigen faces and Fisher faces. The method not only requires lesser training but also demonstrates better time efficiency and performance compared to the conventional method of using mean
FACE RECOGNITION USING DIFFERENT LOCAL FEATURES WITH DIFFERENT DISTANCE TECHN...IJCSEIT Journal
A face recognition system using different local features with different distance measures is proposed in this
paper. Proposed method is fast and gives accurate detection. Feature vector is based on Eigen values,
Eigen vectors, and diagonal vectors of sub images. Images are partitioned into sub images to detect local
features. Sub partitions are rearranged into vertically and horizontally matrices. Eigen values, Eigenvector
and diagonal vectors are computed for these matrices. Global feature vector is generated for face
recognition. Experiments are performed on benchmark face YALE database. Results indicate that the
proposed method gives better recognition performance in terms of average recognized rate and retrieval
time compared to the existing methods.
Possibility fuzzy c means clustering for expression invariant face recognitionIJCI JOURNAL
Face being the most natural method of identification for humans is one of the most significant biometric
modalities and various methods to achieve efficient face recognition have been proposed. However the
changes in face owing to different expressions, pose, makeup, illumination, age bring about marked
variations in the facial image. These changes will inevitably occur and they can be controlled only till a
certain degree beyond which they are bound to happen and will affect the face thereby adversely impacting
the performance of any face recognition system. This paper proposes a strategy to improve the
classification methodology in face recognition by using Possibility Fuzzy C-Means Clustering (PFCM).
This clustering technique was used for face recognition due to its properties like outlier insensitivity which
make it a suitable candidate for use in designing such robust applications.PFCM is a hybridization of
Possibilistic C-Means (PCM) and Fuzzy C-Means (FCM) clustering algorithms. PFCM is a robust
clustering technique and is especially significant for its noise insensitivity. It has also resolved the
coincident clusters problem which is faced by other clustering techniques. Therefore the technique can also
be used to increase the overall robustness of a face recognition system and thereby increase its invariance
and make it a reliably usable biometric modality.
SYMMETRICAL WEIGHTED SUBSPACE HOLISTIC APPROACH FOR EXPRESSION RECOGNITIONijcsit
Human face expression is one of the cognitive activity or attribute to deliver the opinions to others.This paper mainly delivers the performance of appearance based holistic approach subspace methods based on Principal Component Analysis (PCA). In this work texture features are extracted from face images using Gabor filter. It was observed that extracted texture feature vector space has higher dimensional and has
more number of redundant contents. Hence training, testing and classification time becomes more. The expression recognition accuracy rate is also reduced. To overcome this problem Symmetrical Weighted 2DPCA (SW2DPCA) subspace method is introduced. Extracted feature vector space is projected in to subspace by using SW2DPCA method. By implementing weighted principles on odd and even symmetrical
decomposition space of training samples sets proposed method have been formed. Conventional PCA and 2DPCA method yields less recognition rate due to larger variations in expressions and light due to more number of feature space redundant variants. Proposed SW2DPCA method optimizes this problem by reducing redundant contents and discarding unequal variants. In this work a well known JAFFE databases
is used for experiments and tested with proposed SW2DPCA algorithm. From the experimental results it was found that facial recognition accuracy rate of GF+SW2DPCA based feature fusion subspace method has been increased to 95.24% compared to 2DPCA method.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
A novel approach for efficient skull stripping using morphological reconstruc...eSAT Journals
Abstract Brain is the part of the central nervous system located in skull. For the diagnosis of human brain bearing tumour, skull stripping plays an important pre-processing role. Skull stripping is the process separating brain and non-brain tissues of the head which is the critical processing step in the analysis of neuroimaging data. Though various algorithms have been proposed to address this problem, challenges remain. In this paper a new efficient skull stripping method for magnetic resonance images (MRI) is proposed. This method adopts a two-step approach; in the first step an improved systematic application of morphological reconstructions operations is done for the brain image and in the second step, a thresholding based technique is used to extract the brain inside the skull. This paper experimented on Axial PD and FLAIR MRI brain images. Index Terms: Skull stripping, thresholding, morphological reconstruction, Axial PD and FLAIR MRI images of brain.
Face detection using the 3 x3 block rank patterns of gradient magnitude imagessipij
Face detection locates faces prior to various face-
related applications. The objective of face detecti
on is to
determine whether or not there are any faces in an
image and, if any, the location of each face is det
ected.
Face detection in real images is challenging due to
large variability of illumination and face appeara
nces.
This paper proposes a face detection algorithm usin
g the 3×3 block rank patterns of gradient magnitude
images and a geometrical face model. First, the ill
umination-corrected image of the face region is obt
ained
using the brightness plane that is produced using t
he locally minimum brightness of each block. Next,
the
illumination-corrected image is histogram equalized
, the face region is divided into nine (3×3) blocks
, and
two directional (horizontal and vertical) gradient
magnitude images are computed, from which the 3×3
block rank patterns are obtained. For face detectio
n, using the FERET and GT databases three types of
the
3×3 block rank patterns are a priori determined as
templates based on the distribution of the sum of t
he
gradient magnitudes of each block in the face candi
date region that is also composed of nine (3×3) blo
cks.
The 3×3 block rank patterns roughly classify whethe
r the detected face candidate region contains a fac
e or
not. Finally, facial features are detected and used
to validate the face model. The face candidate is
validated as a face if it is matched with the geome
trical face model. The proposed algorithm is tested
on the
Caltech database images and real images. Experiment
al results with a number of test images show the
effectiveness of the proposed algorithm.
Medoid based model for face recognition using eigen and fisher facesijscmcj
Biometric technologies have gained a remarkable impetus in high security applications. Various biometric modalities are widely being used these days. The need for unobtrusive biometric recognition can be fulfilled through Face recognition which is the most natural and non intrusive authentication system. However the vulnerability to changes owing to variations in face due to various factors like pose,
illumination, ageing, emotions, expressions etc make it necessary to have robust face recognition systems.
Various statistical models have been developed so far with varying degree of accuracy and efficiency. This
paper discusses a new approach to utilize Eigen face and Fisher face methodology by using medoid instead
of mean as a statistic in calculating the Eigen faces and Fisher faces. The method not only requires lesser training but also demonstrates better time efficiency and performance compared to the conventional method of using mean
FACE RECOGNITION USING DIFFERENT LOCAL FEATURES WITH DIFFERENT DISTANCE TECHN...IJCSEIT Journal
A face recognition system using different local features with different distance measures is proposed in this
paper. Proposed method is fast and gives accurate detection. Feature vector is based on Eigen values,
Eigen vectors, and diagonal vectors of sub images. Images are partitioned into sub images to detect local
features. Sub partitions are rearranged into vertically and horizontally matrices. Eigen values, Eigenvector
and diagonal vectors are computed for these matrices. Global feature vector is generated for face
recognition. Experiments are performed on benchmark face YALE database. Results indicate that the
proposed method gives better recognition performance in terms of average recognized rate and retrieval
time compared to the existing methods.
Possibility fuzzy c means clustering for expression invariant face recognitionIJCI JOURNAL
Face being the most natural method of identification for humans is one of the most significant biometric
modalities and various methods to achieve efficient face recognition have been proposed. However the
changes in face owing to different expressions, pose, makeup, illumination, age bring about marked
variations in the facial image. These changes will inevitably occur and they can be controlled only till a
certain degree beyond which they are bound to happen and will affect the face thereby adversely impacting
the performance of any face recognition system. This paper proposes a strategy to improve the
classification methodology in face recognition by using Possibility Fuzzy C-Means Clustering (PFCM).
This clustering technique was used for face recognition due to its properties like outlier insensitivity which
make it a suitable candidate for use in designing such robust applications.PFCM is a hybridization of
Possibilistic C-Means (PCM) and Fuzzy C-Means (FCM) clustering algorithms. PFCM is a robust
clustering technique and is especially significant for its noise insensitivity. It has also resolved the
coincident clusters problem which is faced by other clustering techniques. Therefore the technique can also
be used to increase the overall robustness of a face recognition system and thereby increase its invariance
and make it a reliably usable biometric modality.
SYMMETRICAL WEIGHTED SUBSPACE HOLISTIC APPROACH FOR EXPRESSION RECOGNITIONijcsit
Human face expression is one of the cognitive activity or attribute to deliver the opinions to others.This paper mainly delivers the performance of appearance based holistic approach subspace methods based on Principal Component Analysis (PCA). In this work texture features are extracted from face images using Gabor filter. It was observed that extracted texture feature vector space has higher dimensional and has
more number of redundant contents. Hence training, testing and classification time becomes more. The expression recognition accuracy rate is also reduced. To overcome this problem Symmetrical Weighted 2DPCA (SW2DPCA) subspace method is introduced. Extracted feature vector space is projected in to subspace by using SW2DPCA method. By implementing weighted principles on odd and even symmetrical
decomposition space of training samples sets proposed method have been formed. Conventional PCA and 2DPCA method yields less recognition rate due to larger variations in expressions and light due to more number of feature space redundant variants. Proposed SW2DPCA method optimizes this problem by reducing redundant contents and discarding unequal variants. In this work a well known JAFFE databases
is used for experiments and tested with proposed SW2DPCA algorithm. From the experimental results it was found that facial recognition accuracy rate of GF+SW2DPCA based feature fusion subspace method has been increased to 95.24% compared to 2DPCA method.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
A novel approach for efficient skull stripping using morphological reconstruc...eSAT Journals
Abstract Brain is the part of the central nervous system located in skull. For the diagnosis of human brain bearing tumour, skull stripping plays an important pre-processing role. Skull stripping is the process separating brain and non-brain tissues of the head which is the critical processing step in the analysis of neuroimaging data. Though various algorithms have been proposed to address this problem, challenges remain. In this paper a new efficient skull stripping method for magnetic resonance images (MRI) is proposed. This method adopts a two-step approach; in the first step an improved systematic application of morphological reconstructions operations is done for the brain image and in the second step, a thresholding based technique is used to extract the brain inside the skull. This paper experimented on Axial PD and FLAIR MRI brain images. Index Terms: Skull stripping, thresholding, morphological reconstruction, Axial PD and FLAIR MRI images of brain.
Face detection using the 3 x3 block rank patterns of gradient magnitude imagessipij
Face detection locates faces prior to various face-
related applications. The objective of face detecti
on is to
determine whether or not there are any faces in an
image and, if any, the location of each face is det
ected.
Face detection in real images is challenging due to
large variability of illumination and face appeara
nces.
This paper proposes a face detection algorithm usin
g the 3×3 block rank patterns of gradient magnitude
images and a geometrical face model. First, the ill
umination-corrected image of the face region is obt
ained
using the brightness plane that is produced using t
he locally minimum brightness of each block. Next,
the
illumination-corrected image is histogram equalized
, the face region is divided into nine (3×3) blocks
, and
two directional (horizontal and vertical) gradient
magnitude images are computed, from which the 3×3
block rank patterns are obtained. For face detectio
n, using the FERET and GT databases three types of
the
3×3 block rank patterns are a priori determined as
templates based on the distribution of the sum of t
he
gradient magnitudes of each block in the face candi
date region that is also composed of nine (3×3) blo
cks.
The 3×3 block rank patterns roughly classify whethe
r the detected face candidate region contains a fac
e or
not. Finally, facial features are detected and used
to validate the face model. The face candidate is
validated as a face if it is matched with the geome
trical face model. The proposed algorithm is tested
on the
Caltech database images and real images. Experiment
al results with a number of test images show the
effectiveness of the proposed algorithm.
A Review Paper on Stereo Vision Based Depth EstimationIJSRD
Stereo vision is a challenging problem and it is a wide research topic in computer vision. It has got a lot of attraction because it is a cost efficient way in place of using costly sensors. Stereo vision has found a great importance in many fields and applications in today’s world. Some of the applications include robotics, 3-D scanning, 3-D reconstruction, driver assistance systems, forensics, 3-D tracking etc. The main challenge of stereo vision is to generate accurate disparity map. Stereo vision algorithms usually perform four steps: first, matching cost computation; second, cost aggregation; third, disparity computation or optimization; and fourth, disparity refinement. Stereo matching problems are also discussed. A large number of algorithms have been developed for stereo vision. But characterization of their performance has achieved less attraction. This paper gives a brief overview of the existing stereo vision algorithms. After evaluating the papers we can say that focus has been on cost aggregation and multi-step refinement process. Segment-based methods have also attracted attention due to their good performance. Also, using improved filter for cost aggregation in stereo matching achieves better results.
Hierarchical Approach for Total Variation Digital Image InpaintingIJCSEA Journal
The art of recovering an image from damage in an undetectable form is known as inpainting. The manual work of inpainting is most often a very time consum ing process. Due to digitalization of this technique, it is automatic and faster. In this paper, after the user selects the regions to be reconstructed, the algorithm automatically reconstruct the lost regions with the help of the information surrounding them. The existing methods perform very well when the region to be reconstructed is very small, but fails in proper reconstruction as the area increases. This paper describes a Hierarchical method by which the area to be inpainted is reduced in multiple levels and Total Variation(TV) method is used to inpaint in each level. This algorithm gives better performance when compared with other existing algorithms such as nearest neighbor interpolation, Inpainting through Blurring and Sobolev Inpainting.
Engineering Research Publication
Best International Journals, High Impact Journals,
International Journal of Engineering & Technical Research
ISSN : 2321-0869 (O) 2454-4698 (P)
www.erpublication.org
The improved hybrid model for molecular image denoising, proposed by NeST Software, can give a better SNR Molecular Image output. Read more on the proposed hybrid model.
Segmentation of medical images using metric topology – a region growing approachIjrdt Journal
A metric topological approach to the region growing based segmentation is presented in this article. Region based growing techniques has gained a significant importance in the medical image processing field for finest of segregation of tumor detected part in the image. Conventional algorithms were concentrated on segmentation at the coarser level which failed to produce enough evidence for the validity of the algorithm. In this article a novel technique is proposed based on metric topological neighbourhood also with the introduction of new objective measure entropy, apart from the traditional validity measures of Accuracy, PSNR and MSE. This measure is introduced to prove the amount of information lost after segmentation is reduced to greater extent which elucidates the effectiveness of the algorithm. This algorithm is tested on the well known benchmarking of testing in ground truth images in par with the proposed region based growing segmented images. The results validated show the validation of effectiveness of the algorithm.
Image restoration based on morphological operationsijcseit
Image processing including noise suppression, feature extraction, edge detection, image segmentation,
shape recognition, texture analysis, image restoration and reconstruction, image compression etc uses
mathematical morphology which is a method of nonlinear filters.
It is modulated from traditional morphology to order morphology, soft mathematical morphology and fuzzy
soft mathematical morphology. This paper is covers 6 morphological operations which are implemented in
the matlab program, including erosion, dilation, opening, closing, boundary extraction and region filling.
FINGERPRINT CLASSIFICATION BASED ON ORIENTATION FIELDijesajournal
ABSTRACT
This paper introduces an effective method of fingerprint classification based on discriminative feature gathering from orientation field. A nonlinear support vector machines (SVMs) is adopted for the classification. The orientation field is estimated through a pixel-Wise gradient descent method and the percentage of directional block classes is estimated. These percentages are classified into four-dimensional vector considered as a good feature that can be combined with an accurate singular point to classify the fingerprint into one of five classes. This method shows high classification accuracy relative to other spatial domain classifiers.
Face skin color based recognition using local spectral and gray scale featureseSAT Journals
Abstract Human face conveys more information about identity of person. Human face recognition is one of the most challenging problem and it can be used in many applications at different security places in airports, defense and banking sectors etc.In this work used colored features obtained from color segmentation because in real time scenario color provides the more information than gray scale image but it has a drawback. To overcome this drawback gray scale feature extracted from co-occurrence matrix of an image and for efficient face recognition of human Face under different illumination conditions spectral features can be extracted from face texture. These three feature vectors concatenated into a single feature vector and applied Lenc-Kral matching technique to measure similarity between the database and query image, the similarity is high then face is recognized. Keywords: Face recognition, illumination condition, local texture features, color segmentation.
STUDY ANALYSIS ON TEETH SEGMENTATION USING LEVEL SET METHODaciijournal
The three dimensional shape information of teeth from cone beam computed tomography images provides
important assistance for dentist performing implant treatment, orthodontic surgery. This paper describes
the tooth root of both anterior and posterior teeth from CBCT images of head. The segmentation is done
using level set method with five energy functions. The edge energy used to move the curve towards border
of the object. The shape prior energy provides the shape of the contour. The dentine wall energy provides
interaction between the neighboring teeth and prevent shrinkage and leakage problem. The test result for
both segmentation and 3D reconstruction shows that the method can visualize both anterior and posterior
teeth with high accuracy and efficiency.
Extraction of texture features by using gabor filter in wheat crop disease de...eSAT Journals
Abstract
Like country India, there are so many people depending upon agriculture. In this area, many farmers don’t know about new
diseases which are impacting on their farm. As the disease changes, the disease control policy also changes. So many farmers
have very sharp observation on crop diseases, but whenever there is new diseases fall on crops then problems occur. Climate also
changes instantly many of times, because of such reasons farmers unable to understand various diseases.
If farmer unable to predict that diseases quickly then it will affect life of crops. Indirectly it gets affects on total productivity of
farm. As we are well known about that world facing lot of problems due rapid growth in population. So our goal is to increase
agricultural productivity using image processing technology which can help farmer in great extent [7].
In this research work, we are trying that crop disease using Artificial neural network (ANN) which work very effectively. First of
all, we have provided an digital image which is taken by digital camera. That image given to Gaussian filter firstly then
transferred to adaptive median filter to filter out noise present inside image. Gaussian filter removes Gaussian noise which is
present inside image. Adaptive noise filter removes impulsive noise which is present inside image. Also it will reduce distortions
which are present inside images. Then image transferred to segmentation part. In image segmentation we have choose CIELAB
color space method to extract color components properly. For segmentation we have used Gabor filter. After this we distinguish
crop diseases on the basis of texture features which are extracted by Gabor filter [6].
Key Words: Artificial Neural Networks, Image preprocessing, Image Acquisition, and Feature Extraction,
classification etc…
Development of stereo matching algorithm based on sum of absolute RGB color d...IJECEIAES
This article presents local-based stereo matching algorithm which comprises a devel- opment of an algorithm using block matching and two edge preserving filters in the framework. Fundamentally, the matching process consists of several stages which will produce the disparity or depth map. The problem and most challenging work for matching process is to get an accurate corresponding point between two images. Hence, this article proposes an algorithm for stereo matching using improved Sum of Absolute RGB Differences (SAD), gradient matching and edge preserving filters. It is Bilateral Filter (BF) to surge up the accuracy. The SAD and gradient matching will be implemented at the first stage to get the preliminary corresponding result, then the BF works as an edge-preserving filter to remove the noise from the first stage. The second BF is used at the last stage to improve final disparity map and increase the object boundaries. The experimental analysis and validation are using the Middlebury standard benchmarking evaluation system. Based on the results, the proposed work is capable to increase the accuracy and to preserve the object edges. To make the proposed work more reliable with current available methods, the quantitative measurement has been made to compare with other existing methods and it shows the proposed work in this article perform much better.
Face Recognition Using Neural Network Based Fourier Gabor Filters & Random Pr...CSCJournals
Face detection and recognition has many applications in a variety of fields such as authentication, security, video surveillance and human interaction systems. In this paper, we present a neural network system for face recognition. Feature vector based on Fourier Gabor filters is used as input of our classifier, which is a Back Propagation Neural Network (BPNN). The input vector of the network will have large dimension, to reduce its feature subspace we investigate the use of the Random Projection as method of dimensionality reduction. Theory and experiment indicates the robustness of our solution.
LITERATURE SURVEY ON SPARSE REPRESENTATION FOR NEURAL NETWORK BASED FACE DETE...csijjournal
Face detection and recognition is a challenging problem in the field of image processing. In this paper, we reviewed some of the recent research works on face recognition. Issues with the previous face recognition
techniques are , time required is more for face recognition , recognition rate and database required to store the data . To overcome these problems sparse representation based classifier technique can be used.
FACE RECOGNITION USING PRINCIPAL COMPONENT ANALYSIS WITH MEDIAN FOR NORMALIZA...csandit
Recognizing Faces helps to name the various subjects present in the image. This work focuses
on labeling faces on an image which includes faces of humans being of various age group
(heterogeneous set ). Principal component analysis concentrates on finds the mean of the data
set and subtracts the mean value from the data set with an intention to normalize that data.
Normalization with respect to image is the removal of common features from the data set. This
work brings in the novel idea of deploying the median another measure of central tendency for
normalization rather than mean. The above work was implemented using matlab. Results show
that Median is the best measure for normalization for a heterogeneous data set which gives
raise to outliers.
Robust face recognition by applying partitioning around medoids over eigen fa...ijcsa
An unsupervised learning methodology for robust face recognition is proposed for enhancing invariance to
various changes in the face. The area of face recognition in spite of being the most unobtrusive biometric
modality of all has encountered challenges with high performance in uncontrolled environment owing to
frequently occurring, unavoidable variations in the face. These changes may be due to noise, outliers,
changing expressions, emotions, pose, illumination, facial distractions like makeup, spectacles, hair growth
etc. Methods for dealing with these variations have been developed in the past with different success.
However the cost and time efficiency play a crucial role in implementing any methodology in real world.
This paper presents a method to integrate the technique of Partitioning Around Medoids with Eigen Faces
and Fisher Faces to improve the efficiency of face recognition considerably. The system so designed has
higher resistance towards the impact of various changes in the face and performs well in terms of success
rate, cost involved and time complexity. The methodology can therefore be used in developing highly robust
face recognition systems for real time environment.
K-MEDOIDS CLUSTERING USING PARTITIONING AROUND MEDOIDS FOR PERFORMING FACE R...ijscmc
Face recognition is one of the most unobtrusive biometric techniques that can be used for access control as well as surveillance purposes. Various methods for implementing face recognition have been proposed with varying degrees of performance in different scenarios. The most common issue with effective facial biometric systems is high susceptibility of variations in the face owing to different factors like changes in pose, varying illumination, different expression, presence of outliers, noise etc. This paper explores a novel technique for face recognition by performing classification of the face images using unsupervised learning approach through K-Medoids clustering. Partitioning Around Medoids algorithm (PAM) has been used for performing K-Medoids clustering of the data. The results are suggestive of increased robustness to noise and outliers in comparison to other clustering methods. Therefore the technique can also be used to increase the overall robustness of a face recognition system and thereby increase its invariance and make it a reliably usable biometric modality
A Review Paper on Stereo Vision Based Depth EstimationIJSRD
Stereo vision is a challenging problem and it is a wide research topic in computer vision. It has got a lot of attraction because it is a cost efficient way in place of using costly sensors. Stereo vision has found a great importance in many fields and applications in today’s world. Some of the applications include robotics, 3-D scanning, 3-D reconstruction, driver assistance systems, forensics, 3-D tracking etc. The main challenge of stereo vision is to generate accurate disparity map. Stereo vision algorithms usually perform four steps: first, matching cost computation; second, cost aggregation; third, disparity computation or optimization; and fourth, disparity refinement. Stereo matching problems are also discussed. A large number of algorithms have been developed for stereo vision. But characterization of their performance has achieved less attraction. This paper gives a brief overview of the existing stereo vision algorithms. After evaluating the papers we can say that focus has been on cost aggregation and multi-step refinement process. Segment-based methods have also attracted attention due to their good performance. Also, using improved filter for cost aggregation in stereo matching achieves better results.
Hierarchical Approach for Total Variation Digital Image InpaintingIJCSEA Journal
The art of recovering an image from damage in an undetectable form is known as inpainting. The manual work of inpainting is most often a very time consum ing process. Due to digitalization of this technique, it is automatic and faster. In this paper, after the user selects the regions to be reconstructed, the algorithm automatically reconstruct the lost regions with the help of the information surrounding them. The existing methods perform very well when the region to be reconstructed is very small, but fails in proper reconstruction as the area increases. This paper describes a Hierarchical method by which the area to be inpainted is reduced in multiple levels and Total Variation(TV) method is used to inpaint in each level. This algorithm gives better performance when compared with other existing algorithms such as nearest neighbor interpolation, Inpainting through Blurring and Sobolev Inpainting.
Engineering Research Publication
Best International Journals, High Impact Journals,
International Journal of Engineering & Technical Research
ISSN : 2321-0869 (O) 2454-4698 (P)
www.erpublication.org
The improved hybrid model for molecular image denoising, proposed by NeST Software, can give a better SNR Molecular Image output. Read more on the proposed hybrid model.
Segmentation of medical images using metric topology – a region growing approachIjrdt Journal
A metric topological approach to the region growing based segmentation is presented in this article. Region based growing techniques has gained a significant importance in the medical image processing field for finest of segregation of tumor detected part in the image. Conventional algorithms were concentrated on segmentation at the coarser level which failed to produce enough evidence for the validity of the algorithm. In this article a novel technique is proposed based on metric topological neighbourhood also with the introduction of new objective measure entropy, apart from the traditional validity measures of Accuracy, PSNR and MSE. This measure is introduced to prove the amount of information lost after segmentation is reduced to greater extent which elucidates the effectiveness of the algorithm. This algorithm is tested on the well known benchmarking of testing in ground truth images in par with the proposed region based growing segmented images. The results validated show the validation of effectiveness of the algorithm.
Image restoration based on morphological operationsijcseit
Image processing including noise suppression, feature extraction, edge detection, image segmentation,
shape recognition, texture analysis, image restoration and reconstruction, image compression etc uses
mathematical morphology which is a method of nonlinear filters.
It is modulated from traditional morphology to order morphology, soft mathematical morphology and fuzzy
soft mathematical morphology. This paper is covers 6 morphological operations which are implemented in
the matlab program, including erosion, dilation, opening, closing, boundary extraction and region filling.
FINGERPRINT CLASSIFICATION BASED ON ORIENTATION FIELDijesajournal
ABSTRACT
This paper introduces an effective method of fingerprint classification based on discriminative feature gathering from orientation field. A nonlinear support vector machines (SVMs) is adopted for the classification. The orientation field is estimated through a pixel-Wise gradient descent method and the percentage of directional block classes is estimated. These percentages are classified into four-dimensional vector considered as a good feature that can be combined with an accurate singular point to classify the fingerprint into one of five classes. This method shows high classification accuracy relative to other spatial domain classifiers.
Face skin color based recognition using local spectral and gray scale featureseSAT Journals
Abstract Human face conveys more information about identity of person. Human face recognition is one of the most challenging problem and it can be used in many applications at different security places in airports, defense and banking sectors etc.In this work used colored features obtained from color segmentation because in real time scenario color provides the more information than gray scale image but it has a drawback. To overcome this drawback gray scale feature extracted from co-occurrence matrix of an image and for efficient face recognition of human Face under different illumination conditions spectral features can be extracted from face texture. These three feature vectors concatenated into a single feature vector and applied Lenc-Kral matching technique to measure similarity between the database and query image, the similarity is high then face is recognized. Keywords: Face recognition, illumination condition, local texture features, color segmentation.
STUDY ANALYSIS ON TEETH SEGMENTATION USING LEVEL SET METHODaciijournal
The three dimensional shape information of teeth from cone beam computed tomography images provides
important assistance for dentist performing implant treatment, orthodontic surgery. This paper describes
the tooth root of both anterior and posterior teeth from CBCT images of head. The segmentation is done
using level set method with five energy functions. The edge energy used to move the curve towards border
of the object. The shape prior energy provides the shape of the contour. The dentine wall energy provides
interaction between the neighboring teeth and prevent shrinkage and leakage problem. The test result for
both segmentation and 3D reconstruction shows that the method can visualize both anterior and posterior
teeth with high accuracy and efficiency.
Extraction of texture features by using gabor filter in wheat crop disease de...eSAT Journals
Abstract
Like country India, there are so many people depending upon agriculture. In this area, many farmers don’t know about new
diseases which are impacting on their farm. As the disease changes, the disease control policy also changes. So many farmers
have very sharp observation on crop diseases, but whenever there is new diseases fall on crops then problems occur. Climate also
changes instantly many of times, because of such reasons farmers unable to understand various diseases.
If farmer unable to predict that diseases quickly then it will affect life of crops. Indirectly it gets affects on total productivity of
farm. As we are well known about that world facing lot of problems due rapid growth in population. So our goal is to increase
agricultural productivity using image processing technology which can help farmer in great extent [7].
In this research work, we are trying that crop disease using Artificial neural network (ANN) which work very effectively. First of
all, we have provided an digital image which is taken by digital camera. That image given to Gaussian filter firstly then
transferred to adaptive median filter to filter out noise present inside image. Gaussian filter removes Gaussian noise which is
present inside image. Adaptive noise filter removes impulsive noise which is present inside image. Also it will reduce distortions
which are present inside images. Then image transferred to segmentation part. In image segmentation we have choose CIELAB
color space method to extract color components properly. For segmentation we have used Gabor filter. After this we distinguish
crop diseases on the basis of texture features which are extracted by Gabor filter [6].
Key Words: Artificial Neural Networks, Image preprocessing, Image Acquisition, and Feature Extraction,
classification etc…
Development of stereo matching algorithm based on sum of absolute RGB color d...IJECEIAES
This article presents local-based stereo matching algorithm which comprises a devel- opment of an algorithm using block matching and two edge preserving filters in the framework. Fundamentally, the matching process consists of several stages which will produce the disparity or depth map. The problem and most challenging work for matching process is to get an accurate corresponding point between two images. Hence, this article proposes an algorithm for stereo matching using improved Sum of Absolute RGB Differences (SAD), gradient matching and edge preserving filters. It is Bilateral Filter (BF) to surge up the accuracy. The SAD and gradient matching will be implemented at the first stage to get the preliminary corresponding result, then the BF works as an edge-preserving filter to remove the noise from the first stage. The second BF is used at the last stage to improve final disparity map and increase the object boundaries. The experimental analysis and validation are using the Middlebury standard benchmarking evaluation system. Based on the results, the proposed work is capable to increase the accuracy and to preserve the object edges. To make the proposed work more reliable with current available methods, the quantitative measurement has been made to compare with other existing methods and it shows the proposed work in this article perform much better.
Face Recognition Using Neural Network Based Fourier Gabor Filters & Random Pr...CSCJournals
Face detection and recognition has many applications in a variety of fields such as authentication, security, video surveillance and human interaction systems. In this paper, we present a neural network system for face recognition. Feature vector based on Fourier Gabor filters is used as input of our classifier, which is a Back Propagation Neural Network (BPNN). The input vector of the network will have large dimension, to reduce its feature subspace we investigate the use of the Random Projection as method of dimensionality reduction. Theory and experiment indicates the robustness of our solution.
LITERATURE SURVEY ON SPARSE REPRESENTATION FOR NEURAL NETWORK BASED FACE DETE...csijjournal
Face detection and recognition is a challenging problem in the field of image processing. In this paper, we reviewed some of the recent research works on face recognition. Issues with the previous face recognition
techniques are , time required is more for face recognition , recognition rate and database required to store the data . To overcome these problems sparse representation based classifier technique can be used.
FACE RECOGNITION USING PRINCIPAL COMPONENT ANALYSIS WITH MEDIAN FOR NORMALIZA...csandit
Recognizing Faces helps to name the various subjects present in the image. This work focuses
on labeling faces on an image which includes faces of humans being of various age group
(heterogeneous set ). Principal component analysis concentrates on finds the mean of the data
set and subtracts the mean value from the data set with an intention to normalize that data.
Normalization with respect to image is the removal of common features from the data set. This
work brings in the novel idea of deploying the median another measure of central tendency for
normalization rather than mean. The above work was implemented using matlab. Results show
that Median is the best measure for normalization for a heterogeneous data set which gives
raise to outliers.
Robust face recognition by applying partitioning around medoids over eigen fa...ijcsa
An unsupervised learning methodology for robust face recognition is proposed for enhancing invariance to
various changes in the face. The area of face recognition in spite of being the most unobtrusive biometric
modality of all has encountered challenges with high performance in uncontrolled environment owing to
frequently occurring, unavoidable variations in the face. These changes may be due to noise, outliers,
changing expressions, emotions, pose, illumination, facial distractions like makeup, spectacles, hair growth
etc. Methods for dealing with these variations have been developed in the past with different success.
However the cost and time efficiency play a crucial role in implementing any methodology in real world.
This paper presents a method to integrate the technique of Partitioning Around Medoids with Eigen Faces
and Fisher Faces to improve the efficiency of face recognition considerably. The system so designed has
higher resistance towards the impact of various changes in the face and performs well in terms of success
rate, cost involved and time complexity. The methodology can therefore be used in developing highly robust
face recognition systems for real time environment.
K-MEDOIDS CLUSTERING USING PARTITIONING AROUND MEDOIDS FOR PERFORMING FACE R...ijscmc
Face recognition is one of the most unobtrusive biometric techniques that can be used for access control as well as surveillance purposes. Various methods for implementing face recognition have been proposed with varying degrees of performance in different scenarios. The most common issue with effective facial biometric systems is high susceptibility of variations in the face owing to different factors like changes in pose, varying illumination, different expression, presence of outliers, noise etc. This paper explores a novel technique for face recognition by performing classification of the face images using unsupervised learning approach through K-Medoids clustering. Partitioning Around Medoids algorithm (PAM) has been used for performing K-Medoids clustering of the data. The results are suggestive of increased robustness to noise and outliers in comparison to other clustering methods. Therefore the technique can also be used to increase the overall robustness of a face recognition system and thereby increase its invariance and make it a reliably usable biometric modality
Face Recognition System Using Local Ternary Pattern and Signed Number Multipl...inventionjournals
This paper presents a novel approach to face recognition. The task of face recognition is to verify a claimed identity by comparing a claimed image of the individual with other images belonging to the same individual/other individual in a database. The proposed method utilizes Local Ternary Pattern and signed bit multiplication to extract local features of a face. The image is divided into small non-overlapping windows. Processing is carried out on these windows to extract features. Test image’s features are compared with all the training images using Euclidean's distance. The image with lowest Euclidean distance is recognized as the true face image. If the distance between test and all training images is more than threshold then test image is considered as unrecognised image or match not found .The face recognition rate of proposed system is calculated by varying the number of images per person in training database
K-Medoids Clustering Using Partitioning Around Medoids for Performing Face Re...ijscmcj
Face recognition is one of the most unobtrusive biometric techniques that can be used for access control as well as surveillance purposes. Various methods for implementing face recognition have been proposed with varying degrees of performance in different scenarios. The most common issue with effective facial biometric systems is high susceptibility of variations in the face owing to different factors like changes in pose, varying illumination, different expression, presence of outliers, noise etc. This paper explores a novel technique for face recognition by performing classification of the face images using unsupervised learning approach through K-Medoids clustering. Partitioning Around Medoids algorithm (PAM) has been used for performing K-Medoids clustering of the data. The results are suggestive of increased robustness to noise and outliers in comparison to other clustering methods. Therefore the technique can also be used to increase the overall robustness of a face recognition system and thereby increase its invariance and make it a reliably usable biometric modality.
Long-term Face Tracking in the Wild using Deep LearningElaheh Rashedi
This paper investigates long-term face tracking of a specific person given his/her face image in a single frame as a query in a video stream. Through taking advantage of pre-trained deep learning models on big data, a novel system is developed for accurate video face tracking in the unconstrained environments depicting various people and objects moving in and out of the frame. In the proposed system, we present a detection-verification-tracking method (dubbed as 'DVT') which accomplishes the long-term face tracking task through the collaboration of face detection, face verification, and (short-term) face tracking. An offline trained detector based on cascaded convolutional neural networks localizes all faces appeared in the frames, and an offline trained face verifier based on deep convolutional neural networks and similarity metric learning decides if any face or which face corresponds to the queried person. An online trained tracker follows the face from frame to frame. When validated on a sitcom episode and a TV show, the DVT method outperforms tracking-learning-detection (TLD) and face-TLD in terms of recall and precision. The proposed system is also tested on many other types of videos and shows very promising results.
Web image annotation by diffusion maps manifold learning algorithmijfcstjournal
Automatic image annotation is one of the most challenging problems in machine vision areas. The goal of this task is to predict number of keywords automatically for images captured in real data. Many methods are based on visual features in order to calculate similarities between image samples. But the computation cost of these approaches is very high. These methods require many training samples to be stored in memory. To lessen thisburden, a number of techniques have been developed to reduce the number
of features in a dataset. Manifold learning is a popular approach to nonlinear dimensionality reduction. In
this paper, we investigate Diffusion maps manifold learning method for webimage auto-annotation task.Diffusion maps
manifold learning method isused to reduce the dimension of some visual features. Extensive experiments and analysis onNUS-WIDE-LITE web image dataset with
different visual featuresshow how this manifold learning dimensionality reduction method can be applied effectively to image annotation.
ABSTRACT: a rigorous work on static and dynamic appearance based classification systems for face is on but, it is proving to be a challenging task for researchers to design a proper system since human face is complex one. Decades of work was and is focussed on how to classify a face and on how to increase the rate of classification but, little attention was paid to overcome redundancy in image classification. This paper presents a novel idea which focuses on redundancy check and its elimination. The paper after drawing inferences from previous work gives out a novel idea for exact face classification and elimination of redundancy.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Selective local binary pattern with convolutional neural network for facial ...IJECEIAES
Variation in images in terms of head pose and illumination is a challenge in facial expression recognition. This research presents a hybrid approach that combines the conventional and deep learning, to improve facial expression recognition performance and aims to solve the challenge. We propose a selective local binary pattern (SLBP) method to obtain a more stable image representation fed to the learning process in convolutional neural network (CNN). In the preprocessing stage, we use adaptive gamma transformation to reduce illumination variability. The proposed SLBP selects the discriminant features in facial images with head pose variation using the median-based standard deviation of local binary pattern images. We experimented on the Karolinska directed emotional faces (KDEF) dataset containing thousands of images with variations in head pose and illumination and Japanese female facial expression (JAFFE) dataset containing seven facial expressions of Japanese females’ frontal faces. The experiments show that the proposed method is superior compared to the other related approaches with an accuracy of 92.21% on KDEF dataset and 94.28% on JAFFE dataset.
Realtime face matching and gender prediction based on deep learningIJECEIAES
Face analysis is an essential topic in computer vision that dealing with human faces for recognition or prediction tasks. The face is one of the easiest ways to distinguish the identity people. Face recognition is a type of personal identification system that employs a person’s personal traits to determine their identity. Human face recognition scheme generally consists of four steps, namely face detection, alignment, representation, and verification. In this paper, we propose to extract information from human face for several tasks based on recent advanced deep learning framework. The proposed approach outperforms the results in the state-of-the-art.
Human face detection is a significant problem of
image processing and is usually a first step for face
recognition and visual surveillance. This paper presents the
details of face detection approach that is implemented to
achieve accurate face detection in group color images which
are based on facial feature and Support Vector Machine. In
the first step, the proposed approach quickly separates skin
color regions from the background and from non-skin color
regions using YCbCr color space transformation. After the
detection of skin regions, the images are processed with,
wavelet transforms (WT) and discrete cosine transforms
(DCT) as a result of which the 30×30 pixel sub images are
found. These sub images are then assigned to SVM classifier
as an input. The SVM is used to classify non-face regions from
the remaining regions more accurately, that are obtained
from previous steps and having big difference between faces
regions and non-faces regions. The experimental results on
different types of group color images show that this approach
improves the detection speed and minimizes the false
detection rate in less time and detects faces in different color
images.
DeepFace: Closing the Gap to Human-Level Performance in Face VerificationJoão Gabriel Lima
In modern face recognition, the conventional pipeline consists of four stages: detect ⇒ align ⇒ represent ⇒ clas- sify. We revisit both the alignment step and the representa- tion step by employing explicit 3D face modeling in order to apply a piecewise affine transformation, and derive a face representation from a nine-layer deep neural network. This deep network involves more than 120 million parameters using several locally connected layers without weight shar- ing, rather than the standard convolutional layers. Thus we trained it on the largest facial dataset to-date, an iden- tity labeled dataset of four million facial images belong- ing to more than 4,000 identities. The learned representa- tions coupling the accurate model-based alignment with the large facial database generalize remarkably well to faces in unconstrained environments, even with a simple classifier. Our method reaches an accuracy of 97.35% on the Labeled Faces in the Wild (LFW) dataset, reducing the error of the current state of the art by more than 27%, closely approach- ing human-level performance
Awarded presentation of my research activity, PhD Day 2011, February 23th 2011, Cagliari, Italy.
This presentation has been awarded as the best one of the track on information engineering.
Want to know more?
see my publications at
http://prag.diee.unica.it/pra/ita/people/satta
Similar to Learning a multi-center convolutional network for unconstrained face alignment (20)
Acorn Recovery: Restore IT infra within minutesIP ServerOne
Introducing Acorn Recovery as a Service, a simple, fast, and secure managed disaster recovery (DRaaS) by IP ServerOne. A DR solution that helps restore your IT infra within minutes.
Sharpen existing tools or get a new toolbox? Contemporary cluster initiatives...Orkestra
UIIN Conference, Madrid, 27-29 May 2024
James Wilson, Orkestra and Deusto Business School
Emily Wise, Lund University
Madeline Smith, The Glasgow School of Art
Have you ever wondered how search works while visiting an e-commerce site, internal website, or searching through other types of online resources? Look no further than this informative session on the ways that taxonomies help end-users navigate the internet! Hear from taxonomists and other information professionals who have first-hand experience creating and working with taxonomies that aid in navigation, search, and discovery across a range of disciplines.
0x01 - Newton's Third Law: Static vs. Dynamic AbusersOWASP Beja
f you offer a service on the web, odds are that someone will abuse it. Be it an API, a SaaS, a PaaS, or even a static website, someone somewhere will try to figure out a way to use it to their own needs. In this talk we'll compare measures that are effective against static attackers and how to battle a dynamic attacker who adapts to your counter-measures.
About the Speaker
===============
Diogo Sousa, Engineering Manager @ Canonical
An opinionated individual with an interest in cryptography and its intersection with secure software development.
This presentation, created by Syed Faiz ul Hassan, explores the profound influence of media on public perception and behavior. It delves into the evolution of media from oral traditions to modern digital and social media platforms. Key topics include the role of media in information propagation, socialization, crisis awareness, globalization, and education. The presentation also examines media influence through agenda setting, propaganda, and manipulative techniques used by advertisers and marketers. Furthermore, it highlights the impact of surveillance enabled by media technologies on personal behavior and preferences. Through this comprehensive overview, the presentation aims to shed light on how media shapes collective consciousness and public opinion.
This presentation by Morris Kleiner (University of Minnesota), was made during the discussion “Competition and Regulation in Professions and Occupations” held at the Working Party No. 2 on Competition and Regulation on 10 June 2024. More papers and presentations on the topic can be found out at oe.cd/crps.
This presentation was uploaded with the author’s consent.
Competition and Regulation in Professional Services – KLEINER – June 2024 OEC...
Learning a multi-center convolutional network for unconstrained face alignment
1. Learning a Multi-Center
Convolutional Network for
Unconstrained Face Alignment
Zhiwen Shao, Hengliang Zhu, Yangyang Hao,
Min Wang, and Lizhuang Ma
Shanghai Jiao Tong University
5. Methods based on low-level handcrafted features have a
limited capacity to represent highly complex faces
Deep convolutional network
A nonlinear regression problem, which transforms
appearance to shape
Motivation
6. Cascaded CNN [1], Zhou et al. [2], CFAN [3], and CDAN [4]
employ cascaded deep networks to refine predicted shapes
Previous Deep Learning Methods
time-consuming training processes
high model complexity
[1] Y. Sun, X. Wang, and X. Tang, “Deep convolutional network cascade for facial point
detection,” in IEEE Conference on Computer Vision and Pattern Recognition. IEEE, 2013, pp.
3476–3483.[2] E. Zhou, H. Fan, Z. Cao, Y. Jiang, and Q. Yin, “Extensive facial landmark localization with
coarse-to-fine convolutional network cascade,” in IEEE International Conference on Computer
Vision Workshops. IEEE, 2013, pp. 386–391.
[3] J. Zhang, S. Shan, M. Kan, and X. Chen, “Coarse-to-fine auto-encoder networks (cfan) for real-
time face alignment,” in European Conference on Computer Vision. Springer, 2014, pp. 1–16.
[4] R. Weng, J. Lu, Y.-P. Tan, and J. Zhou, “Learning cascaded deep auto-encoder networks for
face alignment,” IEEE Transactions on Multimedia, vol. 18, no. 10, pp. 2066–2078, 2016.
Multiple networks based
7. TCDCN [5] needs extra labels of facial attributes for
samples
one single network without auxiliary information
Previous Deep Learning Methods
limits the universality of this method
Single network based
[5] Z. Zhang, P. Luo, C. C. Loy, and X. Tang, “Learning deep representation for face alignment with
auxiliary attributes,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 38, no.
5, pp. 918–930, 2016.
9. Structural Correlations
Chin is occluded Right contour is invisible
Unconstrained faces with partial occlusion and large pose
Landmarks in the same local region have similar properties
including occlusion and visibility
10. Face Partition
29 landmarks 68 landmarks
Partition of facial landmarks for different labeling patterns
Left eye, right eye, nose, mouth, left contour, chin, and
right contour
12. Network Architecture
Shared layers
• Eight convolutional layers and one fully-connected layer
• Each max-pooling layer follows a stack of two convolutional layers
13. Network Architecture
• Each cluster of facial landmarks is treated as a separate center
• Each layer estimates x and y coordinates of all n facial landmarks
• Focusing on the shape estimation of a specific face region
Multiple center-specific shape prediction layers
14. Loss Function
^ ^
2 2 2
2 1 22 1 2
1
[( ) ( ) ]/ (2 )
n
j j jj j
j
E w f f f f d− −
=
= − + −∑
Weighted inter-ocular distance normalized Euclidean loss
jw weight of the j-th landmark
ground truth coordinatesf predicted coordinates
^
f
d ground truth inter-ocular distance
the first center-specific layer:
larger weights for landmarks around the left eye
21. Weight Computation
Multiple relationship
( ) ( )i c i m
P P
w wη=
( )i c
P set of center-specific landmarks
( )i m
P set of remaining minor landmarks
amplification factor
Different fine-tuning steps have different center-
specific and minor facial landmarks
Consistent with the basic model
( ) ( )
( ) ( )
| | ( | |)i c i m
i c i c
P P
w P w n P n+ − =
| |× number of elements in a set
During the i-th fine-tuning step
22. ( )
( )
( )
( )
/[( 1) | | ]
/[( 1) | | ]
i c
i m
i c
P
i c
P
w n P n
w n P n
η η
η
= − +
= − +
other centers with relatively small weights rather than
zeroutilize implicit structural correlations among different parts
landmarks from the same cluster have similar properties
share an identical weight
search the solution smoothly
Weight Computation
During the i-th fine-tuning step
23. Combined Model
high-level representation
( 1) 1
0 1( , , , ) ( 1024)T D
Dx x x D+ ×
= ∈ =x L ¡
weight matrix ( 1) 2
1 2 2( , , , ) D n
n
+ ×
= ∈W w w wL ¡
0 1( , , , ) , 1, ,2T
k k k Dkw w w k n= =w L L
^
2 12 1
^
22
T
jj
T
jj
f
f
−− =
=
w x
w x
weight matrix of the i-th center-specific layer
i
W
2 1 2 1
2 2
combined i
j j
combined i
j j
− −=
=
w w
w w
( )
1, , , i c
i m j P= ∈L
24. Combined Model
Combined Model S combined
Θ ∪ W
complexity is as same as the basic model
improves the location performance by
exploiting the advantage of each center-specific
solution
Our multi-center learning algorithm takes full advantage of each
stage and searches the optimal solution smoothly
26. Datasets
COFW
occluded dataset in the wild
1345 training images
507 testing images
IBUG
large appearance variations
3148 training images
135 testing images
27. Evaluation Metric
inter-ocular distance normalized mean error
cumulative errors distribution (CED) curves
failure rate
failure: mean error larger than 10%
28. Validation of Multi-Center Learning Algorithm
Method COFW IBUG
Mean Failure Mean Failure
Basic 6.26 3.16 9.23 33.33
Combined 6.08 2.96 8.87 25.93
Mean Error (%) and Failure Rate (%)
improve the accuracy and robustness
good performance of basic model
effectiveness of our network
reinforce the learning for each local face region
34. Comparison with Other Methods
Deep model Speed (FPS) CPU
Cascaded CNN 5 single core, i5-6200U 2.3GHz
CFAN* 43 i7-3770 3.4GHz
CDAN* 50 i5 3.2GHz
TCDCN 50 single core, i5-6200U 2.3GHz
CFT 31 single core, i5-6200U 2.3GHz
MCNet 67 single core, i5-6200U 2.3GHz
Time of face detection is excluded
35. Conclusions
We propose a novel multi-center convolutional network, which
exploits the representation power of each center
We propose the reinforcement for each center to improve the
shape estimation precision of each facial part
Comprehensive experiments demonstrate that our method
achieves real-time and competitive performance compared to
other state-of-the-art techniques
Good morning, everyone. I am Zhiwen Shao. I come from Shanghai Jiao Tong University. In our paper, we propose a Multi-Center Convolutional Network to achieve face alignment.
I first show the background of face alignment
These images illustrate the results of face alignment.
We can observe that these face images are very challenging. They have severe occlusions and large variations of pose, expression, illumination.
Our goal is to develop an efficient method to handle unconstrained faces
Face alignment can be regarded as a nonlinear regression problem, which transforms appearance to shape
Most conventional methods are based on low-level handcrafted features, so they have a limited capacity to represent complex faces
As we all know, a deep convolutional network has an outstanding representation ability. Therefore we use it to model the highly nonlinear function
There are two types of deep learning methods.
The first is multiple networks based.
These methods employ cascaded deep networks to refine predicted shapes successively.
Their training processes are complicated and time-consuming. And they have high computational cost and model complexity due to the use of multiple networks
A very typical method is TCDCN.
It trains only one deep network, but it needs extra labels of facial attributes for training samples.
This limits the universality of this method.
In contrast, our method uses one single network without auxiliary information
Next I introduce our method in details
Partial occlusion and large pose are main characteristics of unconstrained faces.
We discover that each facial landmark is not isolated but highly correlated with adjacent landmarks.
There are two examples.
In the left figure, facial landmarks along the chin are all occluded. And the right figure shows that landmarks on the right side of the face are almost invisible.
Therefore, landmarks in the same local face region have similar properties including occlusion and visibility.
We analyze the structure of a face, and partition it into seven clusters: left eye, right eye, nose, mouth, left contour, chin, and right contour.
As shown in these two figures, different labeling patterns of 29 and 68 facial landmarks are partitioned into 5 and 7 clusters respectively. Each cluster contains structurally relevant facial landmarks.
This is the structure of our multi-center convolutional network.
Our network consists of shared layers and multiple center-specific shape prediction layers.
The shared layers contain eight convolutional layers and one fully-connected layer.
Each max-pooling layer follows a stack of two convolutional layers
The stack of convolutional layers is excellent in feature learning, which is proposed by VGGNet.
According to the evaluation metric, we use weighted inter-ocular distance normalized Euclidean loss
We first pre-train a basic model with shared layers and one shape prediction layer.
Corresponding to Step 1
We further fine-tune each center-specific layer respectively
Corresponding to Step 2 to Step 6
Based on the pre-trained model, our network keeps shared layers and initializes each center-specific layer with the shape prediction parameters. There are m branches of center-specific layers at the end of our network. The fine-tuning of center-specific layers is mutually independent.
Shared layers and integrated shape prediction layer constitute the combined model
Corresponding to Step 7
We obtain the integrated shape prediction layer by combining corresponding parameters from each center-specific layer.
We assume there is a multiple relationship between two weights
To be consistent with the basic model, we keep weights conforming to this formula
The summation of weights is ensured to equal n
By solving two equations, we obtain the respective weights
When emphasizing on the detection of current center, we still consider other centers with relatively small weights rather than zero.
This is beneficial for utilizing implicit structural correlations among different facial parts and searching the solution smoothly
Then I show the experiments
Euclidean distance between two pupil centers
We show the mean error of each cluster for basic model and combined model on COFW dataset
It can be observed that the combined model improves the detection performance of each cluster
We report the results of our method MCNet and previous works.
We can see that our method outperforms most state-of-the-art methods
It is worth noting that TCDCN obtains better performance than our method on IBUG partly owing to their larger training data.
Although occlusions are not detected explicitly, we achieve an outstanding performance on par with Wu et al. on COFW benchmark.
We plot the CED curves for our method and several state-of-the-art methods.
It is observed that our method achieves competitive performance on both two benchmarks.
Our method achieves better performance for high-level normalized mean error. Therefore, our method is strongly robust to unconstrained environments.
There are several images from COFW
We can see our method indicates higher accuracy than RCPR and CFT in the details
Benefiting from utilizing structural correlations among different facial parts, our method is robust to severe occlusions.
We also show example images from IBUG where our method MCNet outperforms LBF and CFSS
Our method also achieves higher accuracy in the details. Therefore our method demonstrates superior capability of handling severe occlusions and complex variations of pose, expression, illumination.
To obtain a more comprehensive comparison, we present the average running speed of different deep learning methods for face alignment
We evaluate these methods on a single core i5-6200U 2.3GHz CPU with 1000 face images. Since CFAN and CDAN do not share their code, we use their published speed results.
Both TCDCN and our method MCNet are based on only one network, so they show relatively quick speed. Cascaded CNN, CFAN and CDAN employ multiple networks, so they cost more running time.
Our method only takes 15 ms on average to process one face, profiting from low model complexity and computational cost of our network. We believe that our method can be extended to real-time facial landmark tracking in unconstrained scenarios.