The document discusses tracking faces using active appearance models (AAMs). It provides an overview of modeling face shape and texture, and fitting the combined model to images using an iterative process. Specifically, it characterizes faces by locating them and representing them with a statistical model that captures variations in shape and texture with a small number of parameters. This model can then be fitted to new images by minimizing the difference between the synthetic and real face.
The document presents Active Appearance Models, which use principal component analysis to create a statistical model that captures appearance variations in images. It discusses how PCA is used to model shape and texture independently, then combined into a single model. The model can generate synthetic images and interpret new images by iteratively adjusting parameters to minimize differences between the input and generated images. The presenter shows the model can successfully converge and interpret images if initial parameter estimates are reasonable.
The document summarizes several papers on Active Shape Models and Active Appearance Models. It discusses how ASMs and AAMs are used to model object shape and appearance variations from a set of labeled training examples. Principal component analysis is applied to capture the main modes of shape and appearance variation within a training set. The models can then be used for tasks like image segmentation and pose estimation by searching for the optimal model parameters that match a new image.
This document provides an overview of using various MATLAB functions to analyze digital image data and pixel values, including histograms, contrast enhancement, and statistics. It discusses using imhist to display histograms, histeq and adapthisteq for contrast adjustment, impixel to view pixel values, improfile for intensity profiles along lines, and imcontour to create contour plots. Examples are given applying these functions to analyze and enhance grayscale and RGB images.
This document summarizes a presentation on digital image processing techniques. It discusses contrast-limited adaptive histogram equalization (CLAHE), histogram equalization, image intensity adjustment, adding noise, and median filtering. Methods like adapthisteq, histeq, imadjust, imnoise and medfilt2 are demonstrated to enhance contrast, adjust values, add noise and reduce salt and pepper noise through median filtering. Examples provided apply these techniques and compare results.
Statistical models of shape and appearancepotaters
This document provides an overview of statistical models for shape and appearance, including statistical shape models, combined appearance models, and matching algorithms like active shape models and active appearance models. It discusses how to build statistical shape models from training images by applying procedures like generalised Procrustes analysis to align shapes and then using principal component analysis to model shape variation. It also explains how to build combined models that correlate shape and texture parameters and how active appearance models work by iteratively updating shape and texture parameters to match a new image.
Digital Image Processing (Lab 09 and 10)Moe Moe Myint
The document discusses digital image processing using MATLAB. It covers topics like linear filtering, transforms, morphological operations and provides examples of using the dct2 and idct2 commands to compute the discrete cosine transform and inverse discrete cosine transform. It also demonstrates commands like imclose to morphologically close an image, imdilate to dilate an image, and imerode to erode an image providing syntax and examples for each. The document is presented by Dr. Moe Moe Myint from Technological University in Myanmar.
The document describes various image processing techniques in MATLAB, including image adjustment, morphological operations, dilation, erosion, opening, closing, and thresholding. Image adjustment maps pixel values to increase contrast. Morphological operations modify images based on shapes and neighborhoods. Dilation expands objects while erosion shrinks them. Opening removes small objects, while closing fills small holes. Thresholding converts an intensity image to binary.
This document presents a new iterative method for view and illumination invariant image matching. The method iteratively estimates the relationship between the relative view and illumination of two images, transforms one image to match the other's view, and normalizes their illumination for accurate matching. This is done by extracting features to estimate the transformation matrix between images, then estimating the illumination relationship and transforming one image's histogram to match the other. The performance is improved over traditional methods and is stable against large changes in view and illumination. The method could fail if the initial view and illumination estimates fail. It provides a new way to evaluate traditional feature detectors based on their valid angle and illumination change ranges.
The document presents Active Appearance Models, which use principal component analysis to create a statistical model that captures appearance variations in images. It discusses how PCA is used to model shape and texture independently, then combined into a single model. The model can generate synthetic images and interpret new images by iteratively adjusting parameters to minimize differences between the input and generated images. The presenter shows the model can successfully converge and interpret images if initial parameter estimates are reasonable.
The document summarizes several papers on Active Shape Models and Active Appearance Models. It discusses how ASMs and AAMs are used to model object shape and appearance variations from a set of labeled training examples. Principal component analysis is applied to capture the main modes of shape and appearance variation within a training set. The models can then be used for tasks like image segmentation and pose estimation by searching for the optimal model parameters that match a new image.
This document provides an overview of using various MATLAB functions to analyze digital image data and pixel values, including histograms, contrast enhancement, and statistics. It discusses using imhist to display histograms, histeq and adapthisteq for contrast adjustment, impixel to view pixel values, improfile for intensity profiles along lines, and imcontour to create contour plots. Examples are given applying these functions to analyze and enhance grayscale and RGB images.
This document summarizes a presentation on digital image processing techniques. It discusses contrast-limited adaptive histogram equalization (CLAHE), histogram equalization, image intensity adjustment, adding noise, and median filtering. Methods like adapthisteq, histeq, imadjust, imnoise and medfilt2 are demonstrated to enhance contrast, adjust values, add noise and reduce salt and pepper noise through median filtering. Examples provided apply these techniques and compare results.
Statistical models of shape and appearancepotaters
This document provides an overview of statistical models for shape and appearance, including statistical shape models, combined appearance models, and matching algorithms like active shape models and active appearance models. It discusses how to build statistical shape models from training images by applying procedures like generalised Procrustes analysis to align shapes and then using principal component analysis to model shape variation. It also explains how to build combined models that correlate shape and texture parameters and how active appearance models work by iteratively updating shape and texture parameters to match a new image.
Digital Image Processing (Lab 09 and 10)Moe Moe Myint
The document discusses digital image processing using MATLAB. It covers topics like linear filtering, transforms, morphological operations and provides examples of using the dct2 and idct2 commands to compute the discrete cosine transform and inverse discrete cosine transform. It also demonstrates commands like imclose to morphologically close an image, imdilate to dilate an image, and imerode to erode an image providing syntax and examples for each. The document is presented by Dr. Moe Moe Myint from Technological University in Myanmar.
The document describes various image processing techniques in MATLAB, including image adjustment, morphological operations, dilation, erosion, opening, closing, and thresholding. Image adjustment maps pixel values to increase contrast. Morphological operations modify images based on shapes and neighborhoods. Dilation expands objects while erosion shrinks them. Opening removes small objects, while closing fills small holes. Thresholding converts an intensity image to binary.
This document presents a new iterative method for view and illumination invariant image matching. The method iteratively estimates the relationship between the relative view and illumination of two images, transforms one image to match the other's view, and normalizes their illumination for accurate matching. This is done by extracting features to estimate the transformation matrix between images, then estimating the illumination relationship and transforming one image's histogram to match the other. The performance is improved over traditional methods and is stable against large changes in view and illumination. The method could fail if the initial view and illumination estimates fail. It provides a new way to evaluate traditional feature detectors based on their valid angle and illumination change ranges.
The document provides an overview of five fundamental machine learning algorithms: linear regression, logistic regression, decision tree learning, k-nearest neighbors, and neural networks. It describes the problem statement, solution, and key aspects of each algorithm. For linear regression, it discusses minimizing the squared error loss to find the optimal regression line. Logistic regression maximizes the likelihood function to find the optimal classification model. Decision tree learning uses an ID3 algorithm to greedily construct a non-parametric model by optimizing the average log-likelihood.
The document describes the implementation of eigenfaces for face recognition using MATLAB. It discusses preparing a training set of face images, calculating the mean image and eigenvectors, projecting new images onto the eigenface space, and recognizing faces by finding the closest match between eigenface coefficients. Key steps include normalizing images, computing the covariance matrix, selecting the most significant eigenvectors, and minimizing Euclidean distances to recognize faces.
Content Based Image Retrieval Using Full Haar SectorizationCSCJournals
Content based image retrieval (CBIR) deals with retrieval of relevant images from the large image database. It works on the features of images extracted. In this paper we are using very innovative idea of sectorization of Full Haar Wavelet transformed images for extracting the features into 4, 8, 12 and 16 sectors. The paper proposes two planes to be sectored i.e. Forward plane (Even plane) and backward plane (Odd plane). Similarity measure is also very essential part of CBIR which lets one to find the closeness of the query image with the database images. We have used two similarity measures namely Euclidean distance (ED) and sum of absolute difference (AD). The overall performance of retrieval of the algorithm has been measured by average precision and recall cross over point and LIRS, LSRR. The paper compares the performance of the methods with respect to type of planes, number of sectors, types of similarity measures and values of LIRS and LSRR.
The document discusses point processing operations in image processing which perform transformations independently on each pixel without considering spatial information. Point processing includes operations like negative, log, power-law transformations, and gamma correction that define a new image as a function of the existing image applied to each pixel. While point processing loses all spatial information, it can be used for basic image enhancement tasks like contrast stretching, histogram equalization, and matching.
This paper proposes a novel approach called R2P (Recomposition and Retargeting of Photographic Images) that can automatically alter the composition of an input source image to match a reference image, while also resizing the recomposed output image to fit the reference. The method first extracts foreground objects from the source and reference images. It then performs recomposition by solving a graph matching optimization to transfer the composition from the reference to the source. Finally, it jointly retargets the recomposed source image by optimizing a mesh warping technique to minimize distortion while fitting the size of the reference image. The approach requires no user input, pre-collected training data, or predetermined composition rules.
This document proposes a new algorithm called 3D continuous dynamic programming (3DCDP) for segmentation-free registration and optimal matching of 3D voxel patterns. 3DCDP allows for matching of full voxels inside 3D objects, not just surfaces. It matches a reference 3D image to parts of an input image in a segmentation-free manner. The algorithm extends 2D continuous dynamic programming to 3D by combining three 2DCDP planes. Experiments show it can accurately match a reference image embedded within an input image.
Eugen Zaharescu-STATEMENT OF RESEARCH INTERESTEugen Zaharescu
- The document is a research statement from Dr. Eugen ZAHARESCU that outlines his interests in mathematical morphology, image analysis, and ontology generation.
- His research has included extending mathematical morphology theory to multivariate images and exploring morphological operators in logarithmic image processing.
- More recently, he has developed algorithms and tools for machine learning, computer vision, and image understanding by applying mathematical concepts from morphology.
This document discusses face recognition and the PCA algorithm for face recognition. It begins with an introduction to face recognition and its uses. It then explains the PCA algorithm for face recognition in 6 steps: 1) converting images to vectors, 2) normalizing the vectors, 3) calculating eigenvectors from the normalized vectors, 4) selecting important eigenvectors, 5) representing faces as combinations of eigenvectors, and 6) recognizing faces. It discusses the strengths and weaknesses of face recognition and lists several applications such as access control, law enforcement, and banking.
This document discusses face recognition using principal component analysis (PCA). It begins by defining face recognition and distinguishing it from face detection. It then outlines the steps of the PCA algorithm for face recognition, including representing faces as vectors, calculating the average face vector, normalizing faces, calculating the covariance matrix, selecting eigenvectors to reduce dimensionality, projecting faces into the reduced eigenface space, and representing faces as a linear combination of eigenfaces. The document focuses on explaining the PCA algorithm and its steps for performing eigenface-based face recognition.
This document discusses variations of the interval linear assignment problem. It begins with an introduction to assignment problems and defines them as problems that assign resources to activities to minimize cost or maximize profit on a one-to-one basis. It then provides the mathematical model for standard assignment problems and discusses variations such as non-square matrices, maximization/minimization objectives, constrained assignments, and alternate optimal solutions. The document also gives examples of managerial applications and provides two numerical examples solving interval linear assignment problems using an interval Hungarian method.
This document compares two image inpainting algorithms: the Fast Marching Method (FMM) and exemplar-based image inpainting. FMM uses structural consistency to fill damaged regions, while exemplar-based uses both structural and textural consistency. FMM is faster but does not preserve edges as well as exemplar-based. Exemplar-based works for both small and large regions but is slower. Both algorithms were tested on photos for tasks like removing objects or adding effects. Exemplar-based was better for large regions and edge preservation, while FMM was better for speed and small regions.
SURVEY ON POLYGONAL APPROXIMATION TECHNIQUES FOR DIGITAL PLANAR CURVESZac Darcy
This document summarizes and compares three techniques for polygonal approximation of digital planar curves:
1) Masood's technique which iteratively deletes redundant points and uses a stabilization process to optimize point locations.
2) Carmona's technique which suppresses redundant points using a breakpoint suppression algorithm and threshold.
3) Tanvir's adaptive optimization algorithm which focuses on high curvature points and applies an optimization procedure.
The techniques are evaluated on standard shapes using measures like number of points, compression ratio, error, and weighted error. Masood's technique generally had lower error while Tanvir's often achieved the highest compression.
Contrast enhancement using various statistical operations and neighborhood pr...sipij
This document proposes a novel contrast enhancement algorithm using various statistical operations and neighborhood processing. It begins with an overview of histogram equalization and some of its limitations. It then discusses related work on other histogram equalization techniques including classical histogram equalization, brightness preserving bi-histogram equalization, recursive mean separate histogram equalization, and background brightness preserving histogram equalization. The proposed method is then described, which applies statistical operations like mean and standard deviation within a neighborhood to locally enhance pixels. Pixels are replaced from an initially equalized image if their difference from the local mean exceeds a threshold. This aims to preserve local brightness features. Finally, metrics for evaluating image quality like PSNR, SSIM, and CNR are defined to analyze results
Determination of Optimal Product Mix for Profit Maximization using Linear Pro...IJERA Editor
This paper demonstrates the use of liner programming methods in order to determine the optimal product mix for
profit maximization. There had been several papers written to demonstrate the use of linear programming in
finding the optimal product mix in various organization. This paper is aimed to show the generic approach to be
taken to find the optimal product mix.
ABSTRACT: a rigorous work on static and dynamic appearance based classification systems for face is on but, it is proving to be a challenging task for researchers to design a proper system since human face is complex one. Decades of work was and is focussed on how to classify a face and on how to increase the rate of classification but, little attention was paid to overcome redundancy in image classification. This paper presents a novel idea which focuses on redundancy check and its elimination. The paper after drawing inferences from previous work gives out a novel idea for exact face classification and elimination of redundancy.
The document discusses using machine learning algorithms like Support Vector Machines (SVM) for classification and Support Vector Regression (SVR) for regression on facial image data. Dimensionality reduction using Locality Preserving Projections is also discussed to reduce computational requirements. SVM classification of gender on a subset of 3000 images achieved over 99% accuracy. SVR is noted to better handle outliers in facial data compared to basic linear regression due to minimizing slope. The goal is to classify gender and regress age from a set of facial images.
This poster presents a feature-level fusion approach for face and palmprint biometrics using improved K-medoids clustering and isometric graph representations. SIFT features are extracted from face and palmprint images and clustered using an improved K-medoids algorithm. Correspondences between feature points are established and represented as an isometric graph. Fused matching scores are obtained using KNN and correlation distances, exhibiting robust performance and increased accuracy over single biometrics.
International Journal of Computational Engineering Research(IJCER)ijceronline
International Journal of Computational Engineering Research(IJCER) is an intentional online Journal in English monthly publishing journal. This Journal publish original research work that contributes significantly to further the scientific knowledge in engineering and Technology.
This is a light-hearted, non-technical presentation about the current state of the art in Artificial Intelligence, particularly the field of Neural Networks and Deep Learning. This talk was presented for the general public in Mission Creak festival in Iowa City, IA on March 3, 2015
Active Appearance Models (AAMs) combine shape and texture models into a single statistical model. AAMs are trained on labeled images to learn the relationship between shape, texture, and model parameters. To interpret a new image, an optimization problem is solved to minimize the difference between the image and one synthesized by the AAM. The model parameters are updated iteratively based on a linear model trained to relate parameter adjustments to image differences. Constrained AAMs incorporate prior information to improve the influence of the starting approximation on the search results.
Computation of Hydrodynamic Characteristics of Ships using CFDNabila Naz
1) The document summarizes research using computational fluid dynamics (CFD) to analyze hydrodynamic characteristics like wave patterns, resistance, and pressure around ship hulls.
2) CFD simulations were conducted using the SHIPFLOW software to model potential, boundary layer, and viscous flow around two ship hulls at varying speeds.
3) Results for wave elevation, resistance coefficients, and streamlines showed good agreement with experimental data, though some discrepancies remained, especially near bow and stern.
Hydrodynamic Modeling of the Physical Dispersion of Radium-Enriched Barite Ai...Donald Carpenter
This document discusses the formation and transport of radium-enriched barite from oil and gas operations. Barite forms as a byproduct and can incorporate radium, becoming a naturally occurring radioactive material (NORM). Millions of barrels of NORM-impacted barite are generated each year. Surface gamma scans can detect radium contamination, but may be complicated over water or with obstructions. Hydrodynamic and particle tracking models can predict how barite will disperse based on its high density and other properties compared to other materials like quartz. Geomorphic assessments that consider landscape features can identify optimal locations to sample for barite accumulation and radium levels in a more efficient manner than random scanning.
The document provides an overview of five fundamental machine learning algorithms: linear regression, logistic regression, decision tree learning, k-nearest neighbors, and neural networks. It describes the problem statement, solution, and key aspects of each algorithm. For linear regression, it discusses minimizing the squared error loss to find the optimal regression line. Logistic regression maximizes the likelihood function to find the optimal classification model. Decision tree learning uses an ID3 algorithm to greedily construct a non-parametric model by optimizing the average log-likelihood.
The document describes the implementation of eigenfaces for face recognition using MATLAB. It discusses preparing a training set of face images, calculating the mean image and eigenvectors, projecting new images onto the eigenface space, and recognizing faces by finding the closest match between eigenface coefficients. Key steps include normalizing images, computing the covariance matrix, selecting the most significant eigenvectors, and minimizing Euclidean distances to recognize faces.
Content Based Image Retrieval Using Full Haar SectorizationCSCJournals
Content based image retrieval (CBIR) deals with retrieval of relevant images from the large image database. It works on the features of images extracted. In this paper we are using very innovative idea of sectorization of Full Haar Wavelet transformed images for extracting the features into 4, 8, 12 and 16 sectors. The paper proposes two planes to be sectored i.e. Forward plane (Even plane) and backward plane (Odd plane). Similarity measure is also very essential part of CBIR which lets one to find the closeness of the query image with the database images. We have used two similarity measures namely Euclidean distance (ED) and sum of absolute difference (AD). The overall performance of retrieval of the algorithm has been measured by average precision and recall cross over point and LIRS, LSRR. The paper compares the performance of the methods with respect to type of planes, number of sectors, types of similarity measures and values of LIRS and LSRR.
The document discusses point processing operations in image processing which perform transformations independently on each pixel without considering spatial information. Point processing includes operations like negative, log, power-law transformations, and gamma correction that define a new image as a function of the existing image applied to each pixel. While point processing loses all spatial information, it can be used for basic image enhancement tasks like contrast stretching, histogram equalization, and matching.
This paper proposes a novel approach called R2P (Recomposition and Retargeting of Photographic Images) that can automatically alter the composition of an input source image to match a reference image, while also resizing the recomposed output image to fit the reference. The method first extracts foreground objects from the source and reference images. It then performs recomposition by solving a graph matching optimization to transfer the composition from the reference to the source. Finally, it jointly retargets the recomposed source image by optimizing a mesh warping technique to minimize distortion while fitting the size of the reference image. The approach requires no user input, pre-collected training data, or predetermined composition rules.
This document proposes a new algorithm called 3D continuous dynamic programming (3DCDP) for segmentation-free registration and optimal matching of 3D voxel patterns. 3DCDP allows for matching of full voxels inside 3D objects, not just surfaces. It matches a reference 3D image to parts of an input image in a segmentation-free manner. The algorithm extends 2D continuous dynamic programming to 3D by combining three 2DCDP planes. Experiments show it can accurately match a reference image embedded within an input image.
Eugen Zaharescu-STATEMENT OF RESEARCH INTERESTEugen Zaharescu
- The document is a research statement from Dr. Eugen ZAHARESCU that outlines his interests in mathematical morphology, image analysis, and ontology generation.
- His research has included extending mathematical morphology theory to multivariate images and exploring morphological operators in logarithmic image processing.
- More recently, he has developed algorithms and tools for machine learning, computer vision, and image understanding by applying mathematical concepts from morphology.
This document discusses face recognition and the PCA algorithm for face recognition. It begins with an introduction to face recognition and its uses. It then explains the PCA algorithm for face recognition in 6 steps: 1) converting images to vectors, 2) normalizing the vectors, 3) calculating eigenvectors from the normalized vectors, 4) selecting important eigenvectors, 5) representing faces as combinations of eigenvectors, and 6) recognizing faces. It discusses the strengths and weaknesses of face recognition and lists several applications such as access control, law enforcement, and banking.
This document discusses face recognition using principal component analysis (PCA). It begins by defining face recognition and distinguishing it from face detection. It then outlines the steps of the PCA algorithm for face recognition, including representing faces as vectors, calculating the average face vector, normalizing faces, calculating the covariance matrix, selecting eigenvectors to reduce dimensionality, projecting faces into the reduced eigenface space, and representing faces as a linear combination of eigenfaces. The document focuses on explaining the PCA algorithm and its steps for performing eigenface-based face recognition.
This document discusses variations of the interval linear assignment problem. It begins with an introduction to assignment problems and defines them as problems that assign resources to activities to minimize cost or maximize profit on a one-to-one basis. It then provides the mathematical model for standard assignment problems and discusses variations such as non-square matrices, maximization/minimization objectives, constrained assignments, and alternate optimal solutions. The document also gives examples of managerial applications and provides two numerical examples solving interval linear assignment problems using an interval Hungarian method.
This document compares two image inpainting algorithms: the Fast Marching Method (FMM) and exemplar-based image inpainting. FMM uses structural consistency to fill damaged regions, while exemplar-based uses both structural and textural consistency. FMM is faster but does not preserve edges as well as exemplar-based. Exemplar-based works for both small and large regions but is slower. Both algorithms were tested on photos for tasks like removing objects or adding effects. Exemplar-based was better for large regions and edge preservation, while FMM was better for speed and small regions.
SURVEY ON POLYGONAL APPROXIMATION TECHNIQUES FOR DIGITAL PLANAR CURVESZac Darcy
This document summarizes and compares three techniques for polygonal approximation of digital planar curves:
1) Masood's technique which iteratively deletes redundant points and uses a stabilization process to optimize point locations.
2) Carmona's technique which suppresses redundant points using a breakpoint suppression algorithm and threshold.
3) Tanvir's adaptive optimization algorithm which focuses on high curvature points and applies an optimization procedure.
The techniques are evaluated on standard shapes using measures like number of points, compression ratio, error, and weighted error. Masood's technique generally had lower error while Tanvir's often achieved the highest compression.
Contrast enhancement using various statistical operations and neighborhood pr...sipij
This document proposes a novel contrast enhancement algorithm using various statistical operations and neighborhood processing. It begins with an overview of histogram equalization and some of its limitations. It then discusses related work on other histogram equalization techniques including classical histogram equalization, brightness preserving bi-histogram equalization, recursive mean separate histogram equalization, and background brightness preserving histogram equalization. The proposed method is then described, which applies statistical operations like mean and standard deviation within a neighborhood to locally enhance pixels. Pixels are replaced from an initially equalized image if their difference from the local mean exceeds a threshold. This aims to preserve local brightness features. Finally, metrics for evaluating image quality like PSNR, SSIM, and CNR are defined to analyze results
Determination of Optimal Product Mix for Profit Maximization using Linear Pro...IJERA Editor
This paper demonstrates the use of liner programming methods in order to determine the optimal product mix for
profit maximization. There had been several papers written to demonstrate the use of linear programming in
finding the optimal product mix in various organization. This paper is aimed to show the generic approach to be
taken to find the optimal product mix.
ABSTRACT: a rigorous work on static and dynamic appearance based classification systems for face is on but, it is proving to be a challenging task for researchers to design a proper system since human face is complex one. Decades of work was and is focussed on how to classify a face and on how to increase the rate of classification but, little attention was paid to overcome redundancy in image classification. This paper presents a novel idea which focuses on redundancy check and its elimination. The paper after drawing inferences from previous work gives out a novel idea for exact face classification and elimination of redundancy.
The document discusses using machine learning algorithms like Support Vector Machines (SVM) for classification and Support Vector Regression (SVR) for regression on facial image data. Dimensionality reduction using Locality Preserving Projections is also discussed to reduce computational requirements. SVM classification of gender on a subset of 3000 images achieved over 99% accuracy. SVR is noted to better handle outliers in facial data compared to basic linear regression due to minimizing slope. The goal is to classify gender and regress age from a set of facial images.
This poster presents a feature-level fusion approach for face and palmprint biometrics using improved K-medoids clustering and isometric graph representations. SIFT features are extracted from face and palmprint images and clustered using an improved K-medoids algorithm. Correspondences between feature points are established and represented as an isometric graph. Fused matching scores are obtained using KNN and correlation distances, exhibiting robust performance and increased accuracy over single biometrics.
International Journal of Computational Engineering Research(IJCER)ijceronline
International Journal of Computational Engineering Research(IJCER) is an intentional online Journal in English monthly publishing journal. This Journal publish original research work that contributes significantly to further the scientific knowledge in engineering and Technology.
This is a light-hearted, non-technical presentation about the current state of the art in Artificial Intelligence, particularly the field of Neural Networks and Deep Learning. This talk was presented for the general public in Mission Creak festival in Iowa City, IA on March 3, 2015
Active Appearance Models (AAMs) combine shape and texture models into a single statistical model. AAMs are trained on labeled images to learn the relationship between shape, texture, and model parameters. To interpret a new image, an optimization problem is solved to minimize the difference between the image and one synthesized by the AAM. The model parameters are updated iteratively based on a linear model trained to relate parameter adjustments to image differences. Constrained AAMs incorporate prior information to improve the influence of the starting approximation on the search results.
Computation of Hydrodynamic Characteristics of Ships using CFDNabila Naz
1) The document summarizes research using computational fluid dynamics (CFD) to analyze hydrodynamic characteristics like wave patterns, resistance, and pressure around ship hulls.
2) CFD simulations were conducted using the SHIPFLOW software to model potential, boundary layer, and viscous flow around two ship hulls at varying speeds.
3) Results for wave elevation, resistance coefficients, and streamlines showed good agreement with experimental data, though some discrepancies remained, especially near bow and stern.
Hydrodynamic Modeling of the Physical Dispersion of Radium-Enriched Barite Ai...Donald Carpenter
This document discusses the formation and transport of radium-enriched barite from oil and gas operations. Barite forms as a byproduct and can incorporate radium, becoming a naturally occurring radioactive material (NORM). Millions of barrels of NORM-impacted barite are generated each year. Surface gamma scans can detect radium contamination, but may be complicated over water or with obstructions. Hydrodynamic and particle tracking models can predict how barite will disperse based on its high density and other properties compared to other materials like quartz. Geomorphic assessments that consider landscape features can identify optimal locations to sample for barite accumulation and radium levels in a more efficient manner than random scanning.
Integrated hydrodynamic and structural analysis webinar presentation tcm4 601490Dai Hung
This document summarizes a webinar on integrated hydrodynamic and structural analysis using the Sesam software. It discusses:
1) An overview of Sesam and its capabilities for hydrostatic, hydrodynamic, and structural analysis of floating structures.
2) A case study analyzing the loads and stresses on an FPSO at different loading conditions, comparing results from the quasi-static and full dynamic solvers. The dynamic solver produced lower pressures, loads, and stresses.
3) Techniques for ensuring accurate load transfer between the hydrodynamic and structural models, including pressure scaling near the waterline and checking the global load balance.
DSD-INT 2016 Hydrodynamic modeling and resource-device suitability analysis o...Deltares
Presentation by Oliver Dan de Luna, University of the Philippines - Marine Science Institute, Philippines, at the Delft3D User Days during Delft Software Days 2016 on Tuesday, 1 November 2016, Delft.
This document reviews techniques for emotion recognition from facial expressions. It begins by outlining the general steps of emotion recognition systems as face detection, feature extraction, and classification. Popular techniques discussed include principal component analysis (PCA), local binary patterns (LBP), active appearance models, and Haar classifiers. PCA and LBP were found to provide higher recognition rates. The paper also reviews the Facial Action Coding System and compares the performance of different techniques based on recognition rate. In conclusion, PCA is identified as having the highest recognition rate and performance for emotion recognition.
This document provides an introduction and overview of a fluid mechanics course taught by Dr. Mohsin Siddique. It outlines the course details including goals, topics, textbook, and assessment methods. The course aims to provide an understanding of fluid statics and dynamics concepts. Key topics covered include fluid properties, fluid statics, fluid flow measurements, dimensional analysis, and fluid flow in pipes and open channels. Students will be evaluated through assignments, quizzes, a midterm exam, and a final exam. The course intends to develop skills relevant to various engineering fields involving fluid mechanics.
This document provides an overview of artificial neural networks (ANN). It discusses the origin of ANNs from biological neural networks. It describes different ANN architectures like multilayer perceptrons and different learning methods like backpropagation. It also outlines some challenging problems that ANNs can help with, such as pattern recognition, clustering, and optimization. The summary states that while the paper gives a good overview of ANNs, more development is needed to show ANNs are better than other methods for most problems.
MIXTURES OF TRAINED REGRESSION CURVES MODELS FOR HANDWRITTEN ARABIC CHARACTER...gerogepatton
In this paper, we demonstrate how regression curves can be used to recognize 2D non-rigid handwritten shapes. Each shape is represented by a set of non-overlapping uniformly distributed landmarks. The underlying models utilize 2nd order of polynomials to model shapes within a training set. To estimate the regression models, we need to extract the required coefficients which describe the variations for a set of shape class. Hence, a least square method is used to estimate such modes. We proceed then, by training these coefficients using the apparatus Expectation Maximization algorithm. Recognition is carried out by finding the least error landmarks displacement with respect to the model curves. Handwritten isolated Arabic characters are used to evaluate our approach.
MIXTURES OF TRAINED REGRESSION CURVESMODELS FOR HANDRITTEN ARABIC CHARACTER R...ijaia
In this paper, we demonstrate how regression curves can be used to recognize 2D non-rigid handwritten shapes. Each shape is represented by a set of non-overlapping uniformly distributed landmarks. The underlying models utilize 2nd order of polynomials to model shapes within a training set. To estimate the regression models, we need to extract the required coefficients which describe the variations for a set of shape class. Hence, a least square method is used to estimate such modes. We proceed then, by training these coefficients using the apparatus Expectation Maximization algorithm. Recognition is carried out by finding the least error landmarks displacement with respect to the model curves. Handwritten isolated Arabic characters are used to evaluate our approach.
Human’s facial parts extraction to recognize facial expressionijitjournal
Real-time facial expression analysis is an important yet challenging task in human computer interaction.
This paper proposes a real-time person independent facial expression recognition system using a
geometrical feature-based approach. The face geometry is extracted using the modified active shape
model. Each part of the face geometry is effectively represented by the Census Transformation (CT) based
feature histogram. The facial expression is classified by the SVM classifier with exponential chi-square
weighted merging kernel. The proposed method was evaluated on the JAFFE database and in real-world
environment. The experimental results show that the approach yields a high recognition rate and is
applicable in real-time facial expression analysis.
Face detection using the 3 x3 block rank patterns of gradient magnitude imagessipij
Face detection locates faces prior to various face-
related applications. The objective of face detecti
on is to
determine whether or not there are any faces in an
image and, if any, the location of each face is det
ected.
Face detection in real images is challenging due to
large variability of illumination and face appeara
nces.
This paper proposes a face detection algorithm usin
g the 3×3 block rank patterns of gradient magnitude
images and a geometrical face model. First, the ill
umination-corrected image of the face region is obt
ained
using the brightness plane that is produced using t
he locally minimum brightness of each block. Next,
the
illumination-corrected image is histogram equalized
, the face region is divided into nine (3×3) blocks
, and
two directional (horizontal and vertical) gradient
magnitude images are computed, from which the 3×3
block rank patterns are obtained. For face detectio
n, using the FERET and GT databases three types of
the
3×3 block rank patterns are a priori determined as
templates based on the distribution of the sum of t
he
gradient magnitudes of each block in the face candi
date region that is also composed of nine (3×3) blo
cks.
The 3×3 block rank patterns roughly classify whethe
r the detected face candidate region contains a fac
e or
not. Finally, facial features are detected and used
to validate the face model. The face candidate is
validated as a face if it is matched with the geome
trical face model. The proposed algorithm is tested
on the
Caltech database images and real images. Experiment
al results with a number of test images show the
effectiveness of the proposed algorithm.
Large sample property of the bayes factor in a spline semiparametric regressi...Alexander Decker
This document summarizes a research paper about investigating the large sample property of the Bayes factor for testing the polynomial component of a spline semiparametric regression model against a fully spline alternative model. It considers a semiparametric regression model where the mean function has two parts - a parametric linear component and a nonparametric penalized spline component. By representing the model as a mixed model, it obtains the closed form of the Bayes factor and proves that the Bayes factor is consistent under certain conditions on the prior and design matrix. It establishes that the Bayes factor converges to infinity under the pure polynomial model and converges to zero almost surely under the spline semiparametric alternative model.
Face Recognition Using Gabor features And PCAIOSR Journals
This document summarizes a research paper on face recognition using Gabor features and principal component analysis (PCA). It begins by providing background on face recognition and discusses challenges like lighting, pose, and orientation. It then describes preprocessing faces using Gabor wavelets to extract discriminative features and reduce variations. PCA is used to further reduce the dimensionality of features into principal components. These components are used for classification, with nearest neighbor classification tested on the Yale face database. Results show the proposed approach improves recognition rates compared to Euclidean distance measures.
This paper proposes a method to jointly match multiple 3D meshes by maximizing pairwise feature affinities and cycle consistency across models. It formulates the matching problem as a low-rank matrix recovery problem and uses nuclear norm relaxation for rank minimization. An alternating minimization algorithm is used to efficiently solve the optimization problem. Experimental results show the method provides an order of magnitude speed-up compared to state-of-the-art algorithms based on semi-definite programming, while achieving competitive performance. It also introduces a distortion term to the pairwise matching to help match reflexive sub-parts of models distinctly.
Joint3DShapeMatching - a fast approach to 3D model matching using MatchALS 3...Mamoon Ismail Khalid
we extend the global optimization-based
approach of jointly matching a set of images to jointly
matching a set of 3D meshes. The estimated correspon
dences simultaneously maximize pairwise feature affini
ties and cycle consistency across multiple models. We
show that the low-rank matrix recovery problem can be
efficiently applied to the 3D meshes as well. The fast
alternating minimization algorithm helps to handle real
world practical problems with thousands of features. Ex
perimental results show that, unlike the state-of-the-art
algorithm which rely on semi-definite programming, our
algorithm provides an order of magnitude speed-up along
with competitive performance. Along with the joint shape
matching we propose an approach to apply a distortion
term in pairwise matching, which helps in successfully
matching the reflexive sub-parts of two models distinc
tively. In the end, we demonstrate the applicability of
the algorithm to match a set of 3D meshes of the SCAPE
benchmark database
Unimodal Multi-Feature Fusion and one-dimensional Hidden Markov Models for Lo...IJECEIAES
The objective of low-resolution face recognition is to identify faces from small size or poor quality images with varying pose, illumination, expression, etc. In this work, we propose a robust low face recognition technique based on one-dimensional Hidden Markov Models. Features of each facial image are extracted using three steps: firstly, both Gabor filters and Histogram of Oriented Gradients (HOG) descriptor are calculated. Secondly, the size of these features is reduced using the Linear Discriminant Analysis (LDA) method in order to remove redundant information. Finally, the reduced features are combined using Canonical Correlation Analysis (CCA) method. Unlike existing techniques using HMMs, in which authors consider each state to represent one facial region (eyes, nose, mouth, etc), the proposed system employs 1D-HMMs without any prior knowledge about the localization of interest regions in the facial image. Performance of the proposed method will be measured using the AR database.
This document describes a framework for 2D pose estimation using active shape models and learned entropy field approximations. A dataset of manually annotated poses was created from NBA footage to train the models. Active shape models use principal component analysis to represent poses as a linear combination of modes of variation learned from the training data. To evaluate pose likelihood, image entropy is proposed as a texture similarity measure and regression is used to learn a function mapping poses to entropy fields, which can be compared to the image entropy. Current results are presented and future work to improve and speed up the approach is discussed.
This document describes a project that uses photometric stereo to reconstruct 3D surfaces from images taken under different lighting conditions on a computer screen. Photometric stereo uses variations in pixel intensities across images to estimate surface normals and reconstruct the 3D shape. The project creates a MATLAB program that performs photometric stereo in real-time by flashing different light patterns on a screen and capturing images with a webcam. By using singular value decomposition, the program can reconstruct surfaces without knowing the exact positions of the light sources, overcoming a limitation of traditional photometric stereo. The reconstruction contains noise but demonstrates photometric stereo in a less controlled environment. Further work could explore tradeoffs of the method and improve efficiency.
This document presents a novel approach for measuring shape similarity and using it for object recognition. The key steps are:
1) Solving the correspondence problem between two shapes by attaching a descriptor called "shape context" to sample points on each shape. Shape context captures the distribution of remaining points relative to the reference point.
2) Using the point correspondences to estimate an aligning transformation between the shapes. This provides a measure of shape similarity as the matching error between corresponding points plus the magnitude of the transformation.
3) Treating recognition as a nearest neighbor problem to find the most similar stored prototype shape. The approach is demonstrated on various datasets including handwritten digits, silhouettes, and 3D objects
The document summarizes two incremental smoothing and mapping algorithms, iSAM and iSAM2. iSAM uses matrix factorization to efficiently update the information matrix when new measurements are obtained. However, periodic batch steps are required to avoid "fill-in", reducing efficiency. iSAM2 uses a novel Bayes tree data structure to represent the factor graph, allowing incremental updates to be made by only modifying the relevant portions of the tree. This provides an exact, incremental solution without periodic batch steps, making it more efficient than iSAM. The document provides examples of applying iSAM2 to a range odometry SLAM problem and a structure from motion application.
GENETIC ALGORITHM FOR FUNCTION APPROXIMATION: AN EXPERIMENTAL INVESTIGATIONijaia
Function Approximation is a popular engineering problems used in system identification or Equation
optimization. Due to the complex search space it requires, AI techniques has been used extensively to spot
the best curves that match the real behavior of the system. Genetic algorithm is known for their fast
convergence and their ability to find an optimal structure of the solution. We propose using a genetic
algorithm as a function approximator. Our attempt will focus on using the polynomial form of the
approximation. After implementing the algorithm, we are going to report our results and compare it with
the real function output.
Based on correlation coefficient in image matchingIJRES Journal
With the development of image technology, the application of image technology in industrial
manufacturing become more and more widely. Image matching is animportantbranch of the image processing,
and also it is very important in the process of industrial detection. The main purpose of this paper is aiming at
the traditional image correlation matching algorithm in the application of detection in industrial, which based
on similarity matching principle. This paper is going to discuss the similarity of the image matching algorithm
on the application of detection in the connector. This paper improves the algorithm to speed up the detection
velocity. In some specific circumstance, while correlation coefficient method is more simple and easy to use, the
software development cycle will be short.
Presentation on Face Recognition Based on 3D Shape EstimationRapidAcademy
Face Recognition Based on 3D Shape Estimation
Quantify faces by parameters specifying their shape and texture.
To recognize faces across a wide range of illumination conditions.
Face recognition needs to be achieved across variations in pose.
Face Pose Classification Method using Image Structural Similarity Indexidescitation
This document proposes a new method for classifying face pose using structural similarity index (SSIM). SSIM is used to measure similarity between a test facial image and images in a database with known poses. The test image is assigned the pose of the database image with the highest SSIM value. Experimental results on the Pointing'04 database show the method can accurately classify poses when many training images are available. Classification confidence decreases when fewer training images are used, as poses may not be directly represented. The method could be useful for applications like driver monitoring that require pose authentication.
Object class recognition by unsupervide scale invariant learning - kunalKunal Kishor Nirala
This document presents an approach for unsupervised scale-invariant object class recognition using a probabilistic model. The model represents objects as flexible constellations of parts with probabilistic representations for shape, appearance, scale, and occlusion. An entropy-based feature detector is used to select image regions and scales. A maximum likelihood algorithm estimates the parameters of the scale-invariant object model from unlabeled training images. The model demonstrates good recognition performance on datasets with geometric and non-geometric object classes.
LEARNING FAIR GRAPH REPRESENTATIONS VIA AUTOMATED DATA AUGMENTATIONS.pptxssuser2624f71
The document proposes Graphair, a method for fair graph representation learning via automated data augmentations. Graphair uses an adversarial process to automatically discover fairness-aware data augmentation strategies for different graph datasets. The augmentation process modifies graphs by perturbing edges and masking node features. Graphair aims to learn representations that achieve demographic parity and equal opportunity while preserving predictive accuracy. Experimental results on node classification tasks show Graphair achieves better fairness metrics than baselines while maintaining high accuracy.
Similar to Tracking Faces using Active Appearance Models (20)
GraphRAG for Life Science to increase LLM accuracyTomaz Bratanic
GraphRAG for life science domain, where you retriever information from biomedical knowledge graphs using LLMs to increase the accuracy and performance of generated answers
“An Outlook of the Ongoing and Future Relationship between Blockchain Technologies and Process-aware Information Systems.” Invited talk at the joint workshop on Blockchain for Information Systems (BC4IS) and Blockchain for Trusted Data Sharing (B4TDS), co-located with with the 36th International Conference on Advanced Information Systems Engineering (CAiSE), 3 June 2024, Limassol, Cyprus.
UiPath Test Automation using UiPath Test Suite series, part 5DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 5. In this session, we will cover CI/CD with devops.
Topics covered:
CI/CD with in UiPath
End-to-end overview of CI/CD pipeline with Azure devops
Speaker:
Lyndsey Byblow, Test Suite Sales Engineer @ UiPath, Inc.
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
Programming Foundation Models with DSPy - Meetup SlidesZilliz
Prompting language models is hard, while programming language models is easy. In this talk, I will discuss the state-of-the-art framework DSPy for programming foundation models with its powerful optimizers and runtime constraint system.
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
Building Production Ready Search Pipelines with Spark and MilvusZilliz
Spark is the widely used ETL tool for processing, indexing and ingesting data to serving stack for search. Milvus is the production-ready open-source vector database. In this talk we will show how to use Spark to process unstructured data to extract vector representations, and push the vectors to Milvus vector database for search serving.
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
UiPath Test Automation using UiPath Test Suite series, part 6
Tracking Faces using Active Appearance Models
1. Tracking Faces
using Active Appearance Models
Steven Mitchell, Ph.D.
Componica, LLC
Iowa City, IA
Thursday, June 20, 13
2. Copyright 2013 - Componica, LLC (http://www.componica.com/)
Overview
Modeling the shape of faces.
Modeling the texture of faces.
Fitting this model to images with faces.
AAMs Revisited
Demo
Thursday, June 20, 13
3. Copyright 2013 - Componica, LLC (http://www.componica.com/)
Characterizing Faces
Locating a face... then
Characterizing the face... for
Face Recognition
Facial Tracking
Augmented Reality
Detecting emotions
Etc.
VIOLA, JONES: ROBUST REAL-TIME OBJECT DETECTION,
IJCV 2001
Thursday, June 20, 13
4. Copyright 2013 - Componica, LLC (http://www.componica.com/)
Faces have Shape
MUCT Facial Database of 3755
faces with landmark data.
Landmarks, a simple way of
encoding shape.
Landmarks must be fixed in
amount and location.
Explicit locations (corners), implicit
(edges, symmetry)
These x/y coordinates can be
represented as a vector:
S. MILBORROW AND J. MORKEL AND F. NICOLLS:
THE MUCT LANDMARKED FACE DATABASE, 2010
Thursday, June 20, 13
5. Copyright 2013 - Componica, LLC (http://www.componica.com/)
Removing Rotation, Scale, Translation
1. Select a face as the mean
shape:
2. Align all faces to the mean
using a least squares method:
3. Compute a new mean:
4. Go to step 2 until convergence.
Thursday, June 20, 13
6. Copyright 2013 - Componica, LLC (http://www.componica.com/)
Principal Component Analysis
PCA is a dimension reducing technique that
approximates data using a lower dimension.
1
x2
x1
x
φ
2
φ
1
x2
x1
x
φ
b
φ2
~x
x
(a) (b)
Figure 3.5: A simple visual depiction of Principal Component Analysis. a) PCA
is applied to a set of 2D vectors which form a cluster around a mean x. This
creates a set of principal components ¡1 and ¡2. b) We can approximate feature
vectors, such as x by eliminating smaller components such as ¡2
benefits of PCA are its ability to create a more compact model of the shape by throw-
ance matrix will exhibit correlation along specific directions. Solving
nvectors, ¡i, and eigenvalues, ∏i, derives these directions:
CΦ = ΛΦ (3.7)
enotes the concatenation of individual eigenvectors ¡i, and Λ is a
atrix of eigenvalues ∏i. By convention, we assume ∏i ∏ ∏i+1.
nd P computed, for any vector x there exists a vector b such that:
x = x + Φb (3.8)
CΦ = ΛΦ
where Φ denotes the concatenation of individual eigenvecto
diagonal matrix of eigenvalues ∏i. By convention, we assume
With L and P computed, for any vector x there exists a vec
x = x + Φb
Conversely there exists a vector b such that:
b = ΦT
(x ° x)
Since the smallest eigenvalues exhibit the least influence onThursday, June 20, 13
7. Copyright 2013 - Componica, LLC (http://www.componica.com/)
Principal Component Analysis
(DEMO...)
In general we can characterize the shape of a
face fairly well using less than 10 numbers.
Thursday, June 20, 13
8. Copyright 2013 - Componica, LLC (http://www.componica.com/)
Faces have Texture
Faces have pixel values. We’ll call them textures.
In order to compare a pixel between faces, warp the face to the average face.
Triangulate the landmark points using Delaunay Triangulation.
Map each triangle to the mean shape using Barycentric Coordinates.
Thursday, June 20, 13
9. Copyright 2013 - Componica, LLC (http://www.componica.com/)
Removing Brightness, Contrast
1. Select a face as the mean
texture:
2. Normalize lighting to the mean:
3. Compute a new mean:
4. Go to step 2 until
convergence.
Thursday, June 20, 13
10. Copyright 2013 - Componica, LLC (http://www.componica.com/)
Average Faces
Here are the average female, combined, and male
faces from the MUCT dataset.
Thursday, June 20, 13
11. Copyright 2013 - Componica, LLC (http://www.componica.com/)
Principal Component Analysis
Treat the textures as
vectors, g, and compute
PCA
Now we have a model for
texture
Combine shape, using
warping, with texture and
we have a model for
appearance.
eters and a combined appearance model with only 80 parameters required to explain 98%
observed variation. The model uses about 10,000 pixel values to make up the face patch.
gures 5.2 and 5.3 show the effects of varying the first two shape and grey-level model
eters through ±3 standard deviations, as determined from the training set. The first
eter corresponds to the largest eigenvalue of the covariance matrix, which gives its variance
the training set. Figure 5.4 shows the effect of varying the first four appearance model
eters, showing changes in identity, pose and expression.
5.2: First two modes of shape
on (±3 sd)
Figure 5.3: First two modes of grey-
level variation (±3 sd)
Approximating a New Example
a new image, labelled with a set of landmarks, we can generate an approximation with the
We follow the steps in the previous section to obtain b, combining the shape and grey-
5.4. APPROXIMATING A NEW EXAMPLE
Figure 5.4: First four modes of appearance variation (±3 sd)
level parameters which match the example. Since Pc is orthogonal, the combined appe
model parameters, c are given by
T.F. COOTES AND C.J.TAYLOR:
STATISTICAL MODELS OF APPEARANCE FOR COMPUTER VISION, 2004
Thursday, June 20, 13
12. Copyright 2013 - Componica, LLC (http://www.componica.com/)
A Model of Faces
We create a model of a face
that statistical captures the
variations of shape and
texture.
This model can both generate
faces and reduce faces to a
vector.
Typically 80-120 values are
sufficient to reconstruct most
faces.
This is a reconstruction of an
unseen image.
The full reconstruction is then given by applying equations (5
normalisation, applying the appropriate pose to the points and pro
into the image.
For example, Figure 5.5 shows a previously unseen image alongs
of the face patch (overlaid on the original image).
Figure 5.5: Example of combined model representation (right) of a
(left)
T.F. COOTES AND C.J.TAYLOR:
STATISTICAL MODELS OF APPEARANCE FOR COMPUTER VISION, 2004
Thursday, June 20, 13
13. Copyright 2013 - Componica, LLC (http://www.componica.com/)
Using this Model to Fit Faces
What do we have:
The input image
A reasonable location of a face via
Viola-Jones
A model that can reconstruct any
face from a small number of
parameters.
Goal:
Adjust the model to best fit the
synthetic face with real face.
From that, we’ve implicitly
characterized the face.
Minimize
Place model
in image
Measure
Difference
Update Model
Iterate
T.F. COOTES AND C.J.TAYLOR:
STATISTICAL MODELS OF APPEARANCE FOR COMPUTER VISION, 2004
INSERT MAGIC
HERE
Thursday, June 20, 13
14. Copyright 2013 - Componica, LLC (http://www.componica.com/)
Using this Model to Fit Faces
Knowing , we’d like to know how to adjust the model, and . In other
words, compute and .
Approximate and by and
Compute and :
In the original face data, compute a face that matches the original.
should be close to zero.
Perturb the model by a small random value and .
Generate pairs of and use least squares to compute and .
Side Note, we also compute four extra parameters for translation, scale and
rotation.
Thursday, June 20, 13
15. Copyright 2013 - Componica, LLC (http://www.componica.com/)
Using this Model to Fit Faces
8.3. LEARNING TO CORRECT MODEL PARAMETERS
Figure 8.2: First mode and displace-
ment weights
Figure 8.3: Third mode and d
placement weights
8.3.2 Perturbing The Face Model
To examine the performance of the prediction, we systematically displaced the face model fr
the true position on a set of 10 test images, and used the model to predict the displacem
given the sampled error vector. Figures 8.4 and 8.5 show the predicted translations against
actual translations. There is a good linear relationship within about 4 pixels of zero. Althou
this breaks down with larger displacements, as long as the prediction has the same sign as
actual error, and does not over-predict too far, an iterative updating scheme should conver
In this case up to 20 pixel displacements in x and about 10 in y should be correctable.
LEARNING TO CORRECT MODEL PARAMETERS 47
re 8.2: First mode and displace-
weights
Figure 8.3: Third mode and dis-
placement weights
2 Perturbing The Face Model
xamine the performance of the prediction, we systematically displaced the face model from
rue position on a set of 10 test images, and used the model to predict the displacement
n the sampled error vector. Figures 8.4 and 8.5 show the predicted translations against the
al translations. There is a good linear relationship within about 4 pixels of zero. Although
breaks down with larger displacements, as long as the prediction has the same sign as the
al error, and does not over-predict too far, an iterative updating scheme should converge.
is case up to 20 pixel displacements in x and about 10 in y should be correctable.
ment weights placement weights
8.3.2 Perturbing The Face Model
To examine the performance of the prediction, we systematically displaced the face model from
the true position on a set of 10 test images, and used the model to predict the displacement
given the sampled error vector. Figures 8.4 and 8.5 show the predicted translations against the
actual translations. There is a good linear relationship within about 4 pixels of zero. Although
this breaks down with larger displacements, as long as the prediction has the same sign as the
actual error, and does not over-predict too far, an iterative updating scheme should converge.
In this case up to 20 pixel displacements in x and about 10 in y should be correctable.
actual dx (pixels)
predicteddx
−40 −30 −20 −10 0 10 20 30 40
−8
−6
−4
−2
0
2
4
6
8
Figure 8.4: Predicted dx vs actual
dx. Errorbars are 1 standard error
actual dx (pixels)
predicteddy
−40 −30 −20 −10 0 10 20 30 40
−8
−6
−4
−2
0
2
4
6
8
Figure 8.5: Predicted dy vs actual
dy. Errorbars are 1 standard error
We can, however, extend this range by building a multi-resolution model of object appear-
ance. We generate Gaussian pyramids for each of our training images, and generate an appear-
Ideally we seek to model a relationship that holds over as large a range errors, δg, as poss
However, the real relationship is found to be linear only over a limited range of values.
experiments on the face model suggest that the optimum perturbation was around 0.5 stand
deviations (over the training set) for each model parameter, about 10% in scale, the equiva
of 3 pixels translation and about 10% in texture scaling.
8.3.1 Results For The Face Model
We applied the above algorithm to the face model described in section 5.3.
We can visualise the effects of the perturbation as follows. If ai is the ith row of the ma
R, the predicted change in the ith parameter, δci is given by
δci = ai.δg (
and ai gives the weight attached to different areas of the sampled patch when estimating
displacement. Figure 8.1 shows the weights corresponding to changes in the pose parame
(sx, sy, tx, ty). Bright areas are positive weights, dark areas negative. As one would exp
the x and y displacement weights are similar to x and y derivative images. Similar results
obtained for weights corresponding to the appearance model parameters
Figure 8.2 and 8.3 show the first and third modes and corresponding displacement weig
The areas which exhibit the largest variations for the mode are assigned the largest weight
the training process.
Figure 8.1: Weights corresponding to changes in the pose parameters, (sx, sy, tx, ty)
T.F. COOTES AND C.J.TAYLOR:
STATISTICAL MODELS OF APPEARANCE FOR COMPUTER VISION, 2004
Thursday, June 20, 13
16. Copyright 2013 - Componica, LLC (http://www.componica.com/)
Using this Model to Fit FacesFigure 8.9: Reconstruction (left) and original (right) given original landmark points
Initial 2 its 8 its 14 its 20 its converged
Figure 8.10: Multi-Resolution search from displaced position
As an example of applying the method to medical images, we built an Appearance Model
T.F. COOTES AND C.J.TAYLOR:
STATISTICAL MODELS OF APPEARANCE FOR COMPUTER VISION, 2004
Thursday, June 20, 13
17. Copyright 2013 - Componica, LLC (http://www.componica.com/)
Using this Model to Fit Faces
IMENTAL RESULTS 51
14 shows the mean intensity error per pixel (for an image using 256 grey-levels)
number of iterations, averaged over a set of searches at a single resolution. In
model was initially displaced by up to 15 pixels. The dotted line gives the mean
n error using the hand marked landmark points, suggesting a good result is obtained
.
15 shows the proportion of 100 multi-resolution searches which converged correctly
g positions displaced from the true position by up to 50 pixels in x and y. The
ys good results with up to 20 pixels (10% of the face width) displacement.
Number of Iterations
Meanintensityerror/pixel
0 2.5 5.0 7.5 10.0 12.5 15.0 17.5 20.0
0
2
4
6
8
10
12
14
Mean intensity error as search progresses. Dotted line is the mean error of the best
dmarks.
y
70
80
90
100
dX
Number of Iterations
0 2.5 5.0 7.5 10.0 12.5 15.0 17.5 20.0
0
Figure 8.14: Mean intensity error as search progresses. Dotted line is the mean er
fit to the landmarks.
Displacement (pixels)
%convergedcorrectly
−50 −40 −30 −20 −10 0 10 20 30 40 50
0
10
20
30
40
50
60
70
80
90
100
dX
dY
Figure 8.15: Proportion of searches which converged from different initial disp
Thursday, June 20, 13
18. Copyright 2013 - Componica, LLC (http://www.componica.com/)
AAMs Revisited
Context:
Tim Cootes, Edward, and Taylor’s original AAM model was created in the
late 90s.
No good face initialization until 2001 with the Cascade Haar face detector.
Had limited use in face recognition and medical imaging.
Generally the algorithm seemed ad hoc.
Redux:
In 2004, Iain Matthews and Simon Baker reexamined AAMs making
significant improvements.
Iain Matthews went on to develop the facial motion tracking for the Movie
Avatar working for Weta Digital in New Zealand and now works for Disney
Research.
Thursday, June 20, 13
19. Copyright 2013 - Componica, LLC (http://www.componica.com/)
AAMs Revisited
Replaced simple gradient decent to more efficient
gradient decent using Gauss-Newton.
Analytically computes the gradients instead of
fitting perturbations between model coefficients
and image differences.
Projects out the texture model from AAM fitting
both simplifying and speeding up the algorithm.
Inverse Warp Composition.
Thursday, June 20, 13
20. Copyright 2013 - Componica, LLC (http://www.componica.com/)
Better Gradient Descent
Cootes’ gradient descent expanding out all the
terms. Notice the k fudge-factor term.
Matthews’ gradient descent via Gauss-Newton.
PRECOMPUTED
Thursday, June 20, 13
21. Copyright 2013 - Componica, LLC (http://www.componica.com/)
Better Gradients
Matthews formulated a direct computation of the
gradient images assuming triangular warping.
Thursday, June 20, 13
23. Copyright 2013 - Componica, LLC (http://www.componica.com/)
Projecting out Texture
Project out the texture variations from the gradient images. This
means the steepest descent images, R, has the texture variations
subtracted out.
The model only needs to fit the shapes ignoring the texture
component of the facial model.
Significantly reduces computation cost as only the shape model is
considered (10 parameters vs. 80 or more parameters).
Thursday, June 20, 13
24. Copyright 2013 - Componica, LLC (http://www.componica.com/)
Demo
http://www.youtube.com/watch?v=VblXShzV2VY
Way cooler live in realtime on my laptop.
Thursday, June 20, 13
25. Copyright 2013 - Componica, LLC (http://www.componica.com/)
Implementation Details
C++ and Open CV for camera capture, face detection, and
matrix operations.
Implementation suckage: 8/10 - Took 4 solid months to
implement with many details left out of the talk.
Working on proprietary features to improve accuracy and
convergence. Potential topics for future presentations.
Porting to iOS & Android with the goal of close to realtime
tracking.
Derive as many products, apps, startups as I can from it to
make money.
Thursday, June 20, 13
26. Copyright 2013 - Componica, LLC (http://www.componica.com/)
Conclusion
Active Appearance Models statistically model shape
and texture of objects.
An optimization scheme fits a model onto an image
implicitly tracking the object.
Original formulated by Cootes et al, AAMs have
recently been enhanced by Matthews et al in 2004.
With the combination of practical face detection and
the speed of the algorithm, AAMs potentially are a
useful step in characterizing and tracking faces.
Thursday, June 20, 13