SlideShare a Scribd company logo
1 of 6
Download to read offline
Electrical & Computer Engineering: An International Journal (ECIJ) Volume 4, Number 4, December 2015
DOI : 10.14810/ecij.2015.4403 23
A LIP LOCALIZATION BASED VISUAL FEATURE
EXTRACTION METHOD
Namrata Dave1
1
Department of Computer Engineering, Gujarat Technological University, India
ABSTRACT
This paper presents a lip localization based visual feature extraction method to segment lip region from
image or video in real time. Lip localization and tracking is useful in many applications such as lip
reading, lip synchronization, visual speech recognition, facial animation etc. To synchronize lip movements
with input audio we need to first segment lip region from input image or video frame. This paper presents a
novel color based approach for localization of lips, which is an early stage for tracking lips in real time. A
phoneme is a basic unit of speech and a viseme is a visual representation of phoneme or shape of mouth
while utterance of particular phoneme. The main goal of our work is to implement a system for
synchronizing lips with the input speech. To extract visual features i.e. visemes from input video frame or
image we have used HSV and YCbCr color model along with various morphological operations. We have
developed algorithm to work with normal lighting conditions and natural facial images of female and male.
KEYWORDS
Lip localization, viseme extraction, visual feature extraction, HSV color model, YCbCr color model.
1. INTRODUCTION
Visual features of speech i.e. visemes are useful in various fields, such as visual speech
recognition [1], [2], audio-video synchronization, or speaker identification/localization in audio-
video scenes [3]. It is observed that mostly visual information related to speech is contained in the
movements of lip. However, a most difficult part in visual speech recognition is to find an
efficient method for extracting visual speech features. It is very difficult to extract visual features
from input image or video which is taken in different situations under varying lighting conditions,
different pose and for different skin tones. Many algorithms have been proposed for lip
localization and viseme extraction. Several groups are working with template models based on
active shape models [4], deformable models [5] and dynamic contours [6].
A lot of work is already done on this topic since years. Many novel approaches have been
proposed in this field. Most approaches are divided into two main groups. First group is using
shape based sophisticated methods like active shape model, deformable models, snake etc. to
extract features from input image or video. Another group is extracting features using different
colour models using different spatial methods which work on pixel intensity, edge and colour
information of lip and/or facial skin.
In this paper, we presented image based approach to extract mouth region and then to
discriminate visual features from it. Here we have tried to extract visemes for digits one to ten.
2. EXISTING METHODS
The Lip segmentation has become an important issue in facial animation, automatic VSR
processing and automatic face recognition. In such systems, the region of interest (ROI) which is
Electrical & Computer Engineering: An International Journal (ECIJ) Volume 4, Number 4, December 2015
24
lips in our case must be detected in each frame. This procedure is normally done by face detection
and extraction of the lip region.
Earlier systems performed the lip segmentation in conjunction with the application of artificial
markers (lipstick) on the lips. The application of lipstick enables the system to detect precisely the
lips in the image data, but this procedure is inappropriate since it is uncomfortable for users and
such systems can be operated only in specific constraint based applications. Thus, the main
research efforts have been concentrated in the development of visual features based algorithms
for lip segmentation. Many studies have shown that colour information of particular region can be
applied to identify the skin or face in digital images. The main idea behind this approach is to
transform the RGB signal into other representation where the mouth region can efficiently
separate from other details in given image, so that segmentation of mouth can be done efficiently.
Due to this fact, a large number of colour representations have been proposed. Coinaiz et al [7]
used the hue component of the HSV representation to highlight the red colour which is assumed
to be associated with the lips in the image. Later, the HSV colour space is used [8] for lip
detection. They used the hue signal to locate the position of the lips in mouth region. The interior
and exterior boundaries of lips are extracted using colour and edge information using a Markov
Random Field (MRF) framework based on the lip area found. Some of the approaches used
YCrCb colour space for the lips detection [9, 10].
Eveno et al [11] proposed a new colour mixture and chromatic transformation for lip
segmentation in 2001. In their approach, a new method for transformation of the RGB colour
space and a chromatic map was applied to separate out the lips and facial skin. They argued that
their proposed approach is able to provide robust lip detection under variable lighting conditions.
Later on Eveno et al introduced a novel method where the pseudo-hue [12] was applied for lip
segmentation that has been embedded in an active contour framework. They applied the proposed
algorithm for visual speech recognition and the results show significant improvement in terms of
accuracy in lip modelling.
Another method for mouth segmentation has been proposed by Liew in 2003 [13]. In their
approach, they used new transformation method to convert given colour image into the CIE-Lab
colour space and CIE-LUV colour space, and then they calculated a lip membership map using
the fuzzy clustering algorithm. The ROI around the mouth can be identified from the face area
after application of morphological filtering on given image.
Guan [14] has given a new approach based on Discrete Hartley Transform (DHT) to improve the
contrast between lip region and the other regions of face. In their paper, they presented method of
extraction of lips by applying wavelet multi-scale edge detection across the C3 component of the
DHT. Their method takes into account both the colour information and the geometric
characteristic of image.
3. VISUAL FEATURE EXTRACTION METHOD.
3.1. Proposed Method
Feature extraction process is preceded by a number of pre-processing steps to be done as shown
in Figure 1. This involves face detection followed by ROI extraction. Then, the lips of the speaker
are tracked in consecutive frames of recorded video. Following these steps, and given an
informative set of features, the visual front-end module can proceed with feature extraction.
Electrical & Computer Engineering: An International Journal (ECIJ) Volume 4, Number 4, December 2015
25
Figure 1. Visual Feature Extraction Process
3.2. Dataset and Implementation
A viseme is the visual appearance of a phoneme or visual unit of speech in spoken language.
However, currently there is no standard viseme table available to be used by all researchers. Due
to this fact, a viseme table for this study was based on the work performed by Lee and Sook in
2002 [15, 16]. This was used in a table which effectively mapped all the possible phonemes to
viseme. They have shown the typical 48 phonemes used in English language including silence
and short pauses, grouped into their 14 viseme classes.
Table 1: List of Phonemes & Viseme Used for Digits One to Five
Word Phoneme Viseme
One w ah n/w an n/v ao n v ao n
Two t uw/d uw t uw
Three th r iy/dh iy th r iy
Four f ao r f ao r
Five f ay v f ay v
Six s ih k s s ih k s
Seven s eh v ax n s eh v ax n
Eight ey t ey t
Nine n ay n n ay n
Ten t eh n t eh n
Face Detection
ROI
Extraction
Mouth/ Lip
Contour
Estimation
Visual Feature
Extraction
Electrical & Computer Engineering: An International Journal (ECIJ) Volume 4, Number 4, December 2015
26
We have recorded videos for 12 different viseme classes as listed in Table1. Table1 shows list of
phonemes and corresponding viseme for digits one to ten. We extracted mouth parameters for
each viseme class for 6 different speakers. Videos recorded for different speakers include female
and male speakers’ videos. Videos are recorded in normal conditions using webcam. No artificial
marker like lipstick is used for detecting lips. Videos are recorded in avi format with 30fps.
In this section, we will discuss quick summary of how exactly lip features are extracted from the
video of speech. Entire video is converted into frames. RGB colour scheme of the image is not
suitable for immediate processing as it contains a lot of mixed information about lightness etc.
Another colour scheme should be used. Very convenient colour scheme is YCbCr as it separates
luminance, blue chrominance and red chrominance. This method is convenient as the mouth
region in image contains high red and low blue components in comparison with other face
regions.
Figure 2: Overview of the Proposed Algorithm for Lip Feature Extraction
The mouth region has a greater value of red chromaticity (Cr) than of blue chromaticity (Cb). The
value of the chromaticity of the Cr component is increased by using its square value. On the other
hand, the region of the mouth has a good response to the ratio Cb/Cr.
Lip Feature Extraction Algorithm steps are described in figure 2. First, Video divided into frames
(frame rate 30 frames per second). On each frame all operations described in algorithm are
performed. We have used YCBCR colour space. Based on luminance and chromaticity
information extracted from frame we are extracting mouth region surrounding Region of interest
i.e. Lips. The ratio Cb/Cr is used to extract face region and eliminate background from input
video frame. From the cropped face image we extracted lip area based on H and S value by using
HSV colour model. On cropped lip image edge detection is applied first. Then to eliminate extra
info morphological operations are applied to get final lip parameters. We have used these features
to estimate mouth region with information such as height and width for each viseme.
3.2. Results
Four video sequences were taken from different subjects, each one having at most than 3X30
frames. Those sequences were analysed using the proposed lip segmentation algorithm. The
Face Region Extraction
Lip features: Lip Width & Height Extracted
Mouth Region (Region of Interest) Extracted
Video divided into still image
Electrical & Computer Engineering: An International Journal (ECIJ) Volume 4, Number 4, December 2015
27
figure 3 shows examples of segmentation results for each subject. The best results were obtained
for the pale skinned, non-bearded subject. We have used s and h value of HSV colour space to
identify lip region based on which all viseme are extracted. Sometimes due to poor illumination
we are not get good results only based on H and S value so we also try-out combining results of
contrast stretched and log applied images to get exactly mouth region.
Using Lip Feature Extraction Algorithm we extracted Viseme for 10 different viseme classes.
Figure 3 shows results of viseme V for four speakers. We tested our algorithm for many frames
for each speaker. For testing purpose for each speaker we selected two videos and extracted
frames to apply algorithm on it. On an average we are getting 85% result for all speakers.
Figure 3: Results viseme 'v'
4. CONCLUSIONS
After the investigation done on existing methods, we have implemented a lip localization based
visual feature extraction method which gives good accuracy for our database. Our algorithm is
suitable for offline applications which can be extended to work in real time applications. Another
improvement to the program can be achieved by trying it for available datasets. We have planned
to use results obtained to use for lip-reading application as future enhancement.
ACKNOWLEDGEMENTS
We would like to thank everyone in my friends, staff members and students who helped me in
recording videos.
Electrical & Computer Engineering: An International Journal (ECIJ) Volume 4, Number 4, December 2015
28
REFERENCES
[1] Tsuhan Chen and R. R, Rao, “Audio-visual interaction in multimedia communication,” in Proc. Of
IEEE Intern. Con$ on Acoustics, Speech and Signal Processing (ICASSP’97), vol. 1, 1997, pp. 179-
182.
[2] X. Zhang and R. M. Mersereau, M. Clements, C. C. Broun, “Visual Speech Feature Extraction for
Improved Speech Recognition,” in Proc. IEEE Intern. Con$ on Acoustics, Speech, and Signal
Processing (ICASSP’O2), vol. 2, 2002, pp. 1993-1996.
[3] D. Lo, R. A. Goubran, J. Gammal, G. Thompson, and D. Schulz, “Robust Joint Audio-Video
Localization in Video Conferencing using Reliability Information,” in Proc. 20th IEEE
Instrumentation and Measurement Technology Conf: (IMTC’O3), vol. 2, May 2003, pp. 141-418.
[4] K. L. Sum, W. H. Lau, S. H. Leung, A. W. C. Liew, and K. W. Tse, “A New Optimization Procedure
for Extracting the Point-based Lip Contour Using Active Shape Model,” in Proc. Of IEEE Intern.
Con. on Acoustics, Speech and Signal Processing (ICASSP ’Ol), vol. 3, May 2001, pp. 1485-1488.
[5] A. W. C Liew, S.H. Leung, and W.H. Lau, “Lip Contour Extraction Using a Deformable Model,” in
Proc. of IEEE Intern. Con. on Image Processing (ICIP’OO), vol. 2, Sept. 2000, pp. 255-258.
[6] R. Kaucic and A. Blake, “Accurate, Real-time, Unadomed Lip Tracking,” in Proc. of 6th International
Conference on Computer Vision, 1998, pp. 370-375.
[7] T. Coianiz, L. Torresani and B. Caprile, “2D Deformable Models for Visual Speech Analysis”,
Proceedings of Springer, Speech reading by Humans and Machines, D.G. stork & M. E. Hennecke
Eds., NY, 1996.
[8] X. Zhang and R.M. Mersereau, “Lip Features extraction Towards an Automatic Speech-reading
System”, Proceedings of ICIP00, Wa07.05, Vancouver, Canada,2000.
[9] M. Heckmann, F. Berthommier and K. Kroschel, “Noise adaptive stream weighting in audio-visual
speech recognition”, Proceedings of EURASIP J. Appl. Signal, vol. 2002, pp. 1260-1273, Nov. 2002.
[10] J. Chaloupka, “Automatic Lips Reading for Audio-Visual Speech Processing and Recognition”,
Proceedings of Of ICSLP, pp. 2505-2508, Jeju Island, Korea, Oct. 2004.
[11] N. Eveno, A. Caplier, P.Y. Coulon. “A new color transformation for lips segmentation”, Proceedings
of IEEE Fourth Workshop on Multimedia Signal, pp. 3-8, Cannes, France, 2001.
[12] N. Eveno, A. Caplier and P. Coulon, “Accurate and Quasi-Automatic Lip Tracking”, Proceedings of
IEEE Trans. Circuits Syst. Video Techn. 14(5): 706-715., 2004.
[13] A.W.C. Liew, S.H. Leung, and W.H. Lau, “Segmentation of color lip images by spatial fuzzy
clustering” Proceedings of IEEE Trans. Fuzzy Syst. vol. 11 no. 4, pp. 542-549, Aug. 2003.
[14] Y. P. Guan, “Automatic Extraction of Lip Based on Wavelet Edge Detection”, Proceedings of the
Eighth international Symposium on Symbolic and Numeric Algorithms, Scientific Computing
SYNASC. IEEE Computer Society, pp. 125-132. Sept. 2006.
[15] S. Lee and D. Yook, “Viseme recognition experiment using context dependent Hidden Markov
Models”, Proceedings of Third Intl. Conf. on Intelligent Data Engineering and Automated learning,
Vol. 2412 pp. 557-561, 2002.
[16] N. Dave, N. M. Patel. "Phoneme and Viseme based Approach for Lip Synchronization.", International
Journal of Signal Processing, Image Processing and Pattern Recognition. 7(3), pp. 385-394, 2014.
Authors
Namrata A Dave
Assistant Professor,
Computer Engineering Department,
G H Patel College of Engg. & Tech.
GUJARAT, INDIA.

More Related Content

Similar to A LIP LOCALIZATION BASED VISUAL FEATURE EXTRACTION METHOD

Abstract Silent Sound Technology
Abstract   Silent Sound TechnologyAbstract   Silent Sound Technology
Abstract Silent Sound Technologyvishnu murthy
 
lips _reading_nagham _salim compute.pptx
lips _reading_nagham _salim compute.pptxlips _reading_nagham _salim compute.pptx
lips _reading_nagham _salim compute.pptxnaghamallella
 
lips _reading _in computer_ vision_n.ppt
lips _reading _in computer_ vision_n.pptlips _reading _in computer_ vision_n.ppt
lips _reading _in computer_ vision_n.pptnaghamallella
 
Speaker independent visual lip activity detection for human - computer inte...
Speaker   independent visual lip activity detection for human - computer inte...Speaker   independent visual lip activity detection for human - computer inte...
Speaker independent visual lip activity detection for human - computer inte...eSAT Journals
 
Speaker independent visual lip activity detection for
Speaker   independent visual lip activity detection forSpeaker   independent visual lip activity detection for
Speaker independent visual lip activity detection foreSAT Publishing House
 
Lip Reading by Using 3-D Discrete Wavelet Transform with Dmey Wavelet
Lip Reading by Using 3-D Discrete Wavelet Transform with Dmey WaveletLip Reading by Using 3-D Discrete Wavelet Transform with Dmey Wavelet
Lip Reading by Using 3-D Discrete Wavelet Transform with Dmey WaveletCSCJournals
 
Survey on Facial Expression Analysis and Recognition
Survey on Facial Expression Analysis and RecognitionSurvey on Facial Expression Analysis and Recognition
Survey on Facial Expression Analysis and RecognitionIRJET Journal
 
Comparative Study of Lip Extraction Feature with Eye Feature Extraction Algor...
Comparative Study of Lip Extraction Feature with Eye Feature Extraction Algor...Comparative Study of Lip Extraction Feature with Eye Feature Extraction Algor...
Comparative Study of Lip Extraction Feature with Eye Feature Extraction Algor...Editor IJCATR
 
IRJET - Automatic Lip Reading: Classification of Words and Phrases using Conv...
IRJET - Automatic Lip Reading: Classification of Words and Phrases using Conv...IRJET - Automatic Lip Reading: Classification of Words and Phrases using Conv...
IRJET - Automatic Lip Reading: Classification of Words and Phrases using Conv...IRJET Journal
 
Ijarcet vol-2-issue-4-1352-1356
Ijarcet vol-2-issue-4-1352-1356Ijarcet vol-2-issue-4-1352-1356
Ijarcet vol-2-issue-4-1352-1356Editor IJARCET
 
Multi-View Algorithm for Face, Eyes and Eye State Detection in Human Image- S...
Multi-View Algorithm for Face, Eyes and Eye State Detection in Human Image- S...Multi-View Algorithm for Face, Eyes and Eye State Detection in Human Image- S...
Multi-View Algorithm for Face, Eyes and Eye State Detection in Human Image- S...IJERA Editor
 
Facial expression identification by using features of salient facial landmarks
Facial expression identification by using features of salient facial landmarksFacial expression identification by using features of salient facial landmarks
Facial expression identification by using features of salient facial landmarkseSAT Journals
 
Facial expression identification by using features of salient facial landmarks
Facial expression identification by using features of salient facial landmarksFacial expression identification by using features of salient facial landmarks
Facial expression identification by using features of salient facial landmarkseSAT Journals
 
Paper id 23201490
Paper id 23201490Paper id 23201490
Paper id 23201490IJRAT
 
A Study On Facial Expression Recognition Using Local Binary Pattern
A Study On Facial Expression Recognition Using Local Binary PatternA Study On Facial Expression Recognition Using Local Binary Pattern
A Study On Facial Expression Recognition Using Local Binary PatternMartha Brown
 
LIP READING: VISUAL SPEECH RECOGNITION USING LIP READING
LIP READING: VISUAL SPEECH RECOGNITION USING LIP READINGLIP READING: VISUAL SPEECH RECOGNITION USING LIP READING
LIP READING: VISUAL SPEECH RECOGNITION USING LIP READINGIRJET Journal
 
SIGN LANGUAGE RECOGNITION USING MACHINE LEARNING
SIGN LANGUAGE RECOGNITION USING MACHINE LEARNINGSIGN LANGUAGE RECOGNITION USING MACHINE LEARNING
SIGN LANGUAGE RECOGNITION USING MACHINE LEARNINGIRJET Journal
 
NEURAL NETWORK APPROACH FOR EYE DETECTION
NEURAL NETWORK APPROACH FOR EYE DETECTIONNEURAL NETWORK APPROACH FOR EYE DETECTION
NEURAL NETWORK APPROACH FOR EYE DETECTIONcscpconf
 

Similar to A LIP LOCALIZATION BASED VISUAL FEATURE EXTRACTION METHOD (20)

Abstract Silent Sound Technology
Abstract   Silent Sound TechnologyAbstract   Silent Sound Technology
Abstract Silent Sound Technology
 
lips _reading_nagham _salim compute.pptx
lips _reading_nagham _salim compute.pptxlips _reading_nagham _salim compute.pptx
lips _reading_nagham _salim compute.pptx
 
lips _reading _in computer_ vision_n.ppt
lips _reading _in computer_ vision_n.pptlips _reading _in computer_ vision_n.ppt
lips _reading _in computer_ vision_n.ppt
 
Speaker independent visual lip activity detection for human - computer inte...
Speaker   independent visual lip activity detection for human - computer inte...Speaker   independent visual lip activity detection for human - computer inte...
Speaker independent visual lip activity detection for human - computer inte...
 
Speaker independent visual lip activity detection for
Speaker   independent visual lip activity detection forSpeaker   independent visual lip activity detection for
Speaker independent visual lip activity detection for
 
Lip Reading by Using 3-D Discrete Wavelet Transform with Dmey Wavelet
Lip Reading by Using 3-D Discrete Wavelet Transform with Dmey WaveletLip Reading by Using 3-D Discrete Wavelet Transform with Dmey Wavelet
Lip Reading by Using 3-D Discrete Wavelet Transform with Dmey Wavelet
 
Survey on Facial Expression Analysis and Recognition
Survey on Facial Expression Analysis and RecognitionSurvey on Facial Expression Analysis and Recognition
Survey on Facial Expression Analysis and Recognition
 
Comparative Study of Lip Extraction Feature with Eye Feature Extraction Algor...
Comparative Study of Lip Extraction Feature with Eye Feature Extraction Algor...Comparative Study of Lip Extraction Feature with Eye Feature Extraction Algor...
Comparative Study of Lip Extraction Feature with Eye Feature Extraction Algor...
 
IRJET - Automatic Lip Reading: Classification of Words and Phrases using Conv...
IRJET - Automatic Lip Reading: Classification of Words and Phrases using Conv...IRJET - Automatic Lip Reading: Classification of Words and Phrases using Conv...
IRJET - Automatic Lip Reading: Classification of Words and Phrases using Conv...
 
Ijarcet vol-2-issue-4-1352-1356
Ijarcet vol-2-issue-4-1352-1356Ijarcet vol-2-issue-4-1352-1356
Ijarcet vol-2-issue-4-1352-1356
 
Multi-View Algorithm for Face, Eyes and Eye State Detection in Human Image- S...
Multi-View Algorithm for Face, Eyes and Eye State Detection in Human Image- S...Multi-View Algorithm for Face, Eyes and Eye State Detection in Human Image- S...
Multi-View Algorithm for Face, Eyes and Eye State Detection in Human Image- S...
 
10.1.1.33.4990
10.1.1.33.499010.1.1.33.4990
10.1.1.33.4990
 
Facial expression identification by using features of salient facial landmarks
Facial expression identification by using features of salient facial landmarksFacial expression identification by using features of salient facial landmarks
Facial expression identification by using features of salient facial landmarks
 
Facial expression identification by using features of salient facial landmarks
Facial expression identification by using features of salient facial landmarksFacial expression identification by using features of salient facial landmarks
Facial expression identification by using features of salient facial landmarks
 
Paper id 23201490
Paper id 23201490Paper id 23201490
Paper id 23201490
 
A Study On Facial Expression Recognition Using Local Binary Pattern
A Study On Facial Expression Recognition Using Local Binary PatternA Study On Facial Expression Recognition Using Local Binary Pattern
A Study On Facial Expression Recognition Using Local Binary Pattern
 
LIP READING: VISUAL SPEECH RECOGNITION USING LIP READING
LIP READING: VISUAL SPEECH RECOGNITION USING LIP READINGLIP READING: VISUAL SPEECH RECOGNITION USING LIP READING
LIP READING: VISUAL SPEECH RECOGNITION USING LIP READING
 
SIGN LANGUAGE RECOGNITION USING MACHINE LEARNING
SIGN LANGUAGE RECOGNITION USING MACHINE LEARNINGSIGN LANGUAGE RECOGNITION USING MACHINE LEARNING
SIGN LANGUAGE RECOGNITION USING MACHINE LEARNING
 
[IJET-V1I5P9] Author: Prutha Gandhi, Dhanashri Dalvi, Pallavi Gaikwad, Shubha...
[IJET-V1I5P9] Author: Prutha Gandhi, Dhanashri Dalvi, Pallavi Gaikwad, Shubha...[IJET-V1I5P9] Author: Prutha Gandhi, Dhanashri Dalvi, Pallavi Gaikwad, Shubha...
[IJET-V1I5P9] Author: Prutha Gandhi, Dhanashri Dalvi, Pallavi Gaikwad, Shubha...
 
NEURAL NETWORK APPROACH FOR EYE DETECTION
NEURAL NETWORK APPROACH FOR EYE DETECTIONNEURAL NETWORK APPROACH FOR EYE DETECTION
NEURAL NETWORK APPROACH FOR EYE DETECTION
 

More from ecij

NEW Current Issue - CALL FOR PAPERS - Electrical and Computer Engineering An ...
NEW Current Issue - CALL FOR PAPERS - Electrical and Computer Engineering An ...NEW Current Issue - CALL FOR PAPERS - Electrical and Computer Engineering An ...
NEW Current Issue - CALL FOR PAPERS - Electrical and Computer Engineering An ...ecij
 
Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)ecij
 
Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)ecij
 
Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)ecij
 
Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)ecij
 
ADOPTING MEASURES TO REDUCE POWER OUTAGES
ADOPTING MEASURES TO REDUCE POWER OUTAGESADOPTING MEASURES TO REDUCE POWER OUTAGES
ADOPTING MEASURES TO REDUCE POWER OUTAGESecij
 
GRID SIDE CONVERTER CONTROL IN DFIG BASED WIND SYSTEM USING ENHANCED HYSTERES...
GRID SIDE CONVERTER CONTROL IN DFIG BASED WIND SYSTEM USING ENHANCED HYSTERES...GRID SIDE CONVERTER CONTROL IN DFIG BASED WIND SYSTEM USING ENHANCED HYSTERES...
GRID SIDE CONVERTER CONTROL IN DFIG BASED WIND SYSTEM USING ENHANCED HYSTERES...ecij
 
Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)ecij
 
PREPARATION OF POROUS AND RECYCLABLE PVA-TIO2HYBRID HYDROGEL
PREPARATION OF POROUS AND RECYCLABLE PVA-TIO2HYBRID HYDROGELPREPARATION OF POROUS AND RECYCLABLE PVA-TIO2HYBRID HYDROGEL
PREPARATION OF POROUS AND RECYCLABLE PVA-TIO2HYBRID HYDROGELecij
 
4th International Conference on Electrical Engineering (ELEC 2020)
4th International Conference on Electrical Engineering (ELEC 2020)4th International Conference on Electrical Engineering (ELEC 2020)
4th International Conference on Electrical Engineering (ELEC 2020)ecij
 
Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)ecij
 
4th International Conference on Bioscience & Engineering (BIEN 2020)
4th International Conference on Bioscience & Engineering (BIEN 2020) 4th International Conference on Bioscience & Engineering (BIEN 2020)
4th International Conference on Bioscience & Engineering (BIEN 2020) ecij
 
Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)ecij
 
Ecij cfp
Ecij cfpEcij cfp
Ecij cfpecij
 
GRID SIDE CONVERTER CONTROL IN DFIG BASED WIND SYSTEM USING ENHANCED HYSTERES...
GRID SIDE CONVERTER CONTROL IN DFIG BASED WIND SYSTEM USING ENHANCED HYSTERES...GRID SIDE CONVERTER CONTROL IN DFIG BASED WIND SYSTEM USING ENHANCED HYSTERES...
GRID SIDE CONVERTER CONTROL IN DFIG BASED WIND SYSTEM USING ENHANCED HYSTERES...ecij
 
UNION OF GRAVITATIONAL AND ELECTROMAGNETIC FIELDS ON THE BASIS OF NONTRADITIO...
UNION OF GRAVITATIONAL AND ELECTROMAGNETIC FIELDS ON THE BASIS OF NONTRADITIO...UNION OF GRAVITATIONAL AND ELECTROMAGNETIC FIELDS ON THE BASIS OF NONTRADITIO...
UNION OF GRAVITATIONAL AND ELECTROMAGNETIC FIELDS ON THE BASIS OF NONTRADITIO...ecij
 
USING MACHINE LEARNING TO BUILD A SEMI-INTELLIGENT BOT
USING MACHINE LEARNING TO BUILD A SEMI-INTELLIGENT BOT USING MACHINE LEARNING TO BUILD A SEMI-INTELLIGENT BOT
USING MACHINE LEARNING TO BUILD A SEMI-INTELLIGENT BOT ecij
 
MODELING AND SIMULATION OF SOLAR PHOTOVOLTAIC APPLICATION BASED MULTILEVEL IN...
MODELING AND SIMULATION OF SOLAR PHOTOVOLTAIC APPLICATION BASED MULTILEVEL IN...MODELING AND SIMULATION OF SOLAR PHOTOVOLTAIC APPLICATION BASED MULTILEVEL IN...
MODELING AND SIMULATION OF SOLAR PHOTOVOLTAIC APPLICATION BASED MULTILEVEL IN...ecij
 
Investigation of Interleaved Boost Converter with Voltage multiplier for PV w...
Investigation of Interleaved Boost Converter with Voltage multiplier for PV w...Investigation of Interleaved Boost Converter with Voltage multiplier for PV w...
Investigation of Interleaved Boost Converter with Voltage multiplier for PV w...ecij
 
A COMPARISON BETWEEN SWARM INTELLIGENCE ALGORITHMS FOR ROUTING PROBLEMS
A COMPARISON BETWEEN SWARM INTELLIGENCE ALGORITHMS FOR ROUTING PROBLEMSA COMPARISON BETWEEN SWARM INTELLIGENCE ALGORITHMS FOR ROUTING PROBLEMS
A COMPARISON BETWEEN SWARM INTELLIGENCE ALGORITHMS FOR ROUTING PROBLEMSecij
 

More from ecij (20)

NEW Current Issue - CALL FOR PAPERS - Electrical and Computer Engineering An ...
NEW Current Issue - CALL FOR PAPERS - Electrical and Computer Engineering An ...NEW Current Issue - CALL FOR PAPERS - Electrical and Computer Engineering An ...
NEW Current Issue - CALL FOR PAPERS - Electrical and Computer Engineering An ...
 
Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)
 
Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)
 
Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)
 
Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)
 
ADOPTING MEASURES TO REDUCE POWER OUTAGES
ADOPTING MEASURES TO REDUCE POWER OUTAGESADOPTING MEASURES TO REDUCE POWER OUTAGES
ADOPTING MEASURES TO REDUCE POWER OUTAGES
 
GRID SIDE CONVERTER CONTROL IN DFIG BASED WIND SYSTEM USING ENHANCED HYSTERES...
GRID SIDE CONVERTER CONTROL IN DFIG BASED WIND SYSTEM USING ENHANCED HYSTERES...GRID SIDE CONVERTER CONTROL IN DFIG BASED WIND SYSTEM USING ENHANCED HYSTERES...
GRID SIDE CONVERTER CONTROL IN DFIG BASED WIND SYSTEM USING ENHANCED HYSTERES...
 
Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)
 
PREPARATION OF POROUS AND RECYCLABLE PVA-TIO2HYBRID HYDROGEL
PREPARATION OF POROUS AND RECYCLABLE PVA-TIO2HYBRID HYDROGELPREPARATION OF POROUS AND RECYCLABLE PVA-TIO2HYBRID HYDROGEL
PREPARATION OF POROUS AND RECYCLABLE PVA-TIO2HYBRID HYDROGEL
 
4th International Conference on Electrical Engineering (ELEC 2020)
4th International Conference on Electrical Engineering (ELEC 2020)4th International Conference on Electrical Engineering (ELEC 2020)
4th International Conference on Electrical Engineering (ELEC 2020)
 
Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)
 
4th International Conference on Bioscience & Engineering (BIEN 2020)
4th International Conference on Bioscience & Engineering (BIEN 2020) 4th International Conference on Bioscience & Engineering (BIEN 2020)
4th International Conference on Bioscience & Engineering (BIEN 2020)
 
Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)Electrical & Computer Engineering: An International Journal (ECIJ)
Electrical & Computer Engineering: An International Journal (ECIJ)
 
Ecij cfp
Ecij cfpEcij cfp
Ecij cfp
 
GRID SIDE CONVERTER CONTROL IN DFIG BASED WIND SYSTEM USING ENHANCED HYSTERES...
GRID SIDE CONVERTER CONTROL IN DFIG BASED WIND SYSTEM USING ENHANCED HYSTERES...GRID SIDE CONVERTER CONTROL IN DFIG BASED WIND SYSTEM USING ENHANCED HYSTERES...
GRID SIDE CONVERTER CONTROL IN DFIG BASED WIND SYSTEM USING ENHANCED HYSTERES...
 
UNION OF GRAVITATIONAL AND ELECTROMAGNETIC FIELDS ON THE BASIS OF NONTRADITIO...
UNION OF GRAVITATIONAL AND ELECTROMAGNETIC FIELDS ON THE BASIS OF NONTRADITIO...UNION OF GRAVITATIONAL AND ELECTROMAGNETIC FIELDS ON THE BASIS OF NONTRADITIO...
UNION OF GRAVITATIONAL AND ELECTROMAGNETIC FIELDS ON THE BASIS OF NONTRADITIO...
 
USING MACHINE LEARNING TO BUILD A SEMI-INTELLIGENT BOT
USING MACHINE LEARNING TO BUILD A SEMI-INTELLIGENT BOT USING MACHINE LEARNING TO BUILD A SEMI-INTELLIGENT BOT
USING MACHINE LEARNING TO BUILD A SEMI-INTELLIGENT BOT
 
MODELING AND SIMULATION OF SOLAR PHOTOVOLTAIC APPLICATION BASED MULTILEVEL IN...
MODELING AND SIMULATION OF SOLAR PHOTOVOLTAIC APPLICATION BASED MULTILEVEL IN...MODELING AND SIMULATION OF SOLAR PHOTOVOLTAIC APPLICATION BASED MULTILEVEL IN...
MODELING AND SIMULATION OF SOLAR PHOTOVOLTAIC APPLICATION BASED MULTILEVEL IN...
 
Investigation of Interleaved Boost Converter with Voltage multiplier for PV w...
Investigation of Interleaved Boost Converter with Voltage multiplier for PV w...Investigation of Interleaved Boost Converter with Voltage multiplier for PV w...
Investigation of Interleaved Boost Converter with Voltage multiplier for PV w...
 
A COMPARISON BETWEEN SWARM INTELLIGENCE ALGORITHMS FOR ROUTING PROBLEMS
A COMPARISON BETWEEN SWARM INTELLIGENCE ALGORITHMS FOR ROUTING PROBLEMSA COMPARISON BETWEEN SWARM INTELLIGENCE ALGORITHMS FOR ROUTING PROBLEMS
A COMPARISON BETWEEN SWARM INTELLIGENCE ALGORITHMS FOR ROUTING PROBLEMS
 

Recently uploaded

Python Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docxPython Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docxRamakrishna Reddy Bijjam
 
Holdier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfHoldier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfagholdier
 
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...ZurliaSoop
 
SOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning PresentationSOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning Presentationcamerronhm
 
PROCESS RECORDING FORMAT.docx
PROCESS      RECORDING        FORMAT.docxPROCESS      RECORDING        FORMAT.docx
PROCESS RECORDING FORMAT.docxPoojaSen20
 
psychiatric nursing HISTORY COLLECTION .docx
psychiatric  nursing HISTORY  COLLECTION  .docxpsychiatric  nursing HISTORY  COLLECTION  .docx
psychiatric nursing HISTORY COLLECTION .docxPoojaSen20
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfciinovamais
 
Unit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxUnit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxVishalSingh1417
 
Spellings Wk 3 English CAPS CARES Please Practise
Spellings Wk 3 English CAPS CARES Please PractiseSpellings Wk 3 English CAPS CARES Please Practise
Spellings Wk 3 English CAPS CARES Please PractiseAnaAcapella
 
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxSKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxAmanpreet Kaur
 
ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.MaryamAhmad92
 
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in DelhiRussian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhikauryashika82
 
The basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxThe basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxheathfieldcps1
 
Micro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfMicro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfPoh-Sun Goh
 
Mixin Classes in Odoo 17 How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17  How to Extend Models Using Mixin ClassesMixin Classes in Odoo 17  How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17 How to Extend Models Using Mixin ClassesCeline George
 
Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)Jisc
 
Seal of Good Local Governance (SGLG) 2024Final.pptx
Seal of Good Local Governance (SGLG) 2024Final.pptxSeal of Good Local Governance (SGLG) 2024Final.pptx
Seal of Good Local Governance (SGLG) 2024Final.pptxnegromaestrong
 
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...Nguyen Thanh Tu Collection
 
ICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptxICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptxAreebaZafar22
 
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptxMaritesTamaniVerdade
 

Recently uploaded (20)

Python Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docxPython Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docx
 
Holdier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfHoldier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdf
 
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
 
SOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning PresentationSOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning Presentation
 
PROCESS RECORDING FORMAT.docx
PROCESS      RECORDING        FORMAT.docxPROCESS      RECORDING        FORMAT.docx
PROCESS RECORDING FORMAT.docx
 
psychiatric nursing HISTORY COLLECTION .docx
psychiatric  nursing HISTORY  COLLECTION  .docxpsychiatric  nursing HISTORY  COLLECTION  .docx
psychiatric nursing HISTORY COLLECTION .docx
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdf
 
Unit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxUnit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptx
 
Spellings Wk 3 English CAPS CARES Please Practise
Spellings Wk 3 English CAPS CARES Please PractiseSpellings Wk 3 English CAPS CARES Please Practise
Spellings Wk 3 English CAPS CARES Please Practise
 
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxSKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
 
ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.
 
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in DelhiRussian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
 
The basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxThe basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptx
 
Micro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfMicro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdf
 
Mixin Classes in Odoo 17 How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17  How to Extend Models Using Mixin ClassesMixin Classes in Odoo 17  How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17 How to Extend Models Using Mixin Classes
 
Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)
 
Seal of Good Local Governance (SGLG) 2024Final.pptx
Seal of Good Local Governance (SGLG) 2024Final.pptxSeal of Good Local Governance (SGLG) 2024Final.pptx
Seal of Good Local Governance (SGLG) 2024Final.pptx
 
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
 
ICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptxICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptx
 
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
 

A LIP LOCALIZATION BASED VISUAL FEATURE EXTRACTION METHOD

  • 1. Electrical & Computer Engineering: An International Journal (ECIJ) Volume 4, Number 4, December 2015 DOI : 10.14810/ecij.2015.4403 23 A LIP LOCALIZATION BASED VISUAL FEATURE EXTRACTION METHOD Namrata Dave1 1 Department of Computer Engineering, Gujarat Technological University, India ABSTRACT This paper presents a lip localization based visual feature extraction method to segment lip region from image or video in real time. Lip localization and tracking is useful in many applications such as lip reading, lip synchronization, visual speech recognition, facial animation etc. To synchronize lip movements with input audio we need to first segment lip region from input image or video frame. This paper presents a novel color based approach for localization of lips, which is an early stage for tracking lips in real time. A phoneme is a basic unit of speech and a viseme is a visual representation of phoneme or shape of mouth while utterance of particular phoneme. The main goal of our work is to implement a system for synchronizing lips with the input speech. To extract visual features i.e. visemes from input video frame or image we have used HSV and YCbCr color model along with various morphological operations. We have developed algorithm to work with normal lighting conditions and natural facial images of female and male. KEYWORDS Lip localization, viseme extraction, visual feature extraction, HSV color model, YCbCr color model. 1. INTRODUCTION Visual features of speech i.e. visemes are useful in various fields, such as visual speech recognition [1], [2], audio-video synchronization, or speaker identification/localization in audio- video scenes [3]. It is observed that mostly visual information related to speech is contained in the movements of lip. However, a most difficult part in visual speech recognition is to find an efficient method for extracting visual speech features. It is very difficult to extract visual features from input image or video which is taken in different situations under varying lighting conditions, different pose and for different skin tones. Many algorithms have been proposed for lip localization and viseme extraction. Several groups are working with template models based on active shape models [4], deformable models [5] and dynamic contours [6]. A lot of work is already done on this topic since years. Many novel approaches have been proposed in this field. Most approaches are divided into two main groups. First group is using shape based sophisticated methods like active shape model, deformable models, snake etc. to extract features from input image or video. Another group is extracting features using different colour models using different spatial methods which work on pixel intensity, edge and colour information of lip and/or facial skin. In this paper, we presented image based approach to extract mouth region and then to discriminate visual features from it. Here we have tried to extract visemes for digits one to ten. 2. EXISTING METHODS The Lip segmentation has become an important issue in facial animation, automatic VSR processing and automatic face recognition. In such systems, the region of interest (ROI) which is
  • 2. Electrical & Computer Engineering: An International Journal (ECIJ) Volume 4, Number 4, December 2015 24 lips in our case must be detected in each frame. This procedure is normally done by face detection and extraction of the lip region. Earlier systems performed the lip segmentation in conjunction with the application of artificial markers (lipstick) on the lips. The application of lipstick enables the system to detect precisely the lips in the image data, but this procedure is inappropriate since it is uncomfortable for users and such systems can be operated only in specific constraint based applications. Thus, the main research efforts have been concentrated in the development of visual features based algorithms for lip segmentation. Many studies have shown that colour information of particular region can be applied to identify the skin or face in digital images. The main idea behind this approach is to transform the RGB signal into other representation where the mouth region can efficiently separate from other details in given image, so that segmentation of mouth can be done efficiently. Due to this fact, a large number of colour representations have been proposed. Coinaiz et al [7] used the hue component of the HSV representation to highlight the red colour which is assumed to be associated with the lips in the image. Later, the HSV colour space is used [8] for lip detection. They used the hue signal to locate the position of the lips in mouth region. The interior and exterior boundaries of lips are extracted using colour and edge information using a Markov Random Field (MRF) framework based on the lip area found. Some of the approaches used YCrCb colour space for the lips detection [9, 10]. Eveno et al [11] proposed a new colour mixture and chromatic transformation for lip segmentation in 2001. In their approach, a new method for transformation of the RGB colour space and a chromatic map was applied to separate out the lips and facial skin. They argued that their proposed approach is able to provide robust lip detection under variable lighting conditions. Later on Eveno et al introduced a novel method where the pseudo-hue [12] was applied for lip segmentation that has been embedded in an active contour framework. They applied the proposed algorithm for visual speech recognition and the results show significant improvement in terms of accuracy in lip modelling. Another method for mouth segmentation has been proposed by Liew in 2003 [13]. In their approach, they used new transformation method to convert given colour image into the CIE-Lab colour space and CIE-LUV colour space, and then they calculated a lip membership map using the fuzzy clustering algorithm. The ROI around the mouth can be identified from the face area after application of morphological filtering on given image. Guan [14] has given a new approach based on Discrete Hartley Transform (DHT) to improve the contrast between lip region and the other regions of face. In their paper, they presented method of extraction of lips by applying wavelet multi-scale edge detection across the C3 component of the DHT. Their method takes into account both the colour information and the geometric characteristic of image. 3. VISUAL FEATURE EXTRACTION METHOD. 3.1. Proposed Method Feature extraction process is preceded by a number of pre-processing steps to be done as shown in Figure 1. This involves face detection followed by ROI extraction. Then, the lips of the speaker are tracked in consecutive frames of recorded video. Following these steps, and given an informative set of features, the visual front-end module can proceed with feature extraction.
  • 3. Electrical & Computer Engineering: An International Journal (ECIJ) Volume 4, Number 4, December 2015 25 Figure 1. Visual Feature Extraction Process 3.2. Dataset and Implementation A viseme is the visual appearance of a phoneme or visual unit of speech in spoken language. However, currently there is no standard viseme table available to be used by all researchers. Due to this fact, a viseme table for this study was based on the work performed by Lee and Sook in 2002 [15, 16]. This was used in a table which effectively mapped all the possible phonemes to viseme. They have shown the typical 48 phonemes used in English language including silence and short pauses, grouped into their 14 viseme classes. Table 1: List of Phonemes & Viseme Used for Digits One to Five Word Phoneme Viseme One w ah n/w an n/v ao n v ao n Two t uw/d uw t uw Three th r iy/dh iy th r iy Four f ao r f ao r Five f ay v f ay v Six s ih k s s ih k s Seven s eh v ax n s eh v ax n Eight ey t ey t Nine n ay n n ay n Ten t eh n t eh n Face Detection ROI Extraction Mouth/ Lip Contour Estimation Visual Feature Extraction
  • 4. Electrical & Computer Engineering: An International Journal (ECIJ) Volume 4, Number 4, December 2015 26 We have recorded videos for 12 different viseme classes as listed in Table1. Table1 shows list of phonemes and corresponding viseme for digits one to ten. We extracted mouth parameters for each viseme class for 6 different speakers. Videos recorded for different speakers include female and male speakers’ videos. Videos are recorded in normal conditions using webcam. No artificial marker like lipstick is used for detecting lips. Videos are recorded in avi format with 30fps. In this section, we will discuss quick summary of how exactly lip features are extracted from the video of speech. Entire video is converted into frames. RGB colour scheme of the image is not suitable for immediate processing as it contains a lot of mixed information about lightness etc. Another colour scheme should be used. Very convenient colour scheme is YCbCr as it separates luminance, blue chrominance and red chrominance. This method is convenient as the mouth region in image contains high red and low blue components in comparison with other face regions. Figure 2: Overview of the Proposed Algorithm for Lip Feature Extraction The mouth region has a greater value of red chromaticity (Cr) than of blue chromaticity (Cb). The value of the chromaticity of the Cr component is increased by using its square value. On the other hand, the region of the mouth has a good response to the ratio Cb/Cr. Lip Feature Extraction Algorithm steps are described in figure 2. First, Video divided into frames (frame rate 30 frames per second). On each frame all operations described in algorithm are performed. We have used YCBCR colour space. Based on luminance and chromaticity information extracted from frame we are extracting mouth region surrounding Region of interest i.e. Lips. The ratio Cb/Cr is used to extract face region and eliminate background from input video frame. From the cropped face image we extracted lip area based on H and S value by using HSV colour model. On cropped lip image edge detection is applied first. Then to eliminate extra info morphological operations are applied to get final lip parameters. We have used these features to estimate mouth region with information such as height and width for each viseme. 3.2. Results Four video sequences were taken from different subjects, each one having at most than 3X30 frames. Those sequences were analysed using the proposed lip segmentation algorithm. The Face Region Extraction Lip features: Lip Width & Height Extracted Mouth Region (Region of Interest) Extracted Video divided into still image
  • 5. Electrical & Computer Engineering: An International Journal (ECIJ) Volume 4, Number 4, December 2015 27 figure 3 shows examples of segmentation results for each subject. The best results were obtained for the pale skinned, non-bearded subject. We have used s and h value of HSV colour space to identify lip region based on which all viseme are extracted. Sometimes due to poor illumination we are not get good results only based on H and S value so we also try-out combining results of contrast stretched and log applied images to get exactly mouth region. Using Lip Feature Extraction Algorithm we extracted Viseme for 10 different viseme classes. Figure 3 shows results of viseme V for four speakers. We tested our algorithm for many frames for each speaker. For testing purpose for each speaker we selected two videos and extracted frames to apply algorithm on it. On an average we are getting 85% result for all speakers. Figure 3: Results viseme 'v' 4. CONCLUSIONS After the investigation done on existing methods, we have implemented a lip localization based visual feature extraction method which gives good accuracy for our database. Our algorithm is suitable for offline applications which can be extended to work in real time applications. Another improvement to the program can be achieved by trying it for available datasets. We have planned to use results obtained to use for lip-reading application as future enhancement. ACKNOWLEDGEMENTS We would like to thank everyone in my friends, staff members and students who helped me in recording videos.
  • 6. Electrical & Computer Engineering: An International Journal (ECIJ) Volume 4, Number 4, December 2015 28 REFERENCES [1] Tsuhan Chen and R. R, Rao, “Audio-visual interaction in multimedia communication,” in Proc. Of IEEE Intern. Con$ on Acoustics, Speech and Signal Processing (ICASSP’97), vol. 1, 1997, pp. 179- 182. [2] X. Zhang and R. M. Mersereau, M. Clements, C. C. Broun, “Visual Speech Feature Extraction for Improved Speech Recognition,” in Proc. IEEE Intern. Con$ on Acoustics, Speech, and Signal Processing (ICASSP’O2), vol. 2, 2002, pp. 1993-1996. [3] D. Lo, R. A. Goubran, J. Gammal, G. Thompson, and D. Schulz, “Robust Joint Audio-Video Localization in Video Conferencing using Reliability Information,” in Proc. 20th IEEE Instrumentation and Measurement Technology Conf: (IMTC’O3), vol. 2, May 2003, pp. 141-418. [4] K. L. Sum, W. H. Lau, S. H. Leung, A. W. C. Liew, and K. W. Tse, “A New Optimization Procedure for Extracting the Point-based Lip Contour Using Active Shape Model,” in Proc. Of IEEE Intern. Con. on Acoustics, Speech and Signal Processing (ICASSP ’Ol), vol. 3, May 2001, pp. 1485-1488. [5] A. W. C Liew, S.H. Leung, and W.H. Lau, “Lip Contour Extraction Using a Deformable Model,” in Proc. of IEEE Intern. Con. on Image Processing (ICIP’OO), vol. 2, Sept. 2000, pp. 255-258. [6] R. Kaucic and A. Blake, “Accurate, Real-time, Unadomed Lip Tracking,” in Proc. of 6th International Conference on Computer Vision, 1998, pp. 370-375. [7] T. Coianiz, L. Torresani and B. Caprile, “2D Deformable Models for Visual Speech Analysis”, Proceedings of Springer, Speech reading by Humans and Machines, D.G. stork & M. E. Hennecke Eds., NY, 1996. [8] X. Zhang and R.M. Mersereau, “Lip Features extraction Towards an Automatic Speech-reading System”, Proceedings of ICIP00, Wa07.05, Vancouver, Canada,2000. [9] M. Heckmann, F. Berthommier and K. Kroschel, “Noise adaptive stream weighting in audio-visual speech recognition”, Proceedings of EURASIP J. Appl. Signal, vol. 2002, pp. 1260-1273, Nov. 2002. [10] J. Chaloupka, “Automatic Lips Reading for Audio-Visual Speech Processing and Recognition”, Proceedings of Of ICSLP, pp. 2505-2508, Jeju Island, Korea, Oct. 2004. [11] N. Eveno, A. Caplier, P.Y. Coulon. “A new color transformation for lips segmentation”, Proceedings of IEEE Fourth Workshop on Multimedia Signal, pp. 3-8, Cannes, France, 2001. [12] N. Eveno, A. Caplier and P. Coulon, “Accurate and Quasi-Automatic Lip Tracking”, Proceedings of IEEE Trans. Circuits Syst. Video Techn. 14(5): 706-715., 2004. [13] A.W.C. Liew, S.H. Leung, and W.H. Lau, “Segmentation of color lip images by spatial fuzzy clustering” Proceedings of IEEE Trans. Fuzzy Syst. vol. 11 no. 4, pp. 542-549, Aug. 2003. [14] Y. P. Guan, “Automatic Extraction of Lip Based on Wavelet Edge Detection”, Proceedings of the Eighth international Symposium on Symbolic and Numeric Algorithms, Scientific Computing SYNASC. IEEE Computer Society, pp. 125-132. Sept. 2006. [15] S. Lee and D. Yook, “Viseme recognition experiment using context dependent Hidden Markov Models”, Proceedings of Third Intl. Conf. on Intelligent Data Engineering and Automated learning, Vol. 2412 pp. 557-561, 2002. [16] N. Dave, N. M. Patel. "Phoneme and Viseme based Approach for Lip Synchronization.", International Journal of Signal Processing, Image Processing and Pattern Recognition. 7(3), pp. 385-394, 2014. Authors Namrata A Dave Assistant Professor, Computer Engineering Department, G H Patel College of Engg. & Tech. GUJARAT, INDIA.