SlideShare a Scribd company logo
1 of 7
Download to read offline
IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 04 Issue: 12 | Dec-2015, Available @ http://www.ijret.org 130
IDENTITY AUTHENTICATION USING VOICE BIOMETRICS
TECHNIQUE
U A Kamalu1
, A. Raji2
, V.I. Nnebedum3
1
University of Port-Harcourt, Nigeria, ugokamax@yahoo.com
2
University of Port-Harcourt, Nigeria, rajoys@yahoo.com
3
University of Port-Harcourt, Nigeria, nnebedum@yahoo.com
Abstract
Identification of people using name, appearances, badges, tags and register may be effective may be in a small organization.
However, as the size of the organization or society increases, these simple ways of identifying individual become ineffective.
Therefore, it may be necessary to employ additional and more sophisticated means of authenticating the identity of people as the
population increases. Voice Biometrics is a method by which individuals can be uniquely identified by evaluating one or more
distinguishing biological traits associated with the voice of such individuals. In this paper, an unconstrained text-independent
recognition system using the Gaussian Mixture Model was applied to match recorded voice to stored voice for the purpose of
identification of individual. Recorded voices were processed and stored in the enrollment phase while probing voices were used
for comparison in the verification/recognition phase of the system.
Keywords: Model, Biometric, verification, enrollment, database, authentication, matching, identity.
--------------------------------------------------------------------***---------------------------------------------------------------------
INTRODUCTION
As individual varies in sizes and shapes, so the voice varies
also. There are many biometrics methods used for
identification purpose. Fingerprint biometrics is perhaps the
most commonly deployed biometrics for authenticating the
identity of people but it has its demerits. Fingerprint is
intrusive and requires the cooperation of the concerned
individual to be effective. More so,injury and wounds to the
finger can negatively affect fingerprint and render the
system useless. In contrast, voice is not subject to injury, it
is non-intrusive and in some cases may not require the
cooperation of individual during verification and
recognition. The voice of a speaker can be recognized and
authenticated in several wayswith Speaker Recognition
Systems (SRS) but the most common methods used are the
Hidden Markov Model (HMM) and Gaussian Mixture
Model (GMM). Text-dependent voice recognition system
uses the HMM while text-independent applicationsemploys
the GMM. HMM uses fixed or prompt sentence of few
words for testing authentication and is slightly intrusive and
could be frustrating to some people. GMM do not use fixed
or prompt sentence but allow any utterances for the
identification of speaker because it uses the behavioral or
physiological characteristics of individual (Yiyin, 2008). In
this paper, GMM was used for the obvious reason of its
flexibility and more friendly appeal.Biometrics techniques
use physiological and/or behavioral characteristics for
authentication and identification of people rather than the
use of token. This technique uses personal characteristics
that cannot be easily duplicated (Tracy et al, 2006). The
biometric identifiers of various individuals are permanent
features unique to them and as such more reliable than token
or knowledge based authentication methods. The categories
of biometrics used for authentication and identification of
people are signature verification, hand geometry,
Fingerprint identification, DNA identification, facial
recognition, and voice recognition (Adewole et al, 2011).
The voice of a person is complex information-bearing
signals that depend on physical and behavioral
characteristics. In order to recognize a person based on
voice, the complexity must be reduced while keeping
sufficient information in the extracted voice feature vector.
Peopleā€™s voices can be distinguished based on the frequency
of the speech, pitch and tone as well as the amplitude. The
frequency of speech is measured based on the speed at
which the individual pronounce a word or make a speech.
The amplitude is based on the pitch or loudness of the
spoken word. Any of the above mentioned characteristics
can be used to distinguish voices of people from one
another. The voices at different mood of an individual can
vary andas such, the state of mood of an individual when the
voice is recorded must be taken into consideration when
comparing the voice to identify a person. To record the
voice of a person, microphone and recorders can beused but
recording can also bedone with software applications. The
availability and characteristics such as sensitivity, signal-to-
noise ratio and size were considered in the choice the
recording unit.
METHODOLOGY
The system involves the enrollment and verification phase.
During the enrollment, voices of individuals in the system
were recorded and the features were extracted and stored as
template in the database for each individual that enrolled.
For the verification phase, the recorded probing voice
sample collected was compared with those in the database to
determine if a match is found to authenticate the identity of
IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 04 Issue: 12 | Dec-2015, Available @ http://www.ijret.org 131
the new voice. Java Luna Eclipse software application tool
was used as the integrated development environment (IDE)
while SQL Server 2000 was used in the design of the
database of the system. The system makes use of the voice
of people to authenticate the identity of individuals that
enrolled in the system. The verification and authentication
of people was achieved from the one-to-one voice
comparison of various individuals who enrolled and that of
the assumed impostors in the system. Therefore, the entire
system comprises of the extraction of features vector from
the recorded voice, the modeling of the features and the
comparison of the voices collected for verification to
determine the identity of an individual in the system.
Voice Signal Recording
The voices of individuals were recorded both during
enrollment and verification phase using the java swing
sound recording tool since the IDE for the system is Java
Luna eclipse. The recorded voices at the enrollment were
then exported to the ā€œhouse audioā€ folder of the system
application for processing while the recorded voice during
verification face were exported to the ā€œmatch audioā€ folder
of the system. The recording tool is shown in figure 1 below
while a sample of one of the recorded voice is as shown in
figure 2.
Figure 1: The java recording tool
Voice Features
There are different categories of voice features that can be
used for physical interpretation of voice in a speaker
recognition system. The common ones are Short-term
spectral features, Spectra-temporal features, Voice source
features, High level features and Prosodic
features(Kinnunen and Li, 2009).Features extraction is a
vital step in the process of any speaker recognition system.
The acquired data (voice signal) from individual is
processed in this section of the system either during the
enrollment or during identification/verification. A set of
feature vectors or parameters from the speech signal
representing some speaker-specific information, which
results from complex transformations occurring at different
levels of the speech production are extracted and stored.
(Mathur et al, 2013). It is good to adopt audio parameters
which are related to voice characteristics that can distinguish
speakers. Voice recognition depends upon both low level
and high level information obtained from a personā€™s speech.
The high level information includes values like dialect,
accent, the talking style, the subject manner of context,
phonetics, prosodic and lexical information (Shriberg,
2007). Low-level features comprise of data such as
fundamental frequency, formant frequency, pitch, intensity,
rhythm, tone, spectral magnitude and bandwidths of an
individualā€™s voice. An ideal feature would possess the
following characteristics:
ā€¢ Robustness against noise and distortion
ā€¢ Easy to measure and determine from the speech signal
ā€¢ Difficult to mimic
ā€¢ Naturally and frequently occurring in the speech
ā€¢ Lower intra-speaker variability and high inter-speaker
variability
ā€¢ Less affected by speakerā€™s health or long term variations
in voice
Voice Signal Processing and Feature Extraction
Thevoice signal that will appeal to the ear of man can range
from 100Hz to about 4000Hz, using 6 bits per sample and
based on Nyquist sampling criterion, the voice were
sampled at least twice the maximum frequency (8000
samples per seconds). The bit rate was obtained using this
formula:
Bit rate = Sampling rate x Number of bit per sample
Bit rate = 8000 x 6 = 48000 bits per seconds.
The recorded raw voice is in waveform format which can be
represented as shown in figure 2. It comprises of several
components of complex signal from the voice.
Figure 2: Sample of recorded voice
In the simplest form, a typical raw voice signal in waveform
format is depicted in figure 3 with the amplitude of the voice
signal shown on the vertical axis while horizontal axis
represents the time.
Figure 3: A Representation of simple waveform of audio
signal X(t)
If the above is therecorded voice signal X(t) andsampled
with Nyquist criterion, then figure 4 can be used to represent
the sampled signal Xs(t) where Ts is the duration of
sampling interval.
IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 04 Issue: 12 | Dec-2015, Available @ http://www.ijret.org 132
Figure 4: Simple Representation of a Sampled signal Xs(t).
The signal depicted in figure 3 is sinusoidal and could be
represented mathematically by the expression [Weiji, 2012].
X(t) = U(t) = a cos(įæ³ t) + b sin(įæ³ t) (1)
The above equation can be transformed into complex form
with Eulerā€™s formula in which case,
U(t) = aejįæ³ t
+ be ā€“jįæ³ t
(2)
The expression in equation (2) represent a simple single
signal and for a complex ones like the voice signal
comprising of several components of these simple signals,
the above expression can be modified to cater for the
summation of the various components of the simple signals
and represented by equation (3) below.
U(t) =āˆ‘ (akejįæ³k t
+ bkeā€“jįæ³k t
) (3)
For periodic signals, the fundamental frequency is given by
ā„¦ = 2Ī /T such that equation (3) can be represented by the
general equation below.
U(t) =āˆ‘ Unejnā„¦t
(4)
Fourier analysis was used to get the value of Un coefficients
for each of the various simple waves that make up the voice
signal. The decomposition of the signal into its fundamental
sine wave using the Fourier transformation makes it possible
to extract features from the voice.The mel-frequency
cepstral coefficients (MFCCs) are common features in audio
processing. It is a popular feature that can be extracted and
used in speech processing and recognition. The MFCCs
features are in frequency domain and more accurate
compare to other features that are in time domain. It
represents the real cepstral of a windowed of a short-time
signal derived from the Fast Fourier Transform (FFT) of that
signal. It extracts parameters from the voice signal which
are similar to those used by human ear for hearing speech.
MFCCs are computed with the aid of a psycho-acoustically
motivated filterbank, followed by logarithmic compression
and discrete cosine transforms (DCT). Denoting the outputs
of an M-channel filterbank as Y(m), for m = 1 ā€¦ā€¦ā€¦ M,
the MFCCs are obtained as shown in the expression below
(Kinnunen and Li, 2009).
Where n is the index of the cepstral coefficient.
The Mel-scale is approximately linear up to 1 KHz and the
relationship between frequency (f) in Hz of the signal and
the Mel-scale is represented with the expression below
[Yiyin, 2008].
The algorithm used for extracting the spectra cepstrum
features in each frame x[n] of the recorded audio signal is as
indicated thus;
Step 1: Start
Step 2: Lunch SRS
Step 3: Perform DFT on x[n] (This computes the
spectrum of recorded voice).
Step 4: Compute the Logarithm of the result
obtained from step 3.
Step5: Perform the inverted DFT operation on the
result obtained from step4.
Step 6: Store the result of Step 5 in the DB
Step 7: Stop
The result of the inverted DFT on the spectrum extracts two
main frequencies which are the spectral envelope (low
frequency) and the excitation signals (high frequency) which
are stored and can be used for recognition during matching.
Figure 5shows variation in frequency spectrum of four
different people speaking the same sentence.
IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 04 Issue: 12 | Dec-2015, Available @ http://www.ijret.org 133
Figure 5: Variation of Frequency Spectrum for four
individuals speaking same thing.
Modeling and Matching
The Gaussian Mixture Models (GMMs) method uses the
speaker feature vectors directly for recognition purpose in a
text-independent system. Feature vectors of individuals from
such system can be viewed as the coefficients extracted
from the mathematical transformation of the voice signals
using the DFT or FFT. The whole feature vectors would
then represent a single point in the n-dimensional room for a
person. These points represent speaker characteristics which
can be used to compare speakers to each other by computing
the distance between the points. When the result of the
distance calculation between two points is small, it means
the probability or degree of similarity of the speakers are
high. GMMs give a statistical chance of belonging to a
speaker to be identified. It is a weighted sum of M-
component Gaussian probability densities functions
[Reynolds, 2005]. The idea behind GMM is that a mixture
of probability functions can be assumed as the combination
of a number of separate probability functions as seen in
Figure 6.
Figure 6: A simple example GMM in onedimension.
The black line represents the GMM, and the others represent
its components.
Using GMM as the classification likelihood function has
some merits which includes computationally inexpensive,
the statistical model is well-understood and for text-
independent tasks, it is insensitive to the temporal aspects of
the speech and model only the underlying distribution of
acoustic observations from a speaker.
The condition for testing if a feature vector belongs to a
speaker voice model is to compute a similarity score when
comparison is done. The similarity score is calculated based
on the observed feature vector, the background model and
the speaker voice model. The probability that a speaker
model belongs to the observed feature vector can be
computed using the equation below [Petter, 2012].
Log[P(m /O)]=Log[P(O / m)]āˆ’Log[P(O /mw )]
Where O is the observed feature vector, m is a speaker
model and mw is the background model. The score
computed from the above equation is referred to as Log-
Likelihood Ratio (LLR) score which solely dependent on
comparisons between the speaker model and the background
model. Any LLR score that is less than 0 indicates a
rejection and a score greater than 0 indicates a possible
match. A person whose voice is been verified will be
rejected or accepted based on a fixed threshold value set for
the value of Log [P(m /O)]. The set value should be
surpassed to be convinced that a speaker model belongs to
the observed feature vector. For this speaker verification
system, the threshold for acceptance was determined
through the mathematical computation and decision of the
software tool application. The algorithm for recognizing a
probe voice as part of the voices already stored in the DB or
rejecting the voice is as shown below.
Step 1: Start
Step 2: Lunch SRS
Step 3: Perform DFT on the probe voice.
Step 4: Compute the Logarithm of the result
obtained from step 3.
Step 5: Perform the inverted DFT operation on the
result obtained from step 4.
Step 6: Compute LLR for the result of step 5.
IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 04 Issue: 12 | Dec-2015, Available @ http://www.ijret.org 134
Step 7: Compare the result of step 6 with values of
voices in DB
Step 8: If the result of step 7ā‰„ 0.95
Then
a. Identify the nearest stored
voice
Else
b. Reject.
Step 9: Stop.
RESULTS AND DISCUSSION
The voice of 22 different peoplecomprises of male and
female individuals were recorded at the enrollment phase of
the system and some of the recorded voices are shown in the
house-audio folder in figure 7 below. However, only 14
probing voices of people who were randomly chosen from
the initial 22 people who participated during enrollment
were later used for matching purpose in the recognition
phase. The voices of these 14 individuals some of which are
shown in figure 8 were recorded at the verification stage and
compared with the earlier recorded 22 voices to determine if
there is a match to identify the person.The voices of various
individuals in the system were stored with an identification
name both at the enrollment and verification phase for the
purpose of recognition whenever there is a match after
comparison. Since the work is on text-independent system
that uses GMM, the individuals that participated were free
to alter any sentence of their choice or sing for considerable
time both during enrollment and verification phase.
Figure 7: Sample of Some Recorded voices during enrollment.
Figure 8: Sample of probing voices recorded for matching purpose
IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 04 Issue: 12 | Dec-2015, Available @ http://www.ijret.org 135
The voice matching code do a sequential one to one
comparison based on frequency of the voices recorded and
stored during enrollment and the probing voice.It compares
the probing voice with each of the 22 voices and returns a
value and the name of the most likely voice from 22 people
that enrolled. The value is the likelihood ratio of how closed
the probing voice is to any of the enrolled voices stored in
the database. Table 1 shows the representation of the
probability value of the likelihood rate that was return for
the matching of eight authentic individuals and that of two
other assumed impostors for each of the authentic person. It
shows value obtained for eight individuals whose voices
were recorded twice, one for enrollment and the second for
matching during recognition. The table also shows the value
obtained when 2 other people voices are used as impostors
on the eight people. The name of real owner of the voice and
the abbreviated name for their matching for each person are
shown. Similarly, the names of the assumed impostor are
also shown.
Table 1: Probability value obtained for matching of genuine
voice and Impostor voices.
S/N
Namesof
Genuinevoice
Genuinevalue
Impostor1name
Impostor1value
Impostor2name
Impostor2value
1
Imabong and
Imbg
1 Rajoys 0.73 Ihebom 0.69
2 Peter and Ptr 1 Ayo Raji 0.67 Susan 0.27
3
Akerele and
Akere
0.92 Ekpetiama 0.95 Ihebom 0.92
4 Ekpetiama&Ekpe 0.98 Nmodeme 0.45 Rhonda 0.36
5 Nmodeme&Nmod 1 Akerele 0.62 Pete lee 0.26
6 Ihebom and Ihbm 0.99 Akpada 0.69 Peter 0.42
7 Akpada and Akpa 0.99 Nmodeme 0.72
Ayo
Raji
0.53
8 H-MAN and HM 1 Akerele 0.45 kathleen 0.22
The second column in the above table indicates the names
and short form of the names of the genuine person whose
voice was recorded. The full name was used at the recording
of the voice during the enrolment phase while the short
abbreviated name was used to record the voice of the same
person during the recognition phase. The result of the
matching of two genuine voices from the same person
recorded at different time is reflected directly in beside each
name in the third column. The fourth and the sixth column
contain the names of the various assumed impostors whose
voices are to be compared with the genuine voice of a
person during recognition. The results of the matching
between the assumed impostorā€™s voices and the genuine
voice are also reflected beside each name of the impostors in
column 5 and column 7 respectively. From the table, the
value for matching of the voices of a genuine person is as
high as 1 for some people but for others, it varies from 0.92
to 0.99. Hence, the criteria for accepting a voice to be
genuine in this work were put at 0.97 and above. The
impostors also score far less than this range of acceptance
from the table. The only exception to this is the case of the
impostors in row 3 in the table where an impostor value is
even higher than the value of the genuine owner of the voice
while the second impostor has exactly the same value of
0.92 with the genuine owner of the voice. This was taken as
error initially when this was noticed during testing, but the
process was repeated again and similar abnormal results
were observed for the genuine owner of voice and the
impostors in row 3. This is a case of false acceptance and
false rejection in the algorithm for the matching of the
voices.
CONCLUSION
Voice signal from people like other biometrics method such
as fingerprint can be used for authenticating the identity of a
person. The identity authentication of people using voice has
some merits. It is easy to use and less intrusive into personal
life especially in forensic applications. It is also not affected
by physical injury and can be deploy for remote
authentication of people with or without their knowledge.
These advantages will make the technology to continue to
gain more popularity in years to come and researches in
these areas seems promising too.
Nevertheless, this technology of authenticating identity of
people from their voices has some downside. These include
the variation in people voice when they are sick, angry or
change in mood. All these conditions must be properly
accounted for both during enrollment and verification
phases to minimize the error rate in the system. In this
paper, voice signal was presented as another biometrics
parameter that can be measured to serves as the basis for
authenticating the identity of an individual. The result
presented for 8 people above was fairly good but cannot be
consider of being perfect authentication means. It may work
for forensic application by narrowing down the number of
suspects but there is need to take other measure to ensure a
reliable authentication especially when security matter of
high importance and huge financial transaction is of interest.
Therefore, the systems cannot be seen as an ultimate
security tool or the perfect solution but rather as attempted
approach to security. Majority of programmes and initiatives
going on in the current Information Technology driven
world involve collection of biometrics data from the citizen.
It can be leveraged on to have a robust biometric database
for the society using voice of individual in the world that is
now a global village.
REFERENCES
[1] Adewole, KS.,Abdulsalam, SO., Jimoh, RG (2011).
Application of Voice Biometrics as an Ecological and
Inexpensive Method of Authentication.
[2] Kinnunen T, Li H (2009) An overview of text-
independent speaker recognition: from features to
supervectors. Speech communication 52: 12-40.
[3] Mathur S, Choudhary SK, Vyas JM (2013) Speaker
Recognition System and its Forensic Implications.
[4] Petter, L. (2012).VoiceR, a System for Speaker
Recognition using Gaussian Mixture Models. Degree
Project in Computer Technology and Software
IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308
_______________________________________________________________________________________
Volume: 04 Issue: 12 | Dec-2015, Available @ http://www.ijret.org 136
Development, First Level Royal Institute of
Technology School of Technology and Health 136 40
Handen, Sweden.
[5] Reynolds, DA., (2005). Gaussian Mixture Models.MIT
Lincoln Laboratory, Lexington, MA USA.
[6] Shriberg E (2007) Higher-level features in speaker
recognition. Speaker classification I, Springer Berlin
Heidelberg 4343:241-259.
[7] Weiji, W. 2012, Introduction to Digital Signal and
System Analysis. Ventus Publishing Aps.
[8] Yiyin, Z. (2008). ā€œSpeaker Identification: E6820 Spring
Final Project Report.A Project Research work in
Columbia University.

More Related Content

What's hot

IRJET- Voice Command Execution with Speech Recognition and Synthesizer
IRJET- Voice Command Execution with Speech Recognition and SynthesizerIRJET- Voice Command Execution with Speech Recognition and Synthesizer
IRJET- Voice Command Execution with Speech Recognition and SynthesizerIRJET Journal
Ā 
Voice
VoiceVoice
Voicereplay21
Ā 
Bk4301345349
Bk4301345349Bk4301345349
Bk4301345349IJERA Editor
Ā 
Adaptive wavelet thresholding with robust hybrid features for text-independe...
Adaptive wavelet thresholding with robust hybrid features  for text-independe...Adaptive wavelet thresholding with robust hybrid features  for text-independe...
Adaptive wavelet thresholding with robust hybrid features for text-independe...IJECEIAES
Ā 
Implementation of Digital Hearing AID for Sensory Neural Impairment
Implementation of Digital Hearing AID for Sensory Neural ImpairmentImplementation of Digital Hearing AID for Sensory Neural Impairment
Implementation of Digital Hearing AID for Sensory Neural Impairmentijtsrd
Ā 
Financial Transactions in ATM Machines using Speech Signals
Financial Transactions in ATM Machines using Speech SignalsFinancial Transactions in ATM Machines using Speech Signals
Financial Transactions in ATM Machines using Speech SignalsIJERA Editor
Ā 
SPEECH ENHANCEMENT USING KERNEL AND NORMALIZED KERNEL AFFINE PROJECTION ALGOR...
SPEECH ENHANCEMENT USING KERNEL AND NORMALIZED KERNEL AFFINE PROJECTION ALGOR...SPEECH ENHANCEMENT USING KERNEL AND NORMALIZED KERNEL AFFINE PROJECTION ALGOR...
SPEECH ENHANCEMENT USING KERNEL AND NORMALIZED KERNEL AFFINE PROJECTION ALGOR...sipij
Ā 
Survey on Different Methods of Digital Audio Watermarking
Survey on Different Methods of Digital Audio WatermarkingSurvey on Different Methods of Digital Audio Watermarking
Survey on Different Methods of Digital Audio WatermarkingIJERA Editor
Ā 
An Optical Character Recognition for Handwritten Devanagari Script
An Optical Character Recognition for Handwritten Devanagari ScriptAn Optical Character Recognition for Handwritten Devanagari Script
An Optical Character Recognition for Handwritten Devanagari ScriptIJERA Editor
Ā 
Video Audio Interface for recognizing gestures of Indian sign Language
Video Audio Interface for recognizing gestures of Indian sign LanguageVideo Audio Interface for recognizing gestures of Indian sign Language
Video Audio Interface for recognizing gestures of Indian sign LanguageCSCJournals
Ā 
Speech Recognition Application for the Speech Impaired using the Android-base...
Speech Recognition Application for the Speech Impaired using the Android-base...Speech Recognition Application for the Speech Impaired using the Android-base...
Speech Recognition Application for the Speech Impaired using the Android-base...TELKOMNIKA JOURNAL
Ā 
Cochlear implant acoustic simulation model based on critical band filters
Cochlear implant acoustic simulation model based on critical band filtersCochlear implant acoustic simulation model based on critical band filters
Cochlear implant acoustic simulation model based on critical band filtersIAEME Publication
Ā 
Fraud Detection Using Signature Recognition
Fraud Detection Using Signature RecognitionFraud Detection Using Signature Recognition
Fraud Detection Using Signature RecognitionTejraj Thakor
Ā 
Fq2510361043
Fq2510361043Fq2510361043
Fq2510361043IJERA Editor
Ā 
HuirongWang_resume
HuirongWang_resumeHuirongWang_resume
HuirongWang_resumeHuirong Wang
Ā 
USER AUTHENTICATION USING NATIVE LANGUAGE PASSWORDS
USER AUTHENTICATION USING NATIVE LANGUAGE PASSWORDSUSER AUTHENTICATION USING NATIVE LANGUAGE PASSWORDS
USER AUTHENTICATION USING NATIVE LANGUAGE PASSWORDSIJNSA Journal
Ā 

What's hot (19)

IRJET- Voice Command Execution with Speech Recognition and Synthesizer
IRJET- Voice Command Execution with Speech Recognition and SynthesizerIRJET- Voice Command Execution with Speech Recognition and Synthesizer
IRJET- Voice Command Execution with Speech Recognition and Synthesizer
Ā 
Ijetcas14 390
Ijetcas14 390Ijetcas14 390
Ijetcas14 390
Ā 
Voice
VoiceVoice
Voice
Ā 
Bk4301345349
Bk4301345349Bk4301345349
Bk4301345349
Ā 
Adaptive wavelet thresholding with robust hybrid features for text-independe...
Adaptive wavelet thresholding with robust hybrid features  for text-independe...Adaptive wavelet thresholding with robust hybrid features  for text-independe...
Adaptive wavelet thresholding with robust hybrid features for text-independe...
Ā 
Implementation of Digital Hearing AID for Sensory Neural Impairment
Implementation of Digital Hearing AID for Sensory Neural ImpairmentImplementation of Digital Hearing AID for Sensory Neural Impairment
Implementation of Digital Hearing AID for Sensory Neural Impairment
Ā 
D017552025
D017552025D017552025
D017552025
Ā 
Financial Transactions in ATM Machines using Speech Signals
Financial Transactions in ATM Machines using Speech SignalsFinancial Transactions in ATM Machines using Speech Signals
Financial Transactions in ATM Machines using Speech Signals
Ā 
30
3030
30
Ā 
SPEECH ENHANCEMENT USING KERNEL AND NORMALIZED KERNEL AFFINE PROJECTION ALGOR...
SPEECH ENHANCEMENT USING KERNEL AND NORMALIZED KERNEL AFFINE PROJECTION ALGOR...SPEECH ENHANCEMENT USING KERNEL AND NORMALIZED KERNEL AFFINE PROJECTION ALGOR...
SPEECH ENHANCEMENT USING KERNEL AND NORMALIZED KERNEL AFFINE PROJECTION ALGOR...
Ā 
Survey on Different Methods of Digital Audio Watermarking
Survey on Different Methods of Digital Audio WatermarkingSurvey on Different Methods of Digital Audio Watermarking
Survey on Different Methods of Digital Audio Watermarking
Ā 
An Optical Character Recognition for Handwritten Devanagari Script
An Optical Character Recognition for Handwritten Devanagari ScriptAn Optical Character Recognition for Handwritten Devanagari Script
An Optical Character Recognition for Handwritten Devanagari Script
Ā 
Video Audio Interface for recognizing gestures of Indian sign Language
Video Audio Interface for recognizing gestures of Indian sign LanguageVideo Audio Interface for recognizing gestures of Indian sign Language
Video Audio Interface for recognizing gestures of Indian sign Language
Ā 
Speech Recognition Application for the Speech Impaired using the Android-base...
Speech Recognition Application for the Speech Impaired using the Android-base...Speech Recognition Application for the Speech Impaired using the Android-base...
Speech Recognition Application for the Speech Impaired using the Android-base...
Ā 
Cochlear implant acoustic simulation model based on critical band filters
Cochlear implant acoustic simulation model based on critical band filtersCochlear implant acoustic simulation model based on critical band filters
Cochlear implant acoustic simulation model based on critical band filters
Ā 
Fraud Detection Using Signature Recognition
Fraud Detection Using Signature RecognitionFraud Detection Using Signature Recognition
Fraud Detection Using Signature Recognition
Ā 
Fq2510361043
Fq2510361043Fq2510361043
Fq2510361043
Ā 
HuirongWang_resume
HuirongWang_resumeHuirongWang_resume
HuirongWang_resume
Ā 
USER AUTHENTICATION USING NATIVE LANGUAGE PASSWORDS
USER AUTHENTICATION USING NATIVE LANGUAGE PASSWORDSUSER AUTHENTICATION USING NATIVE LANGUAGE PASSWORDS
USER AUTHENTICATION USING NATIVE LANGUAGE PASSWORDS
Ā 

Similar to Identity authentication using voice biometrics technique

Speaker recognition in android
Speaker recognition in androidSpeaker recognition in android
Speaker recognition in androidAnshuli Mittal
Ā 
Speech Recognition
Speech RecognitionSpeech Recognition
Speech RecognitionAhmed Moawad
Ā 
Utterance Based Speaker Identification Using ANN
Utterance Based Speaker Identification Using ANNUtterance Based Speaker Identification Using ANN
Utterance Based Speaker Identification Using ANNIJCSEA Journal
Ā 
Utterance Based Speaker Identification Using ANN
Utterance Based Speaker Identification Using ANNUtterance Based Speaker Identification Using ANN
Utterance Based Speaker Identification Using ANNIJCSEA Journal
Ā 
Utterance based speaker identification
Utterance based speaker identificationUtterance based speaker identification
Utterance based speaker identificationIJCSEA Journal
Ā 
A Review Paper on Speech Based Emotion Detection Using Deep Learning
A Review Paper on Speech Based Emotion Detection Using Deep LearningA Review Paper on Speech Based Emotion Detection Using Deep Learning
A Review Paper on Speech Based Emotion Detection Using Deep LearningIRJET Journal
Ā 
A Survey on Speech Recognition with Language Specification
A Survey on Speech Recognition with Language SpecificationA Survey on Speech Recognition with Language Specification
A Survey on Speech Recognition with Language Specificationijtsrd
Ā 
Performance of different classifiers in speech recognition
Performance of different classifiers in speech recognitionPerformance of different classifiers in speech recognition
Performance of different classifiers in speech recognitioneSAT Journals
Ā 
Performance of different classifiers in speech recognition
Performance of different classifiers in speech recognitionPerformance of different classifiers in speech recognition
Performance of different classifiers in speech recognitioneSAT Publishing House
Ā 
Forensic and Automatic Speaker Recognition System
Forensic and Automatic Speaker Recognition System Forensic and Automatic Speaker Recognition System
Forensic and Automatic Speaker Recognition System IJECEIAES
Ā 
Voice-based Login System
Voice-based Login SystemVoice-based Login System
Voice-based Login SystemIRJET Journal
Ā 
IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...
IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...
IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...IRJET Journal
Ā 
Kc3517481754
Kc3517481754Kc3517481754
Kc3517481754IJERA Editor
Ā 
A Robust Speaker Identification System
A Robust Speaker Identification SystemA Robust Speaker Identification System
A Robust Speaker Identification Systemijtsrd
Ā 
Speaker Recognition System
Speaker  Recognition SystemSpeaker  Recognition System
Speaker Recognition SystemKiran Conquer
Ā 
Speech Emotion Recognition Using Machine Learning
Speech Emotion Recognition Using Machine LearningSpeech Emotion Recognition Using Machine Learning
Speech Emotion Recognition Using Machine LearningIRJET Journal
Ā 
Authentication System Based on the Combination of Voice Biometrics and OTP Ge...
Authentication System Based on the Combination of Voice Biometrics and OTP Ge...Authentication System Based on the Combination of Voice Biometrics and OTP Ge...
Authentication System Based on the Combination of Voice Biometrics and OTP Ge...ijtsrd
Ā 
Pattern recognition voice biometrics
Pattern recognition voice biometricsPattern recognition voice biometrics
Pattern recognition voice biometricsMazin Alwaaly
Ā 
Bachelors project summary
Bachelors project summaryBachelors project summary
Bachelors project summaryAditya Deshmukh
Ā 

Similar to Identity authentication using voice biometrics technique (20)

Speaker recognition in android
Speaker recognition in androidSpeaker recognition in android
Speaker recognition in android
Ā 
Speech Recognition
Speech RecognitionSpeech Recognition
Speech Recognition
Ā 
Utterance Based Speaker Identification Using ANN
Utterance Based Speaker Identification Using ANNUtterance Based Speaker Identification Using ANN
Utterance Based Speaker Identification Using ANN
Ā 
Utterance Based Speaker Identification Using ANN
Utterance Based Speaker Identification Using ANNUtterance Based Speaker Identification Using ANN
Utterance Based Speaker Identification Using ANN
Ā 
Utterance based speaker identification
Utterance based speaker identificationUtterance based speaker identification
Utterance based speaker identification
Ā 
A Review Paper on Speech Based Emotion Detection Using Deep Learning
A Review Paper on Speech Based Emotion Detection Using Deep LearningA Review Paper on Speech Based Emotion Detection Using Deep Learning
A Review Paper on Speech Based Emotion Detection Using Deep Learning
Ā 
A Survey on Speech Recognition with Language Specification
A Survey on Speech Recognition with Language SpecificationA Survey on Speech Recognition with Language Specification
A Survey on Speech Recognition with Language Specification
Ā 
Performance of different classifiers in speech recognition
Performance of different classifiers in speech recognitionPerformance of different classifiers in speech recognition
Performance of different classifiers in speech recognition
Ā 
Performance of different classifiers in speech recognition
Performance of different classifiers in speech recognitionPerformance of different classifiers in speech recognition
Performance of different classifiers in speech recognition
Ā 
Forensic and Automatic Speaker Recognition System
Forensic and Automatic Speaker Recognition System Forensic and Automatic Speaker Recognition System
Forensic and Automatic Speaker Recognition System
Ā 
Voice-based Login System
Voice-based Login SystemVoice-based Login System
Voice-based Login System
Ā 
IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...
IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...
IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...
Ā 
Kc3517481754
Kc3517481754Kc3517481754
Kc3517481754
Ā 
A Robust Speaker Identification System
A Robust Speaker Identification SystemA Robust Speaker Identification System
A Robust Speaker Identification System
Ā 
Speaker Recognition System
Speaker  Recognition SystemSpeaker  Recognition System
Speaker Recognition System
Ā 
Dy36749754
Dy36749754Dy36749754
Dy36749754
Ā 
Speech Emotion Recognition Using Machine Learning
Speech Emotion Recognition Using Machine LearningSpeech Emotion Recognition Using Machine Learning
Speech Emotion Recognition Using Machine Learning
Ā 
Authentication System Based on the Combination of Voice Biometrics and OTP Ge...
Authentication System Based on the Combination of Voice Biometrics and OTP Ge...Authentication System Based on the Combination of Voice Biometrics and OTP Ge...
Authentication System Based on the Combination of Voice Biometrics and OTP Ge...
Ā 
Pattern recognition voice biometrics
Pattern recognition voice biometricsPattern recognition voice biometrics
Pattern recognition voice biometrics
Ā 
Bachelors project summary
Bachelors project summaryBachelors project summary
Bachelors project summary
Ā 

More from eSAT Journals

Mechanical properties of hybrid fiber reinforced concrete for pavements
Mechanical properties of hybrid fiber reinforced concrete for pavementsMechanical properties of hybrid fiber reinforced concrete for pavements
Mechanical properties of hybrid fiber reinforced concrete for pavementseSAT Journals
Ā 
Material management in construction ā€“ a case study
Material management in construction ā€“ a case studyMaterial management in construction ā€“ a case study
Material management in construction ā€“ a case studyeSAT Journals
Ā 
Managing drought short term strategies in semi arid regions a case study
Managing drought    short term strategies in semi arid regions  a case studyManaging drought    short term strategies in semi arid regions  a case study
Managing drought short term strategies in semi arid regions a case studyeSAT Journals
Ā 
Life cycle cost analysis of overlay for an urban road in bangalore
Life cycle cost analysis of overlay for an urban road in bangaloreLife cycle cost analysis of overlay for an urban road in bangalore
Life cycle cost analysis of overlay for an urban road in bangaloreeSAT Journals
Ā 
Laboratory studies of dense bituminous mixes ii with reclaimed asphalt materials
Laboratory studies of dense bituminous mixes ii with reclaimed asphalt materialsLaboratory studies of dense bituminous mixes ii with reclaimed asphalt materials
Laboratory studies of dense bituminous mixes ii with reclaimed asphalt materialseSAT Journals
Ā 
Laboratory investigation of expansive soil stabilized with natural inorganic ...
Laboratory investigation of expansive soil stabilized with natural inorganic ...Laboratory investigation of expansive soil stabilized with natural inorganic ...
Laboratory investigation of expansive soil stabilized with natural inorganic ...eSAT Journals
Ā 
Influence of reinforcement on the behavior of hollow concrete block masonry p...
Influence of reinforcement on the behavior of hollow concrete block masonry p...Influence of reinforcement on the behavior of hollow concrete block masonry p...
Influence of reinforcement on the behavior of hollow concrete block masonry p...eSAT Journals
Ā 
Influence of compaction energy on soil stabilized with chemical stabilizer
Influence of compaction energy on soil stabilized with chemical stabilizerInfluence of compaction energy on soil stabilized with chemical stabilizer
Influence of compaction energy on soil stabilized with chemical stabilizereSAT Journals
Ā 
Geographical information system (gis) for water resources management
Geographical information system (gis) for water resources managementGeographical information system (gis) for water resources management
Geographical information system (gis) for water resources managementeSAT Journals
Ā 
Forest type mapping of bidar forest division, karnataka using geoinformatics ...
Forest type mapping of bidar forest division, karnataka using geoinformatics ...Forest type mapping of bidar forest division, karnataka using geoinformatics ...
Forest type mapping of bidar forest division, karnataka using geoinformatics ...eSAT Journals
Ā 
Factors influencing compressive strength of geopolymer concrete
Factors influencing compressive strength of geopolymer concreteFactors influencing compressive strength of geopolymer concrete
Factors influencing compressive strength of geopolymer concreteeSAT Journals
Ā 
Experimental investigation on circular hollow steel columns in filled with li...
Experimental investigation on circular hollow steel columns in filled with li...Experimental investigation on circular hollow steel columns in filled with li...
Experimental investigation on circular hollow steel columns in filled with li...eSAT Journals
Ā 
Experimental behavior of circular hsscfrc filled steel tubular columns under ...
Experimental behavior of circular hsscfrc filled steel tubular columns under ...Experimental behavior of circular hsscfrc filled steel tubular columns under ...
Experimental behavior of circular hsscfrc filled steel tubular columns under ...eSAT Journals
Ā 
Evaluation of punching shear in flat slabs
Evaluation of punching shear in flat slabsEvaluation of punching shear in flat slabs
Evaluation of punching shear in flat slabseSAT Journals
Ā 
Evaluation of performance of intake tower dam for recent earthquake in india
Evaluation of performance of intake tower dam for recent earthquake in indiaEvaluation of performance of intake tower dam for recent earthquake in india
Evaluation of performance of intake tower dam for recent earthquake in indiaeSAT Journals
Ā 
Evaluation of operational efficiency of urban road network using travel time ...
Evaluation of operational efficiency of urban road network using travel time ...Evaluation of operational efficiency of urban road network using travel time ...
Evaluation of operational efficiency of urban road network using travel time ...eSAT Journals
Ā 
Estimation of surface runoff in nallur amanikere watershed using scs cn method
Estimation of surface runoff in nallur amanikere watershed using scs cn methodEstimation of surface runoff in nallur amanikere watershed using scs cn method
Estimation of surface runoff in nallur amanikere watershed using scs cn methodeSAT Journals
Ā 
Estimation of morphometric parameters and runoff using rs & gis techniques
Estimation of morphometric parameters and runoff using rs & gis techniquesEstimation of morphometric parameters and runoff using rs & gis techniques
Estimation of morphometric parameters and runoff using rs & gis techniqueseSAT Journals
Ā 
Effect of variation of plastic hinge length on the results of non linear anal...
Effect of variation of plastic hinge length on the results of non linear anal...Effect of variation of plastic hinge length on the results of non linear anal...
Effect of variation of plastic hinge length on the results of non linear anal...eSAT Journals
Ā 
Effect of use of recycled materials on indirect tensile strength of asphalt c...
Effect of use of recycled materials on indirect tensile strength of asphalt c...Effect of use of recycled materials on indirect tensile strength of asphalt c...
Effect of use of recycled materials on indirect tensile strength of asphalt c...eSAT Journals
Ā 

More from eSAT Journals (20)

Mechanical properties of hybrid fiber reinforced concrete for pavements
Mechanical properties of hybrid fiber reinforced concrete for pavementsMechanical properties of hybrid fiber reinforced concrete for pavements
Mechanical properties of hybrid fiber reinforced concrete for pavements
Ā 
Material management in construction ā€“ a case study
Material management in construction ā€“ a case studyMaterial management in construction ā€“ a case study
Material management in construction ā€“ a case study
Ā 
Managing drought short term strategies in semi arid regions a case study
Managing drought    short term strategies in semi arid regions  a case studyManaging drought    short term strategies in semi arid regions  a case study
Managing drought short term strategies in semi arid regions a case study
Ā 
Life cycle cost analysis of overlay for an urban road in bangalore
Life cycle cost analysis of overlay for an urban road in bangaloreLife cycle cost analysis of overlay for an urban road in bangalore
Life cycle cost analysis of overlay for an urban road in bangalore
Ā 
Laboratory studies of dense bituminous mixes ii with reclaimed asphalt materials
Laboratory studies of dense bituminous mixes ii with reclaimed asphalt materialsLaboratory studies of dense bituminous mixes ii with reclaimed asphalt materials
Laboratory studies of dense bituminous mixes ii with reclaimed asphalt materials
Ā 
Laboratory investigation of expansive soil stabilized with natural inorganic ...
Laboratory investigation of expansive soil stabilized with natural inorganic ...Laboratory investigation of expansive soil stabilized with natural inorganic ...
Laboratory investigation of expansive soil stabilized with natural inorganic ...
Ā 
Influence of reinforcement on the behavior of hollow concrete block masonry p...
Influence of reinforcement on the behavior of hollow concrete block masonry p...Influence of reinforcement on the behavior of hollow concrete block masonry p...
Influence of reinforcement on the behavior of hollow concrete block masonry p...
Ā 
Influence of compaction energy on soil stabilized with chemical stabilizer
Influence of compaction energy on soil stabilized with chemical stabilizerInfluence of compaction energy on soil stabilized with chemical stabilizer
Influence of compaction energy on soil stabilized with chemical stabilizer
Ā 
Geographical information system (gis) for water resources management
Geographical information system (gis) for water resources managementGeographical information system (gis) for water resources management
Geographical information system (gis) for water resources management
Ā 
Forest type mapping of bidar forest division, karnataka using geoinformatics ...
Forest type mapping of bidar forest division, karnataka using geoinformatics ...Forest type mapping of bidar forest division, karnataka using geoinformatics ...
Forest type mapping of bidar forest division, karnataka using geoinformatics ...
Ā 
Factors influencing compressive strength of geopolymer concrete
Factors influencing compressive strength of geopolymer concreteFactors influencing compressive strength of geopolymer concrete
Factors influencing compressive strength of geopolymer concrete
Ā 
Experimental investigation on circular hollow steel columns in filled with li...
Experimental investigation on circular hollow steel columns in filled with li...Experimental investigation on circular hollow steel columns in filled with li...
Experimental investigation on circular hollow steel columns in filled with li...
Ā 
Experimental behavior of circular hsscfrc filled steel tubular columns under ...
Experimental behavior of circular hsscfrc filled steel tubular columns under ...Experimental behavior of circular hsscfrc filled steel tubular columns under ...
Experimental behavior of circular hsscfrc filled steel tubular columns under ...
Ā 
Evaluation of punching shear in flat slabs
Evaluation of punching shear in flat slabsEvaluation of punching shear in flat slabs
Evaluation of punching shear in flat slabs
Ā 
Evaluation of performance of intake tower dam for recent earthquake in india
Evaluation of performance of intake tower dam for recent earthquake in indiaEvaluation of performance of intake tower dam for recent earthquake in india
Evaluation of performance of intake tower dam for recent earthquake in india
Ā 
Evaluation of operational efficiency of urban road network using travel time ...
Evaluation of operational efficiency of urban road network using travel time ...Evaluation of operational efficiency of urban road network using travel time ...
Evaluation of operational efficiency of urban road network using travel time ...
Ā 
Estimation of surface runoff in nallur amanikere watershed using scs cn method
Estimation of surface runoff in nallur amanikere watershed using scs cn methodEstimation of surface runoff in nallur amanikere watershed using scs cn method
Estimation of surface runoff in nallur amanikere watershed using scs cn method
Ā 
Estimation of morphometric parameters and runoff using rs & gis techniques
Estimation of morphometric parameters and runoff using rs & gis techniquesEstimation of morphometric parameters and runoff using rs & gis techniques
Estimation of morphometric parameters and runoff using rs & gis techniques
Ā 
Effect of variation of plastic hinge length on the results of non linear anal...
Effect of variation of plastic hinge length on the results of non linear anal...Effect of variation of plastic hinge length on the results of non linear anal...
Effect of variation of plastic hinge length on the results of non linear anal...
Ā 
Effect of use of recycled materials on indirect tensile strength of asphalt c...
Effect of use of recycled materials on indirect tensile strength of asphalt c...Effect of use of recycled materials on indirect tensile strength of asphalt c...
Effect of use of recycled materials on indirect tensile strength of asphalt c...
Ā 

Recently uploaded

HARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IVHARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IVRajaP95
Ā 
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130Suhani Kapoor
Ā 
ZXCTN 5804 / ZTE PTN / ZTE POTN / ZTE 5804 PTN / ZTE POTN 5804 ( 100/200 GE Z...
ZXCTN 5804 / ZTE PTN / ZTE POTN / ZTE 5804 PTN / ZTE POTN 5804 ( 100/200 GE Z...ZXCTN 5804 / ZTE PTN / ZTE POTN / ZTE 5804 PTN / ZTE POTN 5804 ( 100/200 GE Z...
ZXCTN 5804 / ZTE PTN / ZTE POTN / ZTE 5804 PTN / ZTE POTN 5804 ( 100/200 GE Z...ZTE
Ā 
Sachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
Sachpazis Costas: Geotechnical Engineering: A student's Perspective IntroductionSachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
Sachpazis Costas: Geotechnical Engineering: A student's Perspective IntroductionDr.Costas Sachpazis
Ā 
Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024hassan khalil
Ā 
Oxy acetylene welding presentation note.
Oxy acetylene welding presentation note.Oxy acetylene welding presentation note.
Oxy acetylene welding presentation note.eptoze12
Ā 
GDSC ASEB Gen AI study jams presentation
GDSC ASEB Gen AI study jams presentationGDSC ASEB Gen AI study jams presentation
GDSC ASEB Gen AI study jams presentationGDSCAESB
Ā 
main PPT.pptx of girls hostel security using rfid
main PPT.pptx of girls hostel security using rfidmain PPT.pptx of girls hostel security using rfid
main PPT.pptx of girls hostel security using rfidNikhilNagaraju
Ā 
Artificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptxArtificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptxbritheesh05
Ā 
Biology for Computer Engineers Course Handout.pptx
Biology for Computer Engineers Course Handout.pptxBiology for Computer Engineers Course Handout.pptx
Biology for Computer Engineers Course Handout.pptxDeepakSakkari2
Ā 
VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130
VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130
VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130Suhani Kapoor
Ā 
chaitra-1.pptx fake news detection using machine learning
chaitra-1.pptx  fake news detection using machine learningchaitra-1.pptx  fake news detection using machine learning
chaitra-1.pptx fake news detection using machine learningmisbanausheenparvam
Ā 
Heart Disease Prediction using machine learning.pptx
Heart Disease Prediction using machine learning.pptxHeart Disease Prediction using machine learning.pptx
Heart Disease Prediction using machine learning.pptxPoojaBan
Ā 
HARMONY IN THE HUMAN BEING - Unit-II UHV-2
HARMONY IN THE HUMAN BEING - Unit-II UHV-2HARMONY IN THE HUMAN BEING - Unit-II UHV-2
HARMONY IN THE HUMAN BEING - Unit-II UHV-2RajaP95
Ā 
Current Transformer Drawing and GTP for MSETCL
Current Transformer Drawing and GTP for MSETCLCurrent Transformer Drawing and GTP for MSETCL
Current Transformer Drawing and GTP for MSETCLDeelipZope
Ā 
power system scada applications and uses
power system scada applications and usespower system scada applications and uses
power system scada applications and usesDevarapalliHaritha
Ā 
IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024Mark Billinghurst
Ā 
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdfCCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdfAsst.prof M.Gokilavani
Ā 
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escortsranjana rawat
Ā 

Recently uploaded (20)

HARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IVHARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IV
Ā 
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
Ā 
ZXCTN 5804 / ZTE PTN / ZTE POTN / ZTE 5804 PTN / ZTE POTN 5804 ( 100/200 GE Z...
ZXCTN 5804 / ZTE PTN / ZTE POTN / ZTE 5804 PTN / ZTE POTN 5804 ( 100/200 GE Z...ZXCTN 5804 / ZTE PTN / ZTE POTN / ZTE 5804 PTN / ZTE POTN 5804 ( 100/200 GE Z...
ZXCTN 5804 / ZTE PTN / ZTE POTN / ZTE 5804 PTN / ZTE POTN 5804 ( 100/200 GE Z...
Ā 
Sachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
Sachpazis Costas: Geotechnical Engineering: A student's Perspective IntroductionSachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
Sachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
Ā 
Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024
Ā 
Oxy acetylene welding presentation note.
Oxy acetylene welding presentation note.Oxy acetylene welding presentation note.
Oxy acetylene welding presentation note.
Ā 
GDSC ASEB Gen AI study jams presentation
GDSC ASEB Gen AI study jams presentationGDSC ASEB Gen AI study jams presentation
GDSC ASEB Gen AI study jams presentation
Ā 
main PPT.pptx of girls hostel security using rfid
main PPT.pptx of girls hostel security using rfidmain PPT.pptx of girls hostel security using rfid
main PPT.pptx of girls hostel security using rfid
Ā 
young call girls in Rajiv ChowkšŸ” 9953056974 šŸ” Delhi escort Service
young call girls in Rajiv ChowkšŸ” 9953056974 šŸ” Delhi escort Serviceyoung call girls in Rajiv ChowkšŸ” 9953056974 šŸ” Delhi escort Service
young call girls in Rajiv ChowkšŸ” 9953056974 šŸ” Delhi escort Service
Ā 
Artificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptxArtificial-Intelligence-in-Electronics (K).pptx
Artificial-Intelligence-in-Electronics (K).pptx
Ā 
Biology for Computer Engineers Course Handout.pptx
Biology for Computer Engineers Course Handout.pptxBiology for Computer Engineers Course Handout.pptx
Biology for Computer Engineers Course Handout.pptx
Ā 
VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130
VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130
VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130
Ā 
chaitra-1.pptx fake news detection using machine learning
chaitra-1.pptx  fake news detection using machine learningchaitra-1.pptx  fake news detection using machine learning
chaitra-1.pptx fake news detection using machine learning
Ā 
Heart Disease Prediction using machine learning.pptx
Heart Disease Prediction using machine learning.pptxHeart Disease Prediction using machine learning.pptx
Heart Disease Prediction using machine learning.pptx
Ā 
HARMONY IN THE HUMAN BEING - Unit-II UHV-2
HARMONY IN THE HUMAN BEING - Unit-II UHV-2HARMONY IN THE HUMAN BEING - Unit-II UHV-2
HARMONY IN THE HUMAN BEING - Unit-II UHV-2
Ā 
Current Transformer Drawing and GTP for MSETCL
Current Transformer Drawing and GTP for MSETCLCurrent Transformer Drawing and GTP for MSETCL
Current Transformer Drawing and GTP for MSETCL
Ā 
power system scada applications and uses
power system scada applications and usespower system scada applications and uses
power system scada applications and uses
Ā 
IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024
Ā 
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdfCCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
Ā 
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
Ā 

Identity authentication using voice biometrics technique

  • 1. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308 _______________________________________________________________________________________ Volume: 04 Issue: 12 | Dec-2015, Available @ http://www.ijret.org 130 IDENTITY AUTHENTICATION USING VOICE BIOMETRICS TECHNIQUE U A Kamalu1 , A. Raji2 , V.I. Nnebedum3 1 University of Port-Harcourt, Nigeria, ugokamax@yahoo.com 2 University of Port-Harcourt, Nigeria, rajoys@yahoo.com 3 University of Port-Harcourt, Nigeria, nnebedum@yahoo.com Abstract Identification of people using name, appearances, badges, tags and register may be effective may be in a small organization. However, as the size of the organization or society increases, these simple ways of identifying individual become ineffective. Therefore, it may be necessary to employ additional and more sophisticated means of authenticating the identity of people as the population increases. Voice Biometrics is a method by which individuals can be uniquely identified by evaluating one or more distinguishing biological traits associated with the voice of such individuals. In this paper, an unconstrained text-independent recognition system using the Gaussian Mixture Model was applied to match recorded voice to stored voice for the purpose of identification of individual. Recorded voices were processed and stored in the enrollment phase while probing voices were used for comparison in the verification/recognition phase of the system. Keywords: Model, Biometric, verification, enrollment, database, authentication, matching, identity. --------------------------------------------------------------------***--------------------------------------------------------------------- INTRODUCTION As individual varies in sizes and shapes, so the voice varies also. There are many biometrics methods used for identification purpose. Fingerprint biometrics is perhaps the most commonly deployed biometrics for authenticating the identity of people but it has its demerits. Fingerprint is intrusive and requires the cooperation of the concerned individual to be effective. More so,injury and wounds to the finger can negatively affect fingerprint and render the system useless. In contrast, voice is not subject to injury, it is non-intrusive and in some cases may not require the cooperation of individual during verification and recognition. The voice of a speaker can be recognized and authenticated in several wayswith Speaker Recognition Systems (SRS) but the most common methods used are the Hidden Markov Model (HMM) and Gaussian Mixture Model (GMM). Text-dependent voice recognition system uses the HMM while text-independent applicationsemploys the GMM. HMM uses fixed or prompt sentence of few words for testing authentication and is slightly intrusive and could be frustrating to some people. GMM do not use fixed or prompt sentence but allow any utterances for the identification of speaker because it uses the behavioral or physiological characteristics of individual (Yiyin, 2008). In this paper, GMM was used for the obvious reason of its flexibility and more friendly appeal.Biometrics techniques use physiological and/or behavioral characteristics for authentication and identification of people rather than the use of token. This technique uses personal characteristics that cannot be easily duplicated (Tracy et al, 2006). The biometric identifiers of various individuals are permanent features unique to them and as such more reliable than token or knowledge based authentication methods. The categories of biometrics used for authentication and identification of people are signature verification, hand geometry, Fingerprint identification, DNA identification, facial recognition, and voice recognition (Adewole et al, 2011). The voice of a person is complex information-bearing signals that depend on physical and behavioral characteristics. In order to recognize a person based on voice, the complexity must be reduced while keeping sufficient information in the extracted voice feature vector. Peopleā€™s voices can be distinguished based on the frequency of the speech, pitch and tone as well as the amplitude. The frequency of speech is measured based on the speed at which the individual pronounce a word or make a speech. The amplitude is based on the pitch or loudness of the spoken word. Any of the above mentioned characteristics can be used to distinguish voices of people from one another. The voices at different mood of an individual can vary andas such, the state of mood of an individual when the voice is recorded must be taken into consideration when comparing the voice to identify a person. To record the voice of a person, microphone and recorders can beused but recording can also bedone with software applications. The availability and characteristics such as sensitivity, signal-to- noise ratio and size were considered in the choice the recording unit. METHODOLOGY The system involves the enrollment and verification phase. During the enrollment, voices of individuals in the system were recorded and the features were extracted and stored as template in the database for each individual that enrolled. For the verification phase, the recorded probing voice sample collected was compared with those in the database to determine if a match is found to authenticate the identity of
  • 2. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308 _______________________________________________________________________________________ Volume: 04 Issue: 12 | Dec-2015, Available @ http://www.ijret.org 131 the new voice. Java Luna Eclipse software application tool was used as the integrated development environment (IDE) while SQL Server 2000 was used in the design of the database of the system. The system makes use of the voice of people to authenticate the identity of individuals that enrolled in the system. The verification and authentication of people was achieved from the one-to-one voice comparison of various individuals who enrolled and that of the assumed impostors in the system. Therefore, the entire system comprises of the extraction of features vector from the recorded voice, the modeling of the features and the comparison of the voices collected for verification to determine the identity of an individual in the system. Voice Signal Recording The voices of individuals were recorded both during enrollment and verification phase using the java swing sound recording tool since the IDE for the system is Java Luna eclipse. The recorded voices at the enrollment were then exported to the ā€œhouse audioā€ folder of the system application for processing while the recorded voice during verification face were exported to the ā€œmatch audioā€ folder of the system. The recording tool is shown in figure 1 below while a sample of one of the recorded voice is as shown in figure 2. Figure 1: The java recording tool Voice Features There are different categories of voice features that can be used for physical interpretation of voice in a speaker recognition system. The common ones are Short-term spectral features, Spectra-temporal features, Voice source features, High level features and Prosodic features(Kinnunen and Li, 2009).Features extraction is a vital step in the process of any speaker recognition system. The acquired data (voice signal) from individual is processed in this section of the system either during the enrollment or during identification/verification. A set of feature vectors or parameters from the speech signal representing some speaker-specific information, which results from complex transformations occurring at different levels of the speech production are extracted and stored. (Mathur et al, 2013). It is good to adopt audio parameters which are related to voice characteristics that can distinguish speakers. Voice recognition depends upon both low level and high level information obtained from a personā€™s speech. The high level information includes values like dialect, accent, the talking style, the subject manner of context, phonetics, prosodic and lexical information (Shriberg, 2007). Low-level features comprise of data such as fundamental frequency, formant frequency, pitch, intensity, rhythm, tone, spectral magnitude and bandwidths of an individualā€™s voice. An ideal feature would possess the following characteristics: ā€¢ Robustness against noise and distortion ā€¢ Easy to measure and determine from the speech signal ā€¢ Difficult to mimic ā€¢ Naturally and frequently occurring in the speech ā€¢ Lower intra-speaker variability and high inter-speaker variability ā€¢ Less affected by speakerā€™s health or long term variations in voice Voice Signal Processing and Feature Extraction Thevoice signal that will appeal to the ear of man can range from 100Hz to about 4000Hz, using 6 bits per sample and based on Nyquist sampling criterion, the voice were sampled at least twice the maximum frequency (8000 samples per seconds). The bit rate was obtained using this formula: Bit rate = Sampling rate x Number of bit per sample Bit rate = 8000 x 6 = 48000 bits per seconds. The recorded raw voice is in waveform format which can be represented as shown in figure 2. It comprises of several components of complex signal from the voice. Figure 2: Sample of recorded voice In the simplest form, a typical raw voice signal in waveform format is depicted in figure 3 with the amplitude of the voice signal shown on the vertical axis while horizontal axis represents the time. Figure 3: A Representation of simple waveform of audio signal X(t) If the above is therecorded voice signal X(t) andsampled with Nyquist criterion, then figure 4 can be used to represent the sampled signal Xs(t) where Ts is the duration of sampling interval.
  • 3. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308 _______________________________________________________________________________________ Volume: 04 Issue: 12 | Dec-2015, Available @ http://www.ijret.org 132 Figure 4: Simple Representation of a Sampled signal Xs(t). The signal depicted in figure 3 is sinusoidal and could be represented mathematically by the expression [Weiji, 2012]. X(t) = U(t) = a cos(įæ³ t) + b sin(įæ³ t) (1) The above equation can be transformed into complex form with Eulerā€™s formula in which case, U(t) = aejįæ³ t + be ā€“jįæ³ t (2) The expression in equation (2) represent a simple single signal and for a complex ones like the voice signal comprising of several components of these simple signals, the above expression can be modified to cater for the summation of the various components of the simple signals and represented by equation (3) below. U(t) =āˆ‘ (akejįæ³k t + bkeā€“jįæ³k t ) (3) For periodic signals, the fundamental frequency is given by ā„¦ = 2Ī /T such that equation (3) can be represented by the general equation below. U(t) =āˆ‘ Unejnā„¦t (4) Fourier analysis was used to get the value of Un coefficients for each of the various simple waves that make up the voice signal. The decomposition of the signal into its fundamental sine wave using the Fourier transformation makes it possible to extract features from the voice.The mel-frequency cepstral coefficients (MFCCs) are common features in audio processing. It is a popular feature that can be extracted and used in speech processing and recognition. The MFCCs features are in frequency domain and more accurate compare to other features that are in time domain. It represents the real cepstral of a windowed of a short-time signal derived from the Fast Fourier Transform (FFT) of that signal. It extracts parameters from the voice signal which are similar to those used by human ear for hearing speech. MFCCs are computed with the aid of a psycho-acoustically motivated filterbank, followed by logarithmic compression and discrete cosine transforms (DCT). Denoting the outputs of an M-channel filterbank as Y(m), for m = 1 ā€¦ā€¦ā€¦ M, the MFCCs are obtained as shown in the expression below (Kinnunen and Li, 2009). Where n is the index of the cepstral coefficient. The Mel-scale is approximately linear up to 1 KHz and the relationship between frequency (f) in Hz of the signal and the Mel-scale is represented with the expression below [Yiyin, 2008]. The algorithm used for extracting the spectra cepstrum features in each frame x[n] of the recorded audio signal is as indicated thus; Step 1: Start Step 2: Lunch SRS Step 3: Perform DFT on x[n] (This computes the spectrum of recorded voice). Step 4: Compute the Logarithm of the result obtained from step 3. Step5: Perform the inverted DFT operation on the result obtained from step4. Step 6: Store the result of Step 5 in the DB Step 7: Stop The result of the inverted DFT on the spectrum extracts two main frequencies which are the spectral envelope (low frequency) and the excitation signals (high frequency) which are stored and can be used for recognition during matching. Figure 5shows variation in frequency spectrum of four different people speaking the same sentence.
  • 4. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308 _______________________________________________________________________________________ Volume: 04 Issue: 12 | Dec-2015, Available @ http://www.ijret.org 133 Figure 5: Variation of Frequency Spectrum for four individuals speaking same thing. Modeling and Matching The Gaussian Mixture Models (GMMs) method uses the speaker feature vectors directly for recognition purpose in a text-independent system. Feature vectors of individuals from such system can be viewed as the coefficients extracted from the mathematical transformation of the voice signals using the DFT or FFT. The whole feature vectors would then represent a single point in the n-dimensional room for a person. These points represent speaker characteristics which can be used to compare speakers to each other by computing the distance between the points. When the result of the distance calculation between two points is small, it means the probability or degree of similarity of the speakers are high. GMMs give a statistical chance of belonging to a speaker to be identified. It is a weighted sum of M- component Gaussian probability densities functions [Reynolds, 2005]. The idea behind GMM is that a mixture of probability functions can be assumed as the combination of a number of separate probability functions as seen in Figure 6. Figure 6: A simple example GMM in onedimension. The black line represents the GMM, and the others represent its components. Using GMM as the classification likelihood function has some merits which includes computationally inexpensive, the statistical model is well-understood and for text- independent tasks, it is insensitive to the temporal aspects of the speech and model only the underlying distribution of acoustic observations from a speaker. The condition for testing if a feature vector belongs to a speaker voice model is to compute a similarity score when comparison is done. The similarity score is calculated based on the observed feature vector, the background model and the speaker voice model. The probability that a speaker model belongs to the observed feature vector can be computed using the equation below [Petter, 2012]. Log[P(m /O)]=Log[P(O / m)]āˆ’Log[P(O /mw )] Where O is the observed feature vector, m is a speaker model and mw is the background model. The score computed from the above equation is referred to as Log- Likelihood Ratio (LLR) score which solely dependent on comparisons between the speaker model and the background model. Any LLR score that is less than 0 indicates a rejection and a score greater than 0 indicates a possible match. A person whose voice is been verified will be rejected or accepted based on a fixed threshold value set for the value of Log [P(m /O)]. The set value should be surpassed to be convinced that a speaker model belongs to the observed feature vector. For this speaker verification system, the threshold for acceptance was determined through the mathematical computation and decision of the software tool application. The algorithm for recognizing a probe voice as part of the voices already stored in the DB or rejecting the voice is as shown below. Step 1: Start Step 2: Lunch SRS Step 3: Perform DFT on the probe voice. Step 4: Compute the Logarithm of the result obtained from step 3. Step 5: Perform the inverted DFT operation on the result obtained from step 4. Step 6: Compute LLR for the result of step 5.
  • 5. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308 _______________________________________________________________________________________ Volume: 04 Issue: 12 | Dec-2015, Available @ http://www.ijret.org 134 Step 7: Compare the result of step 6 with values of voices in DB Step 8: If the result of step 7ā‰„ 0.95 Then a. Identify the nearest stored voice Else b. Reject. Step 9: Stop. RESULTS AND DISCUSSION The voice of 22 different peoplecomprises of male and female individuals were recorded at the enrollment phase of the system and some of the recorded voices are shown in the house-audio folder in figure 7 below. However, only 14 probing voices of people who were randomly chosen from the initial 22 people who participated during enrollment were later used for matching purpose in the recognition phase. The voices of these 14 individuals some of which are shown in figure 8 were recorded at the verification stage and compared with the earlier recorded 22 voices to determine if there is a match to identify the person.The voices of various individuals in the system were stored with an identification name both at the enrollment and verification phase for the purpose of recognition whenever there is a match after comparison. Since the work is on text-independent system that uses GMM, the individuals that participated were free to alter any sentence of their choice or sing for considerable time both during enrollment and verification phase. Figure 7: Sample of Some Recorded voices during enrollment. Figure 8: Sample of probing voices recorded for matching purpose
  • 6. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308 _______________________________________________________________________________________ Volume: 04 Issue: 12 | Dec-2015, Available @ http://www.ijret.org 135 The voice matching code do a sequential one to one comparison based on frequency of the voices recorded and stored during enrollment and the probing voice.It compares the probing voice with each of the 22 voices and returns a value and the name of the most likely voice from 22 people that enrolled. The value is the likelihood ratio of how closed the probing voice is to any of the enrolled voices stored in the database. Table 1 shows the representation of the probability value of the likelihood rate that was return for the matching of eight authentic individuals and that of two other assumed impostors for each of the authentic person. It shows value obtained for eight individuals whose voices were recorded twice, one for enrollment and the second for matching during recognition. The table also shows the value obtained when 2 other people voices are used as impostors on the eight people. The name of real owner of the voice and the abbreviated name for their matching for each person are shown. Similarly, the names of the assumed impostor are also shown. Table 1: Probability value obtained for matching of genuine voice and Impostor voices. S/N Namesof Genuinevoice Genuinevalue Impostor1name Impostor1value Impostor2name Impostor2value 1 Imabong and Imbg 1 Rajoys 0.73 Ihebom 0.69 2 Peter and Ptr 1 Ayo Raji 0.67 Susan 0.27 3 Akerele and Akere 0.92 Ekpetiama 0.95 Ihebom 0.92 4 Ekpetiama&Ekpe 0.98 Nmodeme 0.45 Rhonda 0.36 5 Nmodeme&Nmod 1 Akerele 0.62 Pete lee 0.26 6 Ihebom and Ihbm 0.99 Akpada 0.69 Peter 0.42 7 Akpada and Akpa 0.99 Nmodeme 0.72 Ayo Raji 0.53 8 H-MAN and HM 1 Akerele 0.45 kathleen 0.22 The second column in the above table indicates the names and short form of the names of the genuine person whose voice was recorded. The full name was used at the recording of the voice during the enrolment phase while the short abbreviated name was used to record the voice of the same person during the recognition phase. The result of the matching of two genuine voices from the same person recorded at different time is reflected directly in beside each name in the third column. The fourth and the sixth column contain the names of the various assumed impostors whose voices are to be compared with the genuine voice of a person during recognition. The results of the matching between the assumed impostorā€™s voices and the genuine voice are also reflected beside each name of the impostors in column 5 and column 7 respectively. From the table, the value for matching of the voices of a genuine person is as high as 1 for some people but for others, it varies from 0.92 to 0.99. Hence, the criteria for accepting a voice to be genuine in this work were put at 0.97 and above. The impostors also score far less than this range of acceptance from the table. The only exception to this is the case of the impostors in row 3 in the table where an impostor value is even higher than the value of the genuine owner of the voice while the second impostor has exactly the same value of 0.92 with the genuine owner of the voice. This was taken as error initially when this was noticed during testing, but the process was repeated again and similar abnormal results were observed for the genuine owner of voice and the impostors in row 3. This is a case of false acceptance and false rejection in the algorithm for the matching of the voices. CONCLUSION Voice signal from people like other biometrics method such as fingerprint can be used for authenticating the identity of a person. The identity authentication of people using voice has some merits. It is easy to use and less intrusive into personal life especially in forensic applications. It is also not affected by physical injury and can be deploy for remote authentication of people with or without their knowledge. These advantages will make the technology to continue to gain more popularity in years to come and researches in these areas seems promising too. Nevertheless, this technology of authenticating identity of people from their voices has some downside. These include the variation in people voice when they are sick, angry or change in mood. All these conditions must be properly accounted for both during enrollment and verification phases to minimize the error rate in the system. In this paper, voice signal was presented as another biometrics parameter that can be measured to serves as the basis for authenticating the identity of an individual. The result presented for 8 people above was fairly good but cannot be consider of being perfect authentication means. It may work for forensic application by narrowing down the number of suspects but there is need to take other measure to ensure a reliable authentication especially when security matter of high importance and huge financial transaction is of interest. Therefore, the systems cannot be seen as an ultimate security tool or the perfect solution but rather as attempted approach to security. Majority of programmes and initiatives going on in the current Information Technology driven world involve collection of biometrics data from the citizen. It can be leveraged on to have a robust biometric database for the society using voice of individual in the world that is now a global village. REFERENCES [1] Adewole, KS.,Abdulsalam, SO., Jimoh, RG (2011). Application of Voice Biometrics as an Ecological and Inexpensive Method of Authentication. [2] Kinnunen T, Li H (2009) An overview of text- independent speaker recognition: from features to supervectors. Speech communication 52: 12-40. [3] Mathur S, Choudhary SK, Vyas JM (2013) Speaker Recognition System and its Forensic Implications. [4] Petter, L. (2012).VoiceR, a System for Speaker Recognition using Gaussian Mixture Models. Degree Project in Computer Technology and Software
  • 7. IJRET: International Journal of Research in Engineering and Technology eISSN: 2319-1163 | pISSN: 2321-7308 _______________________________________________________________________________________ Volume: 04 Issue: 12 | Dec-2015, Available @ http://www.ijret.org 136 Development, First Level Royal Institute of Technology School of Technology and Health 136 40 Handen, Sweden. [5] Reynolds, DA., (2005). Gaussian Mixture Models.MIT Lincoln Laboratory, Lexington, MA USA. [6] Shriberg E (2007) Higher-level features in speaker recognition. Speaker classification I, Springer Berlin Heidelberg 4343:241-259. [7] Weiji, W. 2012, Introduction to Digital Signal and System Analysis. Ventus Publishing Aps. [8] Yiyin, Z. (2008). ā€œSpeaker Identification: E6820 Spring Final Project Report.A Project Research work in Columbia University.