Professional Documents
Culture Documents
Input
Voice
Speaker
Model
Futures
Extraction
Ref
Model
Decision
Accept
Or
Reject
I.INTRODUCTION
II.METHODOLOGY
The research methodology adopted in this work can be
graphically described in block diagram shown in Fig. 2. As
shown in Fig.2, the first step is the preprocessing step which
detect the silence period of the input voice. To do so, a
multiple level DWT is used to calculate the different threshold
levels. This step is very important step in any front end VRS
and it will be explained in more details in section IV.
Ref. Model
Input voice
output
Futures
Extraction
Preprocessing
Silence detection
Matching
FFT
Windowing
Mel-frequency warping
Threshold
Modified
Data History
Mallat's DWT
Mallat's DWT
DWT
output
Fig. 4 block diagram of the logarithms
III.WAVELET THEORY
Normalized amplitude
0.5
-0.5
-1
1000
2000
3000
4000
5000
Samples
6000
7000
8000
9000
10000
8000
9000
10000
yes word
1
0.8
Normalized amplitude
0.6
0.4
0.2
0
-0.2
-0.4
-0.6
Signa
l
-0.8
-1
1000
2000
3000
4000
5000 6000
Samples
7000
IV.FEATURE EXTRACTION
A block diagram of the structure of logarithms processor is a
shown in Fig. 4.
First, the input voice signal is applied to the silence detection
stage, which is very important in any front end speaker
recognition system. This stage, basically started uttering the
word.
V.CIRCUIT IMPLEMENTION
0.6
0.5
Normalized amplitude
0.4
0.3
0.2
0.1
0
-0.1
-0.2
-0.3
-0.4
1000
2000
3000
4000
5000
6000
Samples
7000
8000
9000
10000
Normalized amplitude
60
50
40
30
20
10
0
1000
2000
3000
4000
5000 6000
Samples
7000
8000
9000
10000
(1)
97
98
99
0.9753
1.5148
0.1110
Welcome
Reject
Welcome
100
0.5141
Welcome
years old
25
number of total
tries
100
number of
accepted tries
95
imposter B
20
imposter C
16
20
19
imposter D
34
15
13
imposter E
45
17
16
imposter F
24
50
CONCLUTION
Admin's MSE
Result
0.8166
Welcome
0.8575
Welcome
0.9105
Welcome
0.7185
Welcome
0.3445
Welcome
0.6574
Welcome
7
8
9
0.3493
1.0240
0.4610
Welcome
Welcome
Welcome
10
0.4566
Welcome
11
0.7713
Welcome
12
0.2649
Welcome
.....
REFERENCES
[1] Wei Han, Cheong-Fat Chan, Chiu-Sing Choy and Kong-Pang Pun, "An
Efficient MFCC Extraction Method in Speech Recognition", IEEE
ISCAS, 2006.
[2] De Krom G., "Consistency and reliability of voice quality ratings for
different types of speech fragments", J Speech Lang Hear Res., Oct.
1994.
[3] Campbell JP., "Speaker recognition: a tutorial", Proceedings of the IEEE,
85(9):143762, Sep., 1997.
[4] Shrawankar U, Thakare VM., "Techniques for feature extraction in speech
recognition system: a comparative study", International Journal of
Computer Applications in Engineering, Technology and Sciences
(IJCAETS), 41218, 2013.
[5] J. Xie and S. Jiang, A simple and fast algorithm for global k means
clustering, in the 2nd international workshop of Education Technology
and Computer Science (ETCS), vol. 2, pp. 3640, 2010
[6] A. Katsamanis, G. Papandreou, and P. Maragos, " Face Active Appearance
Modeling and Speech Acoustic Information to Recover Articulation",
IEEE Transactions on Audio, Speech, and Language Processing, Vol. 17,
No. 3, pp.411-422, 2009.
[7] Bhupinder Singh, Rupinder Kaur, Nidhi Devgun, Ramandeep Kaur, The
process of Feature Extraction in Automatic Speech Recognition System
for Computer Machine Interaction with Humans: A Review, IJARCSSE,
vol. 2, Issue 2, Feb. 2012.
[8] Muda L, Begam KM, Elamvazuthi I., "Voice recognition algorithms using
Mel Frequency Cepstral Coefficient (MFCC) and Dynamic Time
Warping (DTW) techniques", Journal of Computing, vol. 2(3):138143,
2010.
[9] Nearey TM., "Speech perception as pattern recognition", J Acoust. Soc.
Am., 101(6):324154, 1997.