You are on page 1of 3

Implementation of PCA & ICA for Voice ecognition and Separation of Speech

Author A. Nitin Kandpal Author B. B. Madhusudan Rao


Sost Department 121T Embedded Department 121T
Pune, India Pune, India
savera.2020@gmail.com

Abstract-Principle Component Analysis is great to evaluate


the correlation among variable and reduce data dimensionally
without loss of any data. The ability of analyzing the property
of voice, reducing noises and extracting the valuable data of
voice makes PCA an integral part of voice recognition. In
digital signal processing signal estimation is required; signal
may be superimposed by several interfering sources. To find
one desired source signal Independent Component Analysis
can be implemented. ICA recovers a set of independent signal
from a set of measured signals by using statistical analysis of
signal.

I. INTRODUCTION
-0.8,'---_0
'---:---=---'--=---:---::---"
PCA is the method of modem data analysis. When we are
Figure 2. IOlxlO matrix data
observing some complex system like neuroscience, web
indexing, meteorology and oceanography it is tuff to figure
Let's take an example how peA is used to illustrate
out what happing because data appears clouded, unclear and
microarray experiment data shows 101 genes of 101peopie
even redundant.
this can form the matrix of 101x101 so consist 101 axes per
The goal of peA is to figure out new most important
genes due to that data result is cloud values for
basis to re-express the data set. The new basis filters out multidimensional array. From fig 1 data is unclear and
noise and reveals hidden structure [1]. peA is useful when clouded after applying peA data reduced to 101xlO graph
number of variable have to compute and wish to reduce the shows in fig 2. This describes the characteristics of 101
variable by creating artificial variable [2]. people where the most disease genes lie.
leA is based on random statistical computation. The data
is assumed to be linear or nonlinear mixer of some unknown
latent variables. The latent variables are non gaussian and
mutually independent and they called independent
component of observed data [3].

II. ALGORITHM AND MATHEMATICAL SIGNIFICANCE

To analysis the internal characteristics of peA


understanding of matrix and statics is required. By evaluates
matrix property and statics shows the features of peA with
mathematical significance. leA algorithm is based on
Gaussian distributions of random variables by analysis the
maximum kurtosis independent component can be estimated.
Figure 1. IOlxlOI matrix data

978-1-4244-6932-1/10/$26.00 ©2010 IEEE

536
A. PCA Algorithm B. ICA Algorithm
,...----...,

Adjust the data to the


center

'" Rem ove m eanfrom


Calcu1ate Covariance data
matrix

W
VVhiten the data
Eigen value (a,b) &
Eigen vector

W Take initial vector.


Make Eigen matrix such
w(O) of norm I
that@TeaterEigen value
vector com e first .

..v
Letw(k)=E{x(w(k-I)T x)} -
Create features vector
3w(k-l)
K=l initiaU
..v
Final data= Rowfeatures
vector" Row data adjust
transpose. Divide w(k) by its
norm
In step2 first vanance IS measured. The variance of
random variable is measure the statistical dispersion of data
mean how far data from its expected value.
Variance (x) L:(x-ui
=

If u E(x), E(x) is expected value of random variable x.


=

Variance lies on one dimension when variance measure more


than one dimension it's called covariance.
Covariance provides a measure of the strength of the
correlation between two or more sets of random variables.

CO�I,X) coy(x," coY(x,s> Vectorw(k)


( gives independent com ponent)
coY(y,x) c:OY(y.Y) c:ov(y.s)
coY(s.,x) co'l(s,� cOY(z.z) In step 2 centering is used to simplify the data. In step 3
the observed vector x is linearly transform to vector xl. Its
component are uncorrelated and their variances equal to
Covariance matrix unity.
T
The covariance matrix signifies symmetric matrix. The E{xlxI } 1=

fmite dimensional spectrum theorem says that for every real The famous process of whiten data is Eigen value
symmetry matrix A there exist a real orthogonal matrix Q decomposition process.
T I12 T
such that D= Q AQ where D is diagonal matrix. xl ED- E x
=

Third step determines Eigen value and Eigen vector of Where E and D are eigen vectors and eigen values of
covariance matrix. Eigen value is used in linear matrix x[5].
transformation of vector. It is property of matrix; act on a Step 4 assumes a norm vector of I. A norm vector
certain vector in which vector change only it magnitude and function strictly assigns the positive length to all vectors in
direction remain same (principle vector). Eigen value matrix vector space. Norm of one vector matrix is diagonal matrix
2 2
is scalar matrix. Its effect on a matrix is scalar multiplication. (dl, d2, d3 ...) such that (dI +d2 +d3+ ...) 112 1
= .

For a module M over a ring, with the endomorphism algebra In step 5 kurtosis is measured. Kurtosis is measurement
End (M) replacing the algebra of matrix, the analog of scalar of non gaussinaty. Independent signal have highest kurtosis
matrix are scalar transformation. Eigen vectors are if we mix one independent signal with other characteristics
orthogonal to each others. .Orthogonal matrix preserve inner tends to more Gaussian distribution means lesser kurtosis
product so for vector u, v in an n dimensional real inner value.
product space[4].

537
III. EXPERIMENT AND RESULT

PCA is one of the great tool which help to recognize


400J
voices we implemented PCA in matlab platform. 7 speakers
voice is recorded in 8 khz sampling frequency for 2 sec. 4
voice commands used move forward, move backward,
turnleft and turnright for each speaker. When correlation
takes from speaker 1 voice to other speakers voices the
probability of match voice is 0.4. Applied PCA for all 7
.2IlXJ
speakers' voice and create PCA voice. From PCA voice we
took correlation to all 7 speakers the probability of voice
·400J
matches is increased around 0.8 that is very helpful to
recognize voices. Result is shown in tablel.

TABLE I. CORRELATION OF SPEAKERS VOICES

Figure 4. ICA voice


Voice command Correlation of Correlation of
moveforward speakerI with PCA voice with
others others
IV. FUTURE DIRECTIONS
SpeakersI I 0.6989
The future work is to implement PCA and ICA in DSP
Speakers 2 0.4971 0.6167
processor; compares their performance for different
Speakers 3 0.4596 0.8187 algorithms.
Speakers 4 0.4699 0.8359
Speakers 5 0.4764 0.7538 V. CONCLUSION
Speakers 6 0.3751 0.7456 PCA fmds the common characteristic of all speakers'
Speakers 7 0.4965 0.7708 voice. If we say move forward what is the property that we
all values are nonnahzed
hear move forward that property specified by PCA. In ICA
ICA evaluates to separate out one voice with mlxmg experiment with both statistical signal and real signal it has
voices [6]. At initial we mixed sine wave and rectangular been observed that the recovered signal always differ from
wave created new mix signal; applied ICA and recovered original signal in term of amplitude and order.
near about original signal. We recorded 10 sec voice at 8 khz
sampling frequency. In voice lot of noises is added like fan ACKNOWLEDGMENT
sound noise so voice is unclear. I thank to Prof. Rabinder Henry, Prof. Amit Patwardhan,
my seniors and my class friends for helping me in
whichever way possible

400J REFERENCE

[I ] Alok Sherrna, Kuldip K Paliwal and Godfrey C.Onwabolu: Splitting


technique initialization in local PCA . Journal of computer science
2(1) 53-58 2006
[2] Luo Juan, Oubong Gwun: A comparison of sift ,PCA sift and surf.
.2OlJ
ww.Cscjournals.orglcsc/manuscript/journals/lJIIP/volume3/... /IJiP-
51.pdf
.4OOJ
[ 3] TE Won Lee, Kluwer: Independent component analysis publication
•6OCO 1999.
[4] Pradipta Mitra: Entywise bounds for eigen vector of random graph.
Published Oct 31 2009.
[5] J.-F. Cardoso: Eigen structure of the fourth order cumulant tensor
Figure 3. voice with voice noises with application of blind source separation problem.
[6] Qiongfeng PAN and tyseer Aboulnasr : Time domain convolution
We applied ICA to separate out unwanted voice signal. Blind source separation employing selective tap adaptive
algorithmsJournal on audio, speech and music processing, volume
Result is shown in fig. 4. 2007.

538

You might also like