• Tidak ada hasil yang ditemukan

Pattern matching and score normalization

Dalam dokumen and my wife, (Halaman 49-52)

Cholesky decomposition of the inverse of the within-class covariance matrixW as,

W−1=BBT (2.19)

2.3.4 Nuisance attribute projection

In nuisance attribute projection (NAP) method objective is to find the nuisance subspace which represents the channel/session variabilities [56]. The data is projected to the complimentary space of the nuisance subspace for compensation. The projection matrixP is given as,

P =I−RRT (2.20)

whereRis a rectangular matrix containing the eigen vectors corresponding to few of the top eigen values of the within-class covariance matrix.

2.4 Pattern matching and score normalization

The pattern matching module choses the speaker model according to the identity claim input and compute the verification score of the test data on the selected model. This verification score is the measure of confidence of the system in the hypothesis that the input speech utterance is generated by the claimed speaker. The method of computing the verification score depends on the kind of the speaker modeling technique used. The score computation methods that are commonly used with various speaker modeling approaches, discussed in Section 2.2, are described in the following.

2.4.1 Likelihood ratio scoring with GMM modeling

In SV systems using parametric speaker modeling approaches like GMM or GMM-UBM, the likelihood of the test data vectors on the claimed speakers model is used as the score for verification.

Usually, the feature vectors of a test utterance X are assumed to be independent and identically distributed (IID), so the likelihood of a claimed speaker modelθc for a sequence of feature vectors, X ≡ {x1,x2, . . . ,xN}, is computed as,

p(X|θc) =

N

Y

n=1

p(xnc) (2.21)

The likelihoods over the claimed model is often normalized with the likelihood on the UBM model to obtain the log-likelihood ratio (LLR) score as,

LLR Score = log p(X|θc) p(X|θU BM)

= log QN

n=1p(xnc) QN

n=1p(xnU BM)

=

N

X

n=1

logp(xnc)−

N

X

n=1

logp(xnU BM) (2.22)

2.4.2 Cosine distance scoring

With GMM mean supervectors or i-vectors as the speaker utterance representations, SVM can be used for pattern matching [17, 56]. The representations for both the training and test speech utterance are derived and the SVM classifier is used to compute the verification score. Later it was shown in [59] that the use of a simple cosine distance measure between training and test data representations provides better performance compared to the SVM classifier for SV tasks. This is due to the fact that, in most of the publicly available SV databases like NIST SRE datasets, only a few utterances are is available to train each speaker. As a result of this, the training of SVM becomes suboptimal and it fails in achieving its effectiveness in such conditions. The cosine distance score is computed as,

Score = <wˆclm.wˆtst>

kwˆclmk2 kwˆtstk2

(2.23) where ˆwclmand ˆwtstrepresent the representations of the training and the test utterances, respec- tively, and the operators have the usual meaning.

2.4.3 Bayesian classifier with probabilistic linear discriminant analysis

Probabilistic linear discriminant analysis (PLDA) [60] is a stochastic approach which is recently been explored for performing SV task in the i-vector domain [18, 19]. In PLDA, an i-vector w corresponding to a speaker utterances is represented using a generative model as,

w=ρ+Hh+Gg+ε (2.24)

where, ρ is the global mean of i-vector population, H and G are the matrices representing the speaker and channel subspaces respectively, hand g are the speaker and channel factors, respec- tively with each having standard normal prior distribution, and ǫ is the residual factor having

2.4 Pattern matching and score normalization

standard normal prior with diagonal covariance. In [18], heavy-tailed priors are assumed for the latent variables for handling the effect of outliers in the data and the model is known as heavy-tailed PLDA (HPLDA). Later in [19] a simplified version of PLDA has been proposed which ignores the termGg and assigns standard normal prior to the latent variables. The residual term is modeled with Guassian distribution with zero mean and non-diagonal covariance denoted by S. A Gaus- sianization process consisting of whitening followed by length normalization is performed on the i-vectors prior to modeling and testing. This approach is commonly referred to as Gaussian PLDA (GPLDA) and reported to provide performance similar to that of the HPLDA with a much lesser complexity.

The parameters of the GPLDA model{ρ,H,S} are learned with the help of EM algorithm on a development dataset. Given the length normalized i-vectorswclmand wtst corresponding to the claimed speakers training utterance and the test utterance, respectively, the log-likelihood ratio score using the GPLDA model is computed by a hypothesis test as,

LLR Score = log p(wclm,wtsts) p(wclmd)p(wtstd)

=

R p(wclm,wtst,hclm|ρ,S)dhclm Rp(wclmhclm|ρ,S)dhclmR

p(wtst,htst|ρ,S)dhtst

=

Rp(wclm,wtst)|ρ,S)p(hclm)dhclm Rp(wclm|ρ, S)p(hclm)dhclmR

p(wtst|ρ,S)p(htst)dhtst

= N

wTclmwTtstT

|

ρTρTT

,H˜H˜T + ˜S N

wTclmwTtstT

|[ρTρT]T , diag{HHT +S,HHT +S} (2.25) where Γsis the hypothesis that bothwclmandwtstshare the same speaker identity latent variable hwhile Γd is the hypothesis thatwclm and wtst are generated by different speaker identity latent variablesh1andh2. ˜H =

HTHTT

and ˜S =diag{S,S}, where the operatordiag{., .}creates a block diagonal matrix by placing the arguments across the diagonal. The final closed-form solution for the LLR score is derived as,

LLR Score =C+wTclmQwclm+wTtstQwtst+ 2wclmT P wtst (2.26)

where

C= 1

2lnkD0k −1

2lnkD1k+ρT(P +Q)ρ P = ∆−1Φ(∆−Φ∆−1Φ)−1

Q= ∆−1−(∆−Φ∆−1Φ)−1

∆ =HHT +S Φ =HHT

D0=

∆ Φ Φ ∆

, D1 =

∆ 0

0 ∆

Dalam dokumen and my wife, (Halaman 49-52)