• Ingen resultater fundet

Comparison of Fundamental Frequency Estimators

In document IMM, Denmarks Technical University (Sider 39-43)

Choosing and Extracting Feature Sets

3.4 Fundamental Frequency Estimation

3.4.4 Comparison of Fundamental Frequency Estimators

Using each of the three fundamental frequency estimators that are discussed in Sec-tions 3.4.1-3.4.3, an average F0 for each speaker in the reference set is obtained. The estimation of the fundamental frequencies of all six reference speakers is implemented by rst estimating a value for each sentence - all 9 sentences from each speaker are used, including both training and test data. A median value calculated over the estimate for every frame in each sentence is used for the real cepstrum and autocorrelation methods, while the output of the YIN estimator yields a "best" estimate of F0 for the entire sen-tence. This estimate is determined at the dip in the cumulative mean nomalized dierence function discussed in Section 3.4.2 that is found at the minimum lag value. As the other twoF0 estimators return an estimate forF0 for each frame, the median must be calculated to provide one estimate for the entire sentence. For each speaker, the average F0 is found as the mean of the estimates over all 9 sentences. The results for all three estimators are shown in Figure 3.7.

The YIN estimator was implemented with default parameters, as numerous trials with varying threshold values and frame lengths yielded no signicant change in the results.

1 2 3 4 5 6

Figure 3.7: Fundamental frequency estimation for Autocorrelation CC, YIN and Real Cepstrum methods

The lower frequency bound is set at F0,min = 30Hz and the window length set to the sampling frequency divided by this value, see Eq.(3.10), as this is assumed to be enough to determine the signal periodicity. For the speakers in the ELSDSR database, this gives a window length of W = 33ms.

W = Fs

F0,min (3.10)

The optimal frame lengths for the other F0 estimators were determined by trial and error: 30ms for the autocorrelation with center clipping method, and 64ms for the real cepstrum method.

Figure 3.7 shows that the YIN estimator has a tendency to produce higher estimates of the fundamental frequency than the other two estimators. The results from all three estimators, however, show that while the dierences between gender groups are large -this can be seen as the rst 3 speakers are women, the last 3 men - the variation within each gender group is very small, especially for the women, and it is thus unlikely that this feature is well suited for the general speaker identication task. According to the documentation in [48], the deviance between the fundamental frequency estimates are larger between the YIN estimates and the other two sets of data because YIN is more precise.

Results based on all feature sets and an analysis to determine whether the voiced/unvoiced decisions inuence system performance will be discussed in Chapter 9. The time required by each method to return a fundamental frequency estimate is considered here. Averaged over all 7 training sentences and both test sentences for each speaker, these times are shown in Figure 3.8. The training and testing data sets are kept separate because of the dierence in length of the sentences contained in each set. The results are averaged over

all 6 reference speakers.

The Average Computation Time for each Fundamental Frequency Estimator Autocorrelation YIN Real Cepstrum

Figure 3.8: The average computation time for each fundamental frequency estimator From Figure 3.8F0 estimation is seen to be most rapid using the real cepstrum method, while the YIN estimator requires a signicant increase in computational time when com-pared to the other two methods. The choice of estimator, however, will be left until further trials in Chapter 9 have been completed.

The fundamental frequency that has been determined so far has been a single value, aver-aged over the sequence of frames that combined constitute sentences from each speaker.

The way that the fundamental frequency changes as a function of time when a speaker is talking is not represented in this analysis, though this may prove to be interesting as a possible feature for speaker identication. In Figure 3.9, the trajectories of fundamen-tal frequency estimates for entire sequences of frames are shown. The two top speech sequences are of women's voices and the two bottom plots are of male voices. The sen-tence used was arbitrarily chosen, though identical for all speakers to ensure that the trajectories depicted are comparable. Sentence d is used. The original scaling has not been modied in order to achieve a uniformity that would facilitate the comparison of these plots, as the dierences are in some places so signicant that this was not feasible.

It is precisely these dierences, though, that lead to the observation that the range of each speaker's fundamental frequency varies considerably, f.ex. the F0 values for Speaker 1 have a range of roughly 300Hz, while for Speaker 6 they vary within a range of only approximately 130Hz. The number of frames is not equal for all speakers and this shows that the speed with which each speaker utters sentence d is speaker dependent. Despite these dierences, it is easily seen that a large amount of each speaker's fundamental fre-quency estimates lie within the intervals that are dened by the fundamental frequencies of the other speakers. This leads to the assumption that there is little possibility that the trajectory of the fundamental frequency will prove ecient in discriminating between speakers.

In Figure 3.10, more evidence is found that supports the assumption that the sequence of F0 estimates may not be an eective feature vector in speaker identication. The top plot shows the F0 estimates for two training sentences from Speaker 1 and the bottom plot shows a corresponding analysis for two dierent speakers, but for the same sentence.

50 100 150 200 250

Figure 3.9: Fundamental frequency trajectories for dierent speakers The number of frames for each sentence is as follows listed below:

- Speaker 1, sentence a: 169

The variance of F0 estimates for Speaker 1, sentence a and b

F0/Hz

The variance of F0 estimates for Speaker 1 and 2, sentence a

F0/Hz

Frame index

speaker 1 speaker 2

Figure 3.10: Pitch trajectory data, for dierent speakers and sentences

Although Figure 3.10 reveals slightly more overlap between the two sets of points in the top plot, the dierence is not signicant and it is dicult to see how a classier would dierentiate between the speakers if the pitch trajectories were used as features for SID. This feature will be tested, however, as there may be enough variance between some speakers to allow a degree of separation that is greater than seen here.

The feature sets that have been derived in Section 3.4 are representative of the source information in a speech signal and will be tested with dierent classiers in Chapter 9.

The next few sections are dedicated to describing the derivation of other feature sets.

In document IMM, Denmarks Technical University (Sider 39-43)