Abstract
This work begins by examining the correlation between audio and visual speech features and reveals higher correlation to exist within individual phoneme sounds rather than globally across all speech. Utilising this correlation, a visually-derived Wiener filter is proposed in which clean power spectrum estimates are obtained from visual speech features. Two methods of extracting clean power spectrum estimates are made; first from a global estimate using a single Gaussian mixture model (GMM), and second from phoneme-specific estimates using a hidden Markov model (HMM)-GMM structure. Measurement of estimation accuracy reveals that the phoneme-specific (HMM-GMM) system leads to lower estimation errors than the global (GMM) system. Finally, the effectiveness of visually-derived Wiener filtering is examined
Original language | English |
---|---|
Pages | IV-585-IV-588 |
Number of pages | 4 |
DOIs | |
Publication status | Published - 2007 |
Event | IEEE International Conference on Acoustics, Speech and Signal Processing - Honolulu, United States Duration: 15 Apr 2007 → 20 Apr 2007 |
Conference
Conference | IEEE International Conference on Acoustics, Speech and Signal Processing |
---|---|
Country/Territory | United States |
City | Honolulu |
Period | 15/04/07 → 20/04/07 |