Top Page
Masashi UNOKI and Masato AKAGI,
"Vowel segregation in background noise using the model of segregating two acoustic sources,"
Proc. of AI Challenge, pp. 7-14, AOYAMA Gakuin Univ., 3Nov. 1999.
Last modified:
2 June 2001
Abstract

This paper proposes an improved sound segregation model based on auditory scene analysis in order to overcome three disadvantages in our previously proposed model. The improved model solves the problem of segregating two acoustic sources by using constraints related to the heuristic regularities proposed by Bregman. In the improvements, we (1) reconsider the estimation of unknown parameters using Kalman filtering, (2) incorporate a constraint of channel envelopes with periodicity of the fundam ental frequency into the grouping block, and (3) consider a constraint of smoothness of instantaneous amplitudes on channels. Simulations are performed to segregate a real vowel from a noisy vowel and to compare the results of using all or only some constraints. The proposed model can improve our previous model and precisely segregate real speech even in waveforms using all of the constraints related to Bregman's four regularities.

Latex2html

AIC99.html

Download
  • Electrical Paper
    [gziped ps-file] (240 kB)
    [pdf-file] (2.29 MB)
  • Created by M. Unoki, 6 Nov. 2000