next up previous
Next: Auditory sound segregation model Up: Vowel segregation in background Previous: Vowel segregation in background

Introduction

The problem of segregating the desired signal from a noisy signal is an important issue not only in robust speech recognition systems but also in various types of signal processing. This problem has been investigated by many researchers, and many methods have been proposed. For example, an investigation of robust speech recognition [Furui and Sondhi1991], includes noise reduction or suppression [Boll1979] and speech enhancement methods [Junqua and Haton1996]. An investigation of signal processing includes signal estimation using a linear system [Papoulis1977] and signal estimation based on a stochastic process for signals and noise [Papoulis1991]. One recent proposal is Blind Separation [Shamsunder and Giannakis1997], which estimates the inverse-translation-operator (input-output translation function) by using the observed signal to estimate the original input.

However, in practice, it is difficult to segregate each original signal from a mixed signal, because this problem is an ill-posed inverse problem and the signals exist in a concurrent time-frequency region. Furthermore this problem is difficult to solve without using constraints on acoustic sources and the real environment.

On the other hand, the human auditory system can easily segregate the desired signal in a noisy environment that simultaneously contains speech, noise, and reflections. Recently, this ability of the auditory system has been regarded as a function of an active scene analysis system called "Auditory Scene Analysis (ASA)". ASA has become widely known as a result of Bregman's book [Bregman1990]. Bregman has reported that the human auditory system uses four psychoacoustically heuristic regularities related to acoustic events to solve the problem of Auditory Scene Analysis. These regularities are (i) common onset and offset, (ii) gradualness of change, (iii) harmonicity, and (iv) changes occurring in the acoustic event [Bregman1993]. If an auditory sound segregation model were constructed using constraints related to these heuristic regularities, it should be possible to uniquely solve the sound segregation problem (ill-posed inverse problem). In addition, this model should be applicable not only to a preprocessor for robust speech recognition systems but also to various types of signal processing.

Some ASA-based investigations have shown that it is possible to solve the segregation problem by applying constraints to sounds and the environment. These approaches are called "Computational Auditory Scene Analysis (CASA)". Some CASA-based sound segregation models already exist. There are two main types of models, based on either bottom-up or top-down processes. Typical bottom-up models include an auditory sound segregation model based on acoustic events [Cooke1993,Brown1992], a concurrent harmonic sounds segregation model based on the fundamental frequency [de Cheveigne1993], and a sound source separation system with an automatic tone modeling ability [Kashino and Tanaka1993]. Typical top-down models include a segregation model based on psychoacoustic grouping rules [Ellis1996] and a computational model of sound segregation agents [Nakatani et al.1995a,Nakatani et al.1995b]. All of these models use some of the four regularities and the amplitude (or power) spectrum as the acoustic feature. Thus, they cannot completely extract the desired signal from a noisy signal when the signal and noise exist in the same frequency region.

In contrast, we have been tackling the problem of segregating two acoustic sources as a fundamental problem. We believe that this problem can be uniquely solved by using amplitude, phase information, and mathematical constraints related to the four psychoacoustically heuristic regularities [Unoki and Akagi1997,Unoki and Akagi1999a].

This fundamental problem is defined as follows [Unoki and Akagi1997,Unoki and Akagi1999a]. First, only the mixed signal f(t), where f(t)=f1(t)+f2(t), can be observed. Next, f(t) is decomposed into its frequency components by a filterbank (the number of channels is K). The output of the k-th channel Xk(t) is represented by

 \begin{displaymath}X_k(t)=S_k(t)\exp(j\omega_k t + j\phi_k(t)).
\end{displaymath} (1)

Here, if the outputs of the k-th channel X1,k(t) and X2,k(t), which correspond to f1(t) and f2(t), are assumed to be
X1,k(t) = $\displaystyle A_k(t)\exp(j\omega_k t + j\theta_{1k}(t)),$ (2)
X2,k(t) = $\displaystyle B_k(t)\exp(j\omega_k t + j\theta_{2k}(t)),$ (3)

then the instantaneous amplitudes of the two signals Ak(t) and Bk(t) can be determined by
  
Ak(t) = $\displaystyle \frac{S_k(t)\sin(\theta_{2k}(t)-\phi_k(t))}{\sin\theta_k(t)},$ (4)
Bk(t) = $\displaystyle \frac{S_k(t)\sin(\phi_k(t)-\theta_{1k}(t))}{\sin\theta_k(t)},$ (5)

where $\theta_k(t)=\theta_{2k}(t)-\theta_{1k}(t)$, $\theta_k(t)\not= n\pi, n\in{\bf {Z}}$, and $\omega_k$ is the center frequency of the k-th channel. Instantaneous phases $\theta_{1k}(t)$ and $\theta_{2k}(t)$ can be determined by
  
$\displaystyle \theta_{1k}(t)$ = $\displaystyle -\arctan\left( \frac{Y_k(t)\cos\phi_k(t)-\sin\phi_k(t)}{Y_k(t)\sin\phi_k(t)+\cos\phi_k(t)} \right)$  
    $\displaystyle +\arcsin\left(\frac{A_k(t)Y_k(t)}{S_k(t)\sqrt{Y_k(t)^2+1}}\right),$ (6)
$\displaystyle \theta_{2k}(t)$ = $\displaystyle -\arctan\left( \frac{Y_k(t)\cos\phi_k(t)+\sin\phi_k(t)}{Y_k(t)\sin\phi_k(t)-\cos\phi_k(t)} \right)$  
    $\displaystyle +\arcsin\left(-\frac{B_k(t)Y_k(t)}{S_k(t)\sqrt{Y_k(t)^2+1}}\right),$ (7)

where
Yk(t) = $\displaystyle {\sqrt{(2A_k(t)B_k(t))^2-Z_k(t)^2}}/{Z_k(t)},$ (8)
Zk(t) = Sk(t)2-Ak(t)2-Bk(t)2. (9)

Hence, f1(t) and f2(t) can be reconstructed by using the determined pair of [Ak(t) and $\theta_{1k}(t)]$ and the determined pair of [Bk(t) and $\theta_{2k}(t)]$ for all channels. However, Ak(t), Bk(t), $\theta_{1k}(t)$, and $\theta_{2k}(t)$ cannot be uniquely determined without some constraints, as is easily understood from the above equations. Therefore, this problem is an ill-inverse problem.

To overcome this problem, we have tried to construct a basic solution using constraints related to the four regularities [Unoki and Akagi1997,Unoki and Akagi1999a]. Thus, we have proposed a sound segregation model based on auditory scene analysis [Unoki and Akagi1999b]. This model solves the problem of segregating two acoustic sources by using constraints on the continuity of instantaneous phases as well as constraints on the continuity of instantaneous amplitudes and fundamental frequencies. In simulations, we showed that all constraints related to the four regularities are useful in segregating an AM-FM harmonic complex tone from a noisy AM-FM harmonic complex tone. We also showed that the proposed model can precisely segregate a real vowel from a noisy vowel even in waveforms.

However, this model has the following disadvantages:

(1)
the segregation accuracy improvement differs depending on the two types of noise (white or pink noise),
(2)
it often fails to extract the components of envelopes with periodicity at the fundamental frequency by using the constraint of common onset / offset, and
(3)
the segregated vowel feels as if it does not have good hearing quality.

This paper proposes an improved sound segregation model based on auditory scene analysis to overcome the above disadvantages.


next up previous
Next: Auditory sound segregation model Up: Vowel segregation in background Previous: Vowel segregation in background
Masashi Unoki
2000-11-07