This paper proposes an improved sound segregation model based on
auditory scene analysis in order to overcome three disadvantages in our
previously proposed model. The improved model solves the problem of
segregating two acoustic sources by using constraints related to the
heuristic regularities proposed by Bregman.
In the improvements, we (1) reconsider the estimation of unknown
parameters using Kalman filtering, (2) incorporate a constraint of
channel envelopes with periodicity of the fundam ental frequency into
the grouping block, and (3) consider a constraint of smoothness of
instantaneous amplitudes on channels.
Simulations are performed to segregate a real vowel from a noisy vowel
and to compare the results of using all or only some constraints.
The proposed model can improve our previous model and precisely
segregate real speech even in waveforms using all of the constraints
related to Bregman's four regularities.
|