Conference/ProceedingsIEEE Fifth Int. Conf. on Information, Communications and Signal Processing (ICICS '05)
Start date06.12.2005
End date09.12.2005
AddressBangkok, Thailand
Author(s)Juan José Burred, Thomas Sikora
TitleOn the Use of Auditory Representations for Sparsity-Based Sound Source Separation
AbstractSparsity-based source separation algorithms often
rely on a transformation into a sparse domain to improve
mixture disjointness and therefore facilitate separation. To this
end, the most commonly used time-frequency representation has
been the Short Time Fourier Transform (STFT). The purpose of
this paper is to study the use of auditory-based representations
instead of the STFT. We first evaluate the STFT disjointness
properties for the case of speech and music signals, and show
that auditory representations based on the Equal Rectangular
Bandwidth (ERB) and Bark frequency scales can improve the
disjointness of the transformed mixtures.
Key wordssource separation, auditory scales, sparse signals,mixture disjointness