Informed Single Channel Speech Separation with time-frequency exemplar GMM-HMM model

Qi Wang, W. L. Woo, S. S. Dlay, C. S. Chin, Bin Gao

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

1 Citation (Scopus)

Abstract

In recent studies, the problem of Single Channel Speech Separation (SCSS) have been efficiently tackled by introducing additional cues from the original target source in the form of Informed Source Separation (ISS). In this paper, a more realistic situation is considered where an additional user/listener generated Exemplar source is introduced to aid the separation process instead of using the original target source. The Exemplar source consists of patterns that need to be transformed, extracted, regulated and calibrated to generate an utterance dependent (UD) model that could statistically represent the target source. The proposed method uses general speaker independent (SI) features along with the generated UD features are modelled and combined in a joint probability model to achieve reliable separation. Unlike most model-based approaches, the proposed method does not require Speaker Dependent training on individual sources of the mixture, and is therefore much more efficient and less restrictive.

Original languageEnglish
Title of host publication2015 IEEE International Conference on Digital Signal Processing, DSP 2015
PublisherIEEE
Pages1130-1134
Number of pages5
ISBN (Electronic)9781479980581
DOIs
Publication statusPublished - 10 Sep 2015
EventIEEE International Conference on Digital Signal Processing, DSP 2015 - Singapore, Singapore
Duration: 21 Jul 201524 Jul 2015

Conference

ConferenceIEEE International Conference on Digital Signal Processing, DSP 2015
CountrySingapore
CitySingapore
Period21/07/1524/07/15

Fingerprint

Dive into the research topics of 'Informed Single Channel Speech Separation with time-frequency exemplar GMM-HMM model'. Together they form a unique fingerprint.

Cite this