Spectrogram based multi-task audio classification

Yuni Zeng, Hua Mao, Dezhong Peng, Zhang Yi

Research output: Contribution to journalArticlepeer-review

127 Citations (Scopus)
329 Downloads (Pure)

Abstract

Audio classification is regarded as a great challenge in pattern recognition. Although audio classification tasks are always treated as independent tasks, tasks are essentially related to each other such as speakers’ accent and speakers’ identification. In this paper, we propose a Deep Neural Network (DNN)-based multi-task model that exploits such relationships and deals with multiple audio classification tasks simultaneously. We term our model as the gated Residual Networks (GResNets) model since it integrates Deep Residual Networks (ResNets) with a gate mechanism, which extract better representations between tasks compared with Convolutional Neural Networks (CNNs). Specifically, two multiplied convolutional layers are used to replace two feed-forward convolution layers in the ResNets. We tested our model on multiple audio classification tasks and found that our multi-task model achieves higher accuracy than task-specific models which train the models separately.
Original languageEnglish
Pages (from-to)3705-3722
Number of pages18
JournalMultimedia Tools and Applications
Volume78
Issue number3
Early online date26 Dec 2017
DOIs
Publication statusPublished - Feb 2019

Keywords

  • Multi-task learning
  • Convolutional neural networks
  • Deep residual networks
  • Audio classification

Fingerprint

Dive into the research topics of 'Spectrogram based multi-task audio classification'. Together they form a unique fingerprint.

Cite this