Semantic combined network for zero-shot scene parsing

Yinduo Wang, Haofeng Zhang*, Shidong Wang, Yang Long, Longzhi Yang

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

1 Citation (Scopus)
6 Downloads (Pure)

Abstract

Recently, image-based scene parsing has attracted increasing attention due to its wide application. However, conventional models can only be valid on images with the same domain of the training set and are typically trained using discrete and meaningless labels. Inspired by the traditional zero-shot learning methods which employ auxiliary side information to bridge the source and target domains, the authors propose a novel framework called semantic combined network (SCN), which aims at learning a scene parsing model only from the images of the seen classes while targeting on the unseen ones. In addition, with the assistance of semantic embeddings of classes, the proposed SCN can further improve the performances of traditional fully supervised scene parsing methods. Extensive experiments are conducted on the data set Cityscapes, and the results show that the proposed SCN can perform well on both zero-shot scene parsing (ZSSP) and generalised ZSSP settings based on several state-of-the-art scenes parsing architectures. Furthermore, the authors test the proposed model under the traditional fully supervised setting and the results show that the proposed SCN can also significantly improve the performances of the original network models.

Original languageEnglish
Pages (from-to)757-765
Number of pages9
JournalIET Image Processing
Volume14
Issue number4
Early online date27 Nov 2019
DOIs
Publication statusPublished - 27 Mar 2020

Fingerprint

Dive into the research topics of 'Semantic combined network for zero-shot scene parsing'. Together they form a unique fingerprint.

Cite this