TY - JOUR
T1 - GAN-based reactive motion synthesis with class-aware discriminators for human-human interaction
AU - Men, Qianhui
AU - Shum, Hubert P.H.
AU - Ho, Edmond S.L.
AU - Leung, Howard
N1 - Funding information: This project was funded in part by the Royal Society (Ref: IESR2181024).
PY - 2022/2/1
Y1 - 2022/2/1
N2 - Creating realistic characters that can react to the users’ or another character's movement can benefit computer graphics, games and virtual reality hugely. However, synthesizing such reactive motions in human–human interactions is a challenging task due to the many different ways two humans can interact. While there are a number of successful researches in adapting the generative adversarial network (GAN) in synthesizing single human actions, there are very few on modeling human–human interactions. In this paper, we propose a semi-supervised GAN system that synthesizes the reactive motion of a character given the active motion from another character. Our key insights are two-fold. First, to effectively encode the complicated spatial–temporal information of a human motion, we empower the generator with a part-based long short-term memory (LSTM) module, such that the temporal movement of different limbs can be effectively modeled. We further include an attention module such that the temporal significance of the interaction can be learned, which enhances the temporal alignment of the active–reactive motion pair. Second, as the reactive motion of different types of interactions can be significantly different, we introduce a discriminator that not only tells if the generated movement is realistic or not, but also tells the class label of the interaction. This allows the use of such labels in supervising the training of the generator. We experiment with the SBU, the HHOI and the 2C datasets. The high quality of the synthetic motion demonstrates the effective design of our generator, and the discriminability of the synthesis also demonstrates the strength of our discriminator.
AB - Creating realistic characters that can react to the users’ or another character's movement can benefit computer graphics, games and virtual reality hugely. However, synthesizing such reactive motions in human–human interactions is a challenging task due to the many different ways two humans can interact. While there are a number of successful researches in adapting the generative adversarial network (GAN) in synthesizing single human actions, there are very few on modeling human–human interactions. In this paper, we propose a semi-supervised GAN system that synthesizes the reactive motion of a character given the active motion from another character. Our key insights are two-fold. First, to effectively encode the complicated spatial–temporal information of a human motion, we empower the generator with a part-based long short-term memory (LSTM) module, such that the temporal movement of different limbs can be effectively modeled. We further include an attention module such that the temporal significance of the interaction can be learned, which enhances the temporal alignment of the active–reactive motion pair. Second, as the reactive motion of different types of interactions can be significantly different, we introduce a discriminator that not only tells if the generated movement is realistic or not, but also tells the class label of the interaction. This allows the use of such labels in supervising the training of the generator. We experiment with the SBU, the HHOI and the 2C datasets. The high quality of the synthetic motion demonstrates the effective design of our generator, and the discriminability of the synthesis also demonstrates the strength of our discriminator.
KW - Generative adversarial network
KW - Attention
KW - Reactive motion synthesis
UR - http://www.scopus.com/inward/record.url?scp=85117834676&partnerID=8YFLogxK
U2 - 10.1016/j.cag.2021.09.014
DO - 10.1016/j.cag.2021.09.014
M3 - Article
SN - 0097-8493
VL - 102
SP - 634
EP - 645
JO - Computers & Graphics
JF - Computers & Graphics
ER -