TY - GEN
T1 - HARNet
T2 - 34th International Conference on Artificial Neural Networks, ICANN 2025
AU - Li, Jiguang
AU - Şiltu, Meryem Sena
AU - Xu, Meng
AU - Li, Jiawei
AU - Huang, Zhao
AU - Guan, Minglei
PY - 2025/9/11
Y1 - 2025/9/11
N2 - Human Activity Recognition (HAR) is pivotal in various domains, including entertainment, security, and healthcare. Conventional methods often exhibit limitations: hierarchical spatial feature extractors capture local spatial structures but struggle with long-term dependencies, temporal dependency modeling units effectively learn sequential patterns but may lose fine-grained spatial details, and multi-perspective sequential attention modules selectively emphasize critical temporal features yet can overlook subtle local variations. To address these challenges, we propose an advanced framework that synergistically integrates these three components, effectively mitigating spatial constraints and enhancing temporal sensitivity. First, the hierarchical spatial feature extractor autonomously distills multi-level spatial representations from raw skeletal data, ensuring robust spatial encoding. Next, the temporal dependency modeling unit captures long-range temporal correlations, preserving essential motion dynamics across time. Finally, the multi-perspective sequential attention module adaptively assigns significance to different time steps, allowing the model to focus on the most informative elements while suppressing redundant information. Extensive experiments on the AIR-Act2Act dataset demonstrate the superiority of the proposed framework, achieving 99.40% accuracy on dataset 1 and 98.72% on dataset 2, significantly surpassing traditional spatial (92.02%) and temporal models (96.40%) as well as other state-of-the-art approaches (98.0%).
AB - Human Activity Recognition (HAR) is pivotal in various domains, including entertainment, security, and healthcare. Conventional methods often exhibit limitations: hierarchical spatial feature extractors capture local spatial structures but struggle with long-term dependencies, temporal dependency modeling units effectively learn sequential patterns but may lose fine-grained spatial details, and multi-perspective sequential attention modules selectively emphasize critical temporal features yet can overlook subtle local variations. To address these challenges, we propose an advanced framework that synergistically integrates these three components, effectively mitigating spatial constraints and enhancing temporal sensitivity. First, the hierarchical spatial feature extractor autonomously distills multi-level spatial representations from raw skeletal data, ensuring robust spatial encoding. Next, the temporal dependency modeling unit captures long-range temporal correlations, preserving essential motion dynamics across time. Finally, the multi-perspective sequential attention module adaptively assigns significance to different time steps, allowing the model to focus on the most informative elements while suppressing redundant information. Extensive experiments on the AIR-Act2Act dataset demonstrate the superiority of the proposed framework, achieving 99.40% accuracy on dataset 1 and 98.72% on dataset 2, significantly surpassing traditional spatial (92.02%) and temporal models (96.40%) as well as other state-of-the-art approaches (98.0%).
KW - Convolutional Neural Networks
KW - Human Activity Recognition
KW - Long Short-Term Memory
KW - Multi-Head Attention Mechanisms
UR - https://www.scopus.com/pages/publications/105016554043
U2 - 10.1007/978-3-032-04546-1_1
DO - 10.1007/978-3-032-04546-1_1
M3 - Conference contribution
AN - SCOPUS:105016554043
SN - 9783032045454
T3 - Lecture Notes in Computer Science
SP - 1
EP - 13
BT - Artificial Neural Networks and Machine Learning – ICANN 2025 - 34th International Conference on Artificial Neural Networks, 2025, Proceedings
A2 - Senn, Walter
A2 - Sanguineti, Marcello
A2 - Saudargiene, Ausra
A2 - Tetko, Igor V.
A2 - Villa, Alessandro E. P.
A2 - Jirsa, Viktor
A2 - Bengio, Yoshua
PB - Springer
CY - Cham, Switzerland
Y2 - 9 September 2025 through 12 September 2025
ER -