Paper
3 January 2025 STAT-Net: spatiotemporal aggregation transformer network for skeleton-based few-shot action recognition
Dazhi Ren, Shengli Lv, Jinlin Li, Naining Li, Lin Dang
Author Affiliations +
Proceedings Volume 13519, Third International Conference on Communications, Information System, and Data Science (CISDS 2024); 135190F (2025) https://doi.org/10.1117/12.3058329
Event: Third International Conference on Communications, Information System and Data Science 2024, 2024, Nanjing, China
Abstract
Few-shot action recognition predicts new classes without labels and has received widespread attention for practical systems. The skeleton is a sparse representation of human actions, and existing spatiotemporal based models by training a strong encoder network could make the skeleton graph very dense with edges, which may lead to the over-smoothing problem. To address this issue, we propose the Spatio-Temporal Aggregation Transformer Network (STAT-Net) as a general backbone for skeleton-based few-shot action recognition. In the spatiotemporal aggregation transformer modules, the spatial multi-head self attention for modeling the connection of different joints in the same frame, while the temporal multi-head self attention for modeling the skeleton sequence between two adjacent frames. The extracted features between the three parts are aggregated by Adaptive Fusion technique to obtain a high dimensional embedding. Extensive experiments on two benchmarks demonstrate that our proposed model achieves better recognition results com- pared with other existing methods.
© (2025) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Dazhi Ren, Shengli Lv, Jinlin Li, Naining Li, and Lin Dang "STAT-Net: spatiotemporal aggregation transformer network for skeleton-based few-shot action recognition", Proc. SPIE 13519, Third International Conference on Communications, Information System, and Data Science (CISDS 2024), 135190F (3 January 2025); https://doi.org/10.1117/12.3058329
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
Back to Top