Fusing Higher-Order Features in Graph Neural Networks for Skeleton-Based Action Recognition

Zhenyue Qin*, Yang Liu*, Pan Ji, Dongwoo Kim, Lei Wang, R. I. McKay, Saeed Anwar, Tom Gedeon

*Corresponding author for this work

    Research output: Contribution to journalArticlepeer-review

    41 Citations (Scopus)

    Abstract

    Skeleton sequences are lightweight and compact and thus are ideal candidates for action recognition on edge devices. Recent skeleton-based action recognition methods extract features from 3-D joint coordinates as spatial-temporal cues, using these representations in a graph neural network for feature fusion to boost recognition performance. The use of first- and second-order features, that is, joint and bone representations, has led to high accuracy. Nonetheless, many models are still confused by actions that have similar motion trajectories. To address these issues, we propose fusing higher-order features in the form of angular encoding (AGE) into modern architectures to robustly capture the relationships between joints and body parts. This simple fusion with popular spatial-temporal graph neural networks achieves new state-of-the-art accuracy in two large benchmarks, including NTU60 and NTU120, while employing fewer parameters and reduced run time. Our source code is publicly available at: https://github.com/ZhenyueQin/Angular-Skeleton-Encoding.

    Original languageEnglish
    Pages (from-to)4783-4797
    Number of pages15
    JournalIEEE Transactions on Neural Networks and Learning Systems
    Volume35
    Issue number4
    DOIs
    Publication statusPublished - 1 Apr 2024

    Fingerprint

    Dive into the research topics of 'Fusing Higher-Order Features in Graph Neural Networks for Skeleton-Based Action Recognition'. Together they form a unique fingerprint.

    Cite this