SAGCN: Self-Attention Graph Convolutional Network for Human Pose Embedding

Accurate human pose embedding is crucial for action recognition. While traditional convolutional neural networks (CNNs) have advanced pose feature extraction, they struggle to model structural relationships and long-range dependencies between keypoints, and are less robust to occlusions. To address...

Full description

Bibliographic Details
Published in:IEEE Access
Main Authors: Zhongxiong Xu, Jiajun Hong, Yicong Yu, Chengzhu Lin, Linfei Yu, Meixian Xu
Format: Article
Language:English
Published: IEEE 2025-01-01
Subjects:
Online Access:https://ieeexplore.ieee.org/document/11119677/
Description
Summary:Accurate human pose embedding is crucial for action recognition. While traditional convolutional neural networks (CNNs) have advanced pose feature extraction, they struggle to model structural relationships and long-range dependencies between keypoints, and are less robust to occlusions. To address these limitations, we present SAGCN, a novel model integrating graph convolutional network (GCN) with self-attention. SAGCN leverages GCN to preserve keypoint structure and self-attention to capture long-range dependencies. We further introduce probabilistic pose embeddings to represent inherent multi-view pose uncertainty. Evaluated on Human 3.6M and MPI-INF-3DHP for cross-view retrieval, and on PenAction for sequence alignment, SAGCN outperforms existing methods in retrieval and achieves competitive alignment results compared to specialized approaches.
ISSN:2169-3536