3.8 Proceedings Paper

TokenPose: Learning Keypoint Tokens for Human Pose Estimation

出版社

IEEE
DOI: 10.1109/ICCV48922.2021.01112

关键词

-

资金

  1. National Key R&D Plan of the Ministry of Science and Technology [2020AAA0104400]
  2. National Key Research and Development Program of China [2018YFB1800204]
  3. National Natural Science Foundation of China [61773117, 61771273]
  4. R&D Program of Shenzhen [JCYJ20180508152204044]

向作者/读者索取更多资源

This paper introduces a novel approach for human pose estimation based on Token representation, which can learn constraint relationships and appearance cues simultaneously, achieving comparable performance with existing methods in experiments.
Human pose estimation deeply relies on visual clues and anatomical constraints between parts to locate keypoints. Most existing CNN-based methods do well in visual representation, however, lacking in the ability to explicitly learn the constraint relationships between keypoints. In this paper, we propose a novel approach based on Token representation for human Pose estimation (TokenPose). In detail, each keypoint is explicitly embedded as a token to simultaneously learn constraint relationships and appearance cues from images. Extensive experiments show that the small and large TokenPose models are on par with state-of-the-art CNN-based counterparts while being more lightweight. Specifically, our TokenPose-S and TokenPose-L achieve 72.5 AP and 75.8 AP on COCO validation dataset respectively, with significant reduction in parameters (down arrow 80.6%; down arrow 56.8%) and GFLOPs (down arrow 75.3%; down arrow 24.7%). Code is publicly available(1).

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

3.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据