4.7 Article

Joint Hand-Object 3D Reconstruction From a Single Image With Cross-Branch Feature Fusion

Journal

IEEE TRANSACTIONS ON IMAGE PROCESSING
Volume 30, Issue -, Pages 4008-4021

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TIP.2021.3068645

Keywords

Three-dimensional displays; Shape; Image reconstruction; Task analysis; Solid modeling; Image coding; Pose estimation; Hand pose and shape estimation; 3D object reconstruction; hand-object interaction

Funding

  1. Fundamental Research Funds for the Central Universities [2042020KF0016]
  2. Wuhan University-Infinova project [2019010019]

Ask authors/readers for more resources

Accurately reconstructing 3D hand and object shapes is crucial for understanding human-object interaction. Unlike traditional bare hand pose estimation, the interaction between hand and object imposes constraints on both, suggesting that considering hand configuration as contextual information for the object is important. Current approaches often use separate branches to reconstruct the hand and object, with little communication between them. This study proposes a joint approach in feature space, exploring the reciprocity between hand and object branches. Through cross-branch feature fusion architectures and an auxiliary depth estimation module, the proposed method outperforms existing approaches in terms of reconstruction accuracy.
Accurate 3D reconstruction of the hand and object shape from a hand-object image is important for understanding human-object interaction as well as human daily activities. Different from bare hand pose estimation, hand-object interaction poses a strong constraint on both the hand and its manipulated object, which suggests that hand configuration may be crucial contextual information for the object, and vice versa. However, current approaches address this task by training a two-branch network to reconstruct the hand and object separately with little communication between the two branches. In this work, we propose to consider hand and object jointly in feature space and explore the reciprocity of the two branches. We extensively investigate cross-branch feature fusion architectures with MLP or LSTM units. Among the investigated architectures, a variant with LSTM units that enhances object feature with hand feature shows the best performance gain. Moreover, we employ an auxiliary depth estimation module to augment the input RGB image with the estimated depth map, which further improves the reconstruction accuracy. Experiments conducted on public datasets demonstrate that our approach significantly outperforms existing approaches in terms of the reconstruction accuracy of objects.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available