4.7 Article

Motion Decoupling Network for Intra-Operative Motion Estimation Under Occlusion

Journal

IEEE TRANSACTIONS ON MEDICAL IMAGING
Volume 42, Issue 10, Pages 2924-2935

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TMI.2023.3268774

Keywords

Optical flow; Surgery; Instruments; Estimation; Task analysis; Videos; Motion estimation; Computer-assisted surgery; motion estimation; optical flow; self-supervised learning; surgical images

Ask authors/readers for more resources

This paper proposes a novel unsupervised learning framework to accurately estimate the motion of instruments and soft tissue from intra-operative images. It addresses the challenge of heavy occlusion in the surgical scene and achieves better accuracy compared to other unsupervised methods.
In recent intelligent-robot-assisted surgery studies, an urgent issue is how to detect the motion of instruments and soft tissue accurately from intra-operative images. Although optical flow technology from computer vision is a powerful solution to the motion-tracking problem, it has difficulty obtaining the pixel-wise optical flow ground truth of real surgery videos for supervised learning. Thus, unsupervised learning methods are critical. However, current unsupervised methods face the challenge of heavy occlusion in the surgical scene. This paper proposes a novel unsupervised learning framework to estimate the motion from surgical images under occlusion. The framework consists of a Motion Decoupling Network to estimate the tissue and the instrument motion with different constraints. Notably, the network integrates a segmentation subnet that estimates the segmentation map of instruments in an unsupervised manner to obtain the occlusion region and improve the dual motion estimation. Additionally, a hybrid self-supervised strategy with occlusion completion is introduced to recover realistic vision clues. Extensive experiments on two surgical datasets show that the proposed method achieves accurate motion estimation for intra-operative scenes and outperforms other unsupervised methods, with a margin of 15% in accuracy. The average estimation error for tissue is less than 2.2 pixels on average for both surgical datasets.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available