4.7 Article

Location Sensitive Network for Human Instance Segmentation

Journal

IEEE TRANSACTIONS ON IMAGE PROCESSING
Volume 30, Issue -, Pages 7649-7662

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TIP.2021.3107210

Keywords

Image segmentation; Prototypes; Heating systems; Task analysis; Semantics; Feature extraction; Detectors; Human instance segmentation; spatial invariance; coordinates encoding; points representation

Funding

  1. National Key Research and Development Program of China [2017YFA0700800]
  2. Natural Science Foundation of China (NSFC) [61876171, 61976203]

Ask authors/readers for more resources

Location is crucial in instance segmentation, and LSNet integrates instance-specific location information for distinguishing instances. The proposed model also utilizes the Keypoints Sensitive Combination operation to effectively reduce mis-classified pixels, achieving superior performance compared to its peers, especially in cases of severe occlusion.
Location is an important distinguishing information for instance segmentation. In this paper, we propose a novel model, called Location Sensitive Network (LSNet), for human instance segmentation. LSNet integrates instance-specific location information into one-stage segmentation framework. Specifically, in the segmentation branch, Pose Attention Module (PAM) encodes the location information into the attention regions through coordinates encoding. Based on the location information provided by PAM, the segmentation branch is able to effectively distinguish instances in feature-level. Moreover, we propose a combination operation named Keypoints Sensitive Combination (KSCom) to utilize the location information from multiple sampling points. These sampling points construct the points representation for instances via human keypoints and random points. Human keypoints provide the spatial locations and semantic information of the instances, and random points expand the receptive fields. Based on the points representation for each instance, KSCom effectively reduces the mis-classified pixels. Our method is validated by the experiments on public datasets. LSNet-5 achieves 56.2 mAP at 18.5 FPS on COCOPersons. Besides, the proposed method is significantly superior to its peers in the case of severe occlusion.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available