Keypoint Aware Robust Representation for Transformer Based Re Identification of Occluded Person

Keypoint Aware Robust Representation for Transformer Based Re Identification of Occluded Person

Abstract:

Occluded person Re-identification is a challenging task which aims to find or distinguish a specific person when the human body is occluded by obstacles, other persons or oneself. Some recent state-of-the-art works adopting a transformer and/or pose-guided methods have improved the feature representation and performances, but are still in trouble with both weak representation and heavy structure. In this paper, we suggest the novel methods of transformer-based Re-identification for the occluded person as follows. First, in data augmentation, instead of deleting an arbitrary area, only a part containing the keypoint features of a person is deleted for effective learning in occlusion. Second, we suggest a unique hierarchical patch and feature based cross attention combining the reliability-enhanced heatmaps and the output of the transformer intermediate layer, which can more effectively pay attention to the non-occluded human region. Third, we propose an entire-partial loss function with non-similarity grouping for more robust feature representation. We compare our approach and various existing methods for the mAP and Rank-1 performances on the Occluded-Duke, Occluded-ReID, Market-1501 and DukeMTMC datasets. Experimental results demonstrate that our proposed model not only outperforms the existing methods, but also has the smallest scale among state-of-the-art methods.