Part Representation Learning with Teacher-Student Decoder for Occluded Person Re-identification (2312.09797v1)
Abstract: Occluded person re-identification (ReID) is a very challenging task due to the occlusion disturbance and incomplete target information. Leveraging external cues such as human pose or parsing to locate and align part features has been proven to be very effective in occluded person ReID. Meanwhile, recent Transformer structures have a strong ability of long-range modeling. Considering the above facts, we propose a Teacher-Student Decoder (TSD) framework for occluded person ReID, which utilizes the Transformer decoder with the help of human parsing. More specifically, our proposed TSD consists of a Parsing-aware Teacher Decoder (PTD) and a Standard Student Decoder (SSD). PTD employs human parsing cues to restrict Transformer's attention and imparts this information to SSD through feature distillation. Thereby, SSD can learn from PTD to aggregate information of body parts automatically. Moreover, a mask generator is designed to provide discriminative regions for better ReID. In addition, existing occluded person ReID benchmarks utilize occluded samples as queries, which will amplify the role of alleviating occlusion interference and underestimate the impact of the feature absence issue. Contrastively, we propose a new benchmark with non-occluded queries, serving as a complement to the existing benchmark. Extensive experiments demonstrate that our proposed method is superior and the new benchmark is essential. The source codes are available at https://github.com/hh23333/TSD.
- “Pose-guided feature alignment for occluded person re-identification,” in ICCV, 2019, pp. 542–551.
- “An image is worth 16x16 words: Transformers for image recognition at scale,” arXiv:2010.11929, 2010.
- “Deep residual learning for image recognition,” in CVPR, 2016, pp. 770–778.
- “Deep high-resolution representation learning for visual recognition,” IEEE TPAMI, vol. 43, no. 10, pp. 3349–3364, 2020.
- “Pose-guided visible part matching for occluded person reid,” in CVPR, 2020, pp. 11744–11752.
- “Body part-based representation learning for occluded person re-identification,” in WACV, 2023, pp. 1613–1623.
- “Foreground-aware pyramid reconstruction for alignment-free occluded person re-identification,” in ICCV, 2019, pp. 8450–8459.
- “Learning disentangled representation implicitly via transformer for occluded person re-identification,” IEEE TMM, 2022.
- “Pose-guided feature disentangling for occluded person re-identification based on transformer,” in AAAI, 2022, vol. 36, pp. 2540–2549.
- “End-to-end object detection with transformers,” in ECCV. Springer, 2020, pp. 213–229.
- “Unihcp: A unified model for human-centric perceptions,” in CVPR, 2023, pp. 17840–17852.
- “Masked-attention mask transformer for universal image segmentation,” in CVPR, 2022, pp. 1290–1299.
- “Hat: Hierarchical aggregation transformers for person re-identification,” in ACM MM, 2021, pp. 516–525.
- “Semi-attention partition for occluded person re-identification,” in AAAI, 2023, vol. 37, pp. 998–1006.
- “Layer normalization,” arXiv:1607.06450, 2016.
- “Self-correction for human parsing,” IEEE TPAMI, vol. 44, no. 6, pp. 3260–3271, 2020.
- “Bag of tricks and a strong baseline for deep person re-identification,” in CVPRW, 2019, pp. 0–0.
- “In defense of the triplet loss for person re-identification,” arXiv:1703.07737, 2017.
- “Focal loss for dense object detection,” in ICCV, 2017, pp. 2980–2988.
- “Unlabeled samples generated by gan improve the person re-identification baseline in vitro,” in ICCV, 2017, pp. 3754–3762.
- “Random erasing data augmentation,” in AAAI, 2020, vol. 34, pp. 13001–13008.
- “Transreid: Transformer-based object re-identification,” in ICCV, 2021, pp. 15013–15022.
- “Feature erasing and diffusion network for occluded person re-identification,” in CVPR, 2022, pp. 4754–4763.
- “Dynamic prototype mask for occluded person re-identification,” in ACM MM, 2022, pp. 531–540.
- Shang Gao (75 papers)
- Chenyang Yu (14 papers)
- Pingping Zhang (70 papers)
- Huchuan Lu (199 papers)