Progressive Feature Learning for Realistic Cloth-Changing Gait Recognition (2207.11720v2)
Abstract: Gait recognition is instrumental in crime prevention and social security, for it can be conducted at a long distance to figure out the identity of persons. However, existing datasets and methods cannot satisfactorily deal with the most challenging cloth-changing problem in practice. Specifically, the practical gait models are usually trained on automatically labeled data, in which the sequences' views and cloth conditions of each person have some restrictions. To be concrete, the cross-view sub-dataset only has normal walking condition without cloth-changing, while the cross-cloth sub-dataset has cloth-changing sequences but only in front views. As a result, the cloth-changing accuracy cannot meet practical requirements. In this work, we formulate the problem as Realistic Cloth-Changing Gait Recognition (abbreviated as RCC-GR) and we construct two benchmarks: CASIA-BN-RCC and OUMVLP-RCC, to simulate the above setting. Furthermore, we propose a new framework called Progressive Feature Learning that can be applied with off-the-shelf backbones to improve their performance in RCC-GR. Specifically, in our framework, we design Progressive Mapping and Progressive Uncertainty to extract cross-view features and then extract cross-cloth features on the basis. In this way, the feature from the cross-view sub-dataset can first dominate the feature space and relieve the uneven distribution caused by the adverse effect from the cross-cloth sub-dataset. The experiments on our benchmarks show that our framework can effectively improve recognition performance, especially in the cloth-changing conditions.
- P. Xu and X. Zhu, “Deepchange: A large long-term person re-identification benchmark with clothes change,” arXiv preprint arXiv:2105.14685, 2021.
- P. K. Larsen, E. B. Simonsen, and N. Lynnerup, “Gait analysis in forensic medicine,” Journal of forensic sciences, vol. 53, no. 5, pp. 1149–1153, 2008.
- I. Bouchrika, M. Goffredo, J. Carter, and M. Nixon, “On using gait in forensic biometrics,” Journal of forensic sciences, vol. 56, no. 4, pp. 882–889, 2011.
- S. Black, M. Wall, R. Abboud, R. Baker, and J. Stebbins, “Forensic gait analysis: A primer for courts,” 2017.
- S. Yu, D. Tan, and T. Tan, “A framework for evaluating the effect of view angle, clothing and carrying condition on gait recognition,” in 18th international conference on pattern recognition (ICPR’06), vol. 4. IEEE, 2006, pp. 441–444.
- R. Liao, C. Cao, E. B. Garcia, S. Yu, and Y. Huang, “Pose-based temporal-spatial network (ptsn) for gait recognition with carrying and clothing variations,” in Chinese Conference on biometric recognition. Springer, 2017, pp. 474–483.
- N. Takemura, Y. Makihara, D. Muramatsu, T. Echigo, and Y. Yagi, “Multi-view large population gait dataset and its performance evaluation for cross-view gait recognition,” IPSJ Transactions on Computer Vision and Applications, vol. 10, no. 1, pp. 1–14, 2018.
- P. Connor and A. Ross, “Biometric recognition by gait: A survey of modalities and features,” Computer Vision and Image Understanding, vol. 167, pp. 1–27, 2018.
- A. Zhao, J. Dong, J. Li, L. Qi, and H. Zhou, “Associated spatio-temporal capsule network for gait recognition,” IEEE Transactions on Multimedia, vol. 24, pp. 846–860, 2022.
- Z. Zhu, X. Guo, T. Yang, J. Huang, J. Deng, G. Huang, D. Du, J. Lu, and J. Zhou, “Gait recognition in the wild: A benchmark,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 14 789–14 799.
- Y.-C. Chen, X. Zhu, W.-S. Zheng, and J.-H. Lai, “Person re-identification by camera correlation aware feature augmentation,” IEEE transactions on pattern analysis and machine intelligence, vol. 40, no. 2, pp. 392–408, 2017.
- L. Zheng, Y. Yang, and A. G. Hauptmann, “Person re-identification: Past, present and future,” arXiv preprint arXiv:1610.02984, 2016.
- J. Deng, J. Guo, N. Xue, and S. Zafeiriou, “Arcface: Additive angular margin loss for deep face recognition,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 4690–4699.
- J. Deng, J. Guo, E. Ververas, I. Kotsia, and S. Zafeiriou, “Retinaface: Single-shot multi-level face localisation in the wild,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 5203–5212.
- J. Liu, Y. Sun, C. Han, Z. Dou, and W. Li, “Deep representation learning on long-tailed data: A learnable embedding augmentation perspective,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 2970–2979.
- R. Liao, S. Yu, W. An, and Y. Huang, “A model-based gait recognition method with body pose and human prior knowledge,” Pattern Recognition, vol. 98, p. 107069, 2020.
- X. Li, Y. Makihara, C. Xu, Y. Yagi, S. Yu, and M. Ren, “End-to-end model-based gait recognition,” in Proceedings of the Asian conference on computer vision, 2020.
- X. Li, Y. Makihara, C. Xu, and Y. Yagi, “End-to-end model-based gait recognition using synchronized multi-view pose constraint,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 4106–4115.
- N. Li and X. Zhao, “A strong and robust skeleton-based gait recognition method with gait periodicity priors,” IEEE Transactions on Multimedia, pp. 1–1, 2022.
- K. Xu, X. Jiang, and T. Sun, “Gait recognition based on local graphical skeleton descriptor with pairwise similarity network,” IEEE Transactions on Multimedia, vol. 24, pp. 3265–3275, 2022.
- K. Shiraga, Y. Makihara, D. Muramatsu, T. Echigo, and Y. Yagi, “Geinet: View-invariant gait recognition using a convolutional neural network,” in 2016 international conference on biometrics (ICB). IEEE, 2016, pp. 1–8.
- B. Hu, Y. Gao, Y. Guan, Y. Long, N. Lane, and T. Ploetz, “Robust cross-view gait identification with evidence: A discriminant gait gan (diggan) approach on 10000 people,” arXiv e-prints, pp. arXiv–1811, 2018.
- Y. He, J. Zhang, H. Shan, and L. Wang, “Multi-task gans for view-specific feature learning in gait recognition,” IEEE Transactions on Information Forensics and Security, vol. 14, no. 1, pp. 102–113, 2018.
- W. Liu, C. Zhang, H. Ma, and S. Li, “Learning efficient spatial-temporal gait features with deep learning for human identification,” Neuroinformatics, vol. 16, no. 3, pp. 457–471, 2018.
- T. Wolf, M. Babaee, and G. Rigoll, “Multi-view gait recognition using 3d convolutional neural networks,” in 2016 IEEE international conference on image processing (ICIP). IEEE, 2016, pp. 4165–4169.
- S. Li, W. Liu, and H. Ma, “Attentive spatial–temporal summary networks for feature learning in irregular gait recognition,” IEEE Transactions on Multimedia, vol. 21, no. 9, pp. 2361–2375, 2019.
- B. Lin, S. Zhang, and F. Bao, “Gait recognition with multiple-temporal-scale 3d convolutional neural network,” in ACM MM, 2020, pp. 3054–3062.
- B. Lin, S. Zhang, and X. Yu, “Gait recognition via effective global-local feature representation and local temporal aggregation,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 14 648–14 656.
- L. Yao, W. Kusakunniran, P. Zhang, Q. Wu, and J. Zhang, “Improving disentangled representation learning for gait recognition using group supervision,” IEEE Transactions on Multimedia, pp. 1–1, 2022.
- H. Chao, Y. He, J. Zhang, and J. Feng, “Gaitset: Regarding gait as a set for cross-view gait recognition,” in Proceedings of the AAAI conference on artificial intelligence, vol. 33, no. 01, 2019, pp. 8126–8133.
- C. Fan, Y. Peng, C. Cao, X. Liu, S. Hou, J. Chi, Y. Huang, Q. Li, and Z. He, “Gaitpart: Temporal part-based model for gait recognition,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 14 225–14 233.
- S. Hou, C. Cao, X. Liu, and Y. Huang, “Gait lateral network: Learning discriminative and compact representations for gait recognition,” in Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part IX. Springer, 2020, pp. 382–398.
- S. Hou, X. Liu, C. Cao, and Y. Huang, “Set residual network for silhouette-based gait recognition,” IEEE Transactions on Biometrics, Behavior, and Identity Science, 2021.
- Y. Shi and A. K. Jain, “Probabilistic face embeddings,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 6902–6911.
- T. Yu, D. Li, Y. Yang, T. M. Hospedales, and T. Xiang, “Robust person re-identification by modelling feature uncertainty,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 552–561.
- P. Wang, C. Ding, W. Tan, M. Gong, K. Jia, and D. Tao, “Uncertainty-aware clustering for unsupervised domain adaptive object re-identification,” IEEE Transactions on Multimedia, pp. 1–1, 2022.
- X. Song and Z. Jin, “Robust label rectifying with consistent contrastive-learning for domain adaptive person re-identification,” IEEE Transactions on Multimedia, vol. 24, pp. 3229–3239, 2022.
- X. Yang, Y. Gao, H. Luo, C. Liao, and K.-T. Cheng, “Bayesian denet: Monocular depth prediction and frame-wise fusion with synchronized uncertainty,” IEEE Transactions on Multimedia, vol. 21, no. 11, pp. 2701–2713, 2019.
- D. Guan, J. Huang, A. Xiao, S. Lu, and Y. Cao, “Uncertainty-aware unsupervised domain adaptation in object detection,” IEEE Transactions on Multimedia, vol. 24, pp. 2502–2514, 2022.
- J. Zhuo, S. Wang, and Q. Huang, “Uncertainty modeling for robust domain adaptation under noisy environments,” IEEE Transactions on Multimedia, pp. 1–14, 2022.
- Y. Cui, W. Deng, X. Xu, Z. Liu, Z. Liu, M. Pietikäinen, and L. Liu, “Uncertainty-guided semi-supervised few-shot class-incremental learning with knowledge distillation,” IEEE Transactions on Multimedia, pp. 1–14, 2022.
- J. Hong, W. Zhang, Z. Feng, and W. Zhang, “Dual cross-attention for video object segmentation via uncertainty refinement,” IEEE Transactions on Multimedia, pp. 1–16, 2022.
- J. Chang, Z. Lan, C. Cheng, and Y. Wei, “Data uncertainty learning in face recognition,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 5710–5719.
- Y. Shi, X. Yu, K. Sohn, M. Chandraker, and A. K. Jain, “Towards universal representation learning for deep face recognition,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 6817–6826.
- K. Chen, Q. Lv, T. Yi, and Z. Yi, “Reliable probabilistic face embeddings in the wild,” arXiv preprint arXiv:2102.04075, 2021.
- Y. Shi, W. Tian, H. Ling, Z. Li, and P. Li, “Spatial-wise and channel-wise feature uncertainty for occluded person re-identification,” Neurocomputing, vol. 486, pp. 237–249, 2022.
- D. P. Kingma and M. Welling, “Auto-encoding variational bayes,” arXiv preprint arXiv:1312.6114, 2013.
- A. Paszke, S. Gross, F. Massa, A. Lerer, J. Bradbury, G. Chanan, T. Killeen, Z. Lin, N. Gimelshein, L. Antiga et al., “Pytorch: An imperative style, high-performance deep learning library,” Advances in neural information processing systems, vol. 32, 2019.
- A. Hermans, L. Beyer, and B. Leibe, “In defense of the triplet loss for person re-identification,” arXiv preprint arXiv:1703.07737, 2017.