High-Order Structure Based Middle-Feature Learning for Visible-Infrared Person Re-Identification (2312.07853v2)
Abstract: Visible-infrared person re-identification (VI-ReID) aims to retrieve images of the same persons captured by visible (VIS) and infrared (IR) cameras. Existing VI-ReID methods ignore high-order structure information of features while being relatively difficult to learn a reasonable common feature space due to the large modality discrepancy between VIS and IR images. To address the above problems, we propose a novel high-order structure based middle-feature learning network (HOS-Net) for effective VI-ReID. Specifically, we first leverage a short- and long-range feature extraction (SLE) module to effectively exploit both short-range and long-range features. Then, we propose a high-order structure learning (HSL) module to successfully model the high-order relationship across different local features of each person image based on a whitened hypergraph network.This greatly alleviates model collapse and enhances feature representations. Finally, we develop a common feature space learning (CFL) module to learn a discriminative and reasonable common feature space based on middle features generated by aligning features from different modalities and ranges. In particular, a modality-range identity-center contrastive (MRIC) loss is proposed to reduce the distances between the VIS, IR, and middle features, smoothing the training process. Extensive experiments on the SYSU-MM01, RegDB, and LLCM datasets show that our HOS-Net achieves superior state-of-the-art performance. Our code is available at \url{https://github.com/Jaulaucoeng/HOS-Net}.
- Structure-aware positional Transformer for visible-infrared person re-identification. IEEE Transactions on Image Processing, 31: 2352–2364.
- Mobile-former: Bridging MobileNet and Transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 5270–5279.
- Hierarchical Cross-Modality Disentanglement for Visible-Infrared Person Re-Identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 13–19.
- Cross-modality person re-identification with generative adversarial training. In Proceedings of the International Joint Conference on Artificial Intelligence, 1–7.
- Hypergraph neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, 3558–3565.
- Graph attention tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 9543–9552.
- In defense of the triplet loss for person re-identification. arXiv preprint arXiv:1703.07737.
- Mean field analysis of hypergraph contagion models. SIAM Journal on Applied Mathematics, 82(6): 1987–2007.
- Exploring modality-shared appearance features and modality-invariant relation features for cross-modality person re-identification. Pattern Recognition, 135: 109145.
- Cross-modality person re-identification via multi-task learning. Pattern Recognition, 128: 108653.
- Cloth-changing person re-identification from a single image with gait prediction and regularization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 14278–14287.
- Infrared-visible cross-modal person re-identification with an x modality. In Proceedings of the AAAI Conference on Artificial Intelligence, 4610–4617.
- Decoupled pose and similarity based graph neural network for video person re-identification. IEEE Signal Processing Letters, 29: 264–268.
- Parameter sharing exploration and hetero-center triplet loss for visible-thermal person re-identification. IEEE Transactions on Multimedia, 23: 4414–4425.
- Learning memory-augmented unidirectional metrics for cross-modality person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 19366–19375.
- Revisiting modality-specific feature compensation for visible-infrared person re-identification. IEEE Transactions on Circuits and Systems for Video Technology, 32(10): 7226–7240.
- Learning progressive modality-shared Transformers for effective visible-infrared person re-identification. In Proceedings of the AAAI Conference on Artificial Intelligence, 1835–1843.
- Exploring high-order spatio-temporal correlations from skeleton for person re-identification. IEEE Transactions on Image Processing, 32: 949–963.
- Person recognition system based on a combination of body images from visible light and thermal cameras. Sensors, 17(3): 605.
- Learning by aligning: Visible-infrared person re-identification using cross-modal correspondences. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 12046–12055.
- Visualizing data using t-SNE. Journal of Machine Learning Research, 9(11).
- D22{\mathrm{}^{2}}start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPTDA: Geometry-guided dual-alignment learning for RGB-infrared person re-identification. Pattern Recognition, 135: 109150.
- Cross-modality paired-images generation for RGB-infrared person re-identification. In Proceedings of the AAAI Conference on Artificial Intelligence, 12144–12151.
- Learning to reduce dual-level discrepancy for infrared-visible person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 618–626.
- Rbdf: Reciprocal bidirectional framework for visible infrared person reidentification. IEEE Transactions on Cybernetics, 52(10): 10988–10998.
- RGB-IR person re-identification by cross-modality similarity preservation. International Journal of Computer Vision, 128: 1765–1785.
- Occluded person re-identification with single-scale global representations. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 11875–11884.
- Top-K Visual Tokens Transformer: Selecting Tokens for Visible-Infrared Person Re-Identification. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, 1–5.
- Channel Augmented Joint Learning for Visible-Infrared Recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 13567–13576.
- Dynamic dual-attentive aggregation learning for visible-infrared person re-identification. In Proceedings of the European Conference on Computer Vision, 229–247.
- Deep learning for person re-identification: A survey and outlook. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(6): 2872–2893.
- Dual mutual learning for cross-modality person re-identification. IEEE Transactions on Circuits and Systems for Video Technology, 32(8): 5361–5373.
- Parc-Net: Position aware circular convolution with merits from convnets and Transformer. In Proceedings of the European Conference on Computer Vision, 613–630.
- Diverse embedding expansion network and low-light cross-modality benchmark for visible-infrared person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2153–2162.
- Towards a unified middle modality learning for visible-infrared person re-identification. In Proceedings of the ACM International Conference on Multimedia, 788–796.
- CM-NAS: Cross-modality neural architecture search for visible-infrared person re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 11823–11832.
- Cross-modality person re-identification with shared-specific feature transfer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 13379–13389.
- RGB-infrared cross-modality person re-identification via joint pixel and feature alignment. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 3623–3632.
- RGB-infrared cross-modality person re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 5380–5389.
- Discover cross-modality nuances for visible-infrared person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 4330–4339.
- FMCNet: Feature-level modality compensation for visible-infrared person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 7349–7358.
- Liuxiang Qiu (1 paper)
- Si Chen (83 papers)
- Yan Yan (242 papers)
- Da-Han Wang (8 papers)
- Shunzhi Zhu (2 papers)
- Jing-Hao Xue (54 papers)