Unified-modal Salient Object Detection via Adaptive Prompt Learning (2311.16835v5)
Abstract: Existing single-modal and multi-modal salient object detection (SOD) methods focus on designing specific architectures tailored for their respective tasks. However, developing completely different models for different tasks leads to labor and time consumption, as well as high computational and practical deployment costs. In this paper, we attempt to address both single-modal and multi-modal SOD in a unified framework called UniSOD, which fully exploits the overlapping prior knowledge between different tasks. Nevertheless, assigning appropriate strategies to modality variable inputs is challenging. To this end, UniSOD learns modality-aware prompts with task-specific hints through adaptive prompt learning, which are plugged into the proposed pre-trained baseline SOD model to handle corresponding tasks, while only requiring few learnable parameters compared to training the entire model. Each modality-aware prompt is generated from a switchable prompt generation block, which adaptively performs structural switching based on single-modal and multi-modal inputs without human intervention. Through end-to-end joint training, UniSOD achieves overall performance improvement on 14 benchmark datasets for RGB, RGB-D, and RGB-T SOD, which demonstrates that our method effectively and efficiently unifies single-modal and multi-modal SOD tasks.The code and results are available at https://github.com/Angknpng/UniSOD.
- H. Li, D. Zhang, N. Liu, L. Cheng, Y. Dai, C. Zhang, X. Wang, and J. Han, “Boosting low-data instance segmentation by unsupervised pre-training with saliency prompt,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 15 485–15 494.
- P. Zhang, W. Liu, D. Wang, Y. Lei, H. Wang, and H. Lu, “Non-rigid object tracking via deep multi-scale spatial-temporal discriminative saliency maps,” Pattern Recognit., vol. 100, p. 107130, 2020.
- R. Cong, J. Lei, H. Fu, F. Porikli, Q. Huang, and C. Hou, “Video saliency detection via sparsity-based reconstruction and propagation,” IEEE Trans. Image Process., vol. 28, no. 10, pp. 4819–4831, 2019.
- D. Fan, W. Wang, M. Cheng, and J. Shen, “Shifting more attention to video salient object detection,” in IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 8554–8564.
- H. Zhou, B. Qiao, L. Yang, J. Lai, and X. Xie, “Texture-guided saliency distilling for unsupervised salient object detection,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 7257–7267.
- M. Ma, C. Xia, C. Xie, X. Chen, and J. Li, “Boosting broader receptive fields for salient object detection,” IEEE Trans. Image Process., vol. 32, pp. 1026–1038, 2023.
- X. Tian, J. Zhang, M. Xiang, and Y. Dai, “Modeling the distributional uncertainty for salient object detection models,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 19 660–19 670.
- W. Ji, G. Yan, J. Li, Y. Piao, S. Yao, M. Zhang, L. Cheng, and H. Lu, “DMRA: depth-induced multi-scale recurrent attention network for RGB-D saliency detection,” IEEE Trans. Image Process., vol. 31, pp. 2321–2336, 2022.
- J. Zhao, Y. Cao, D. Fan, M. Cheng, X. Li, and L. Zhang, “Contrast prior and fluid pyramid integration for RGBD salient object detection,” in IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 3927–3936.
- K. Fu, D. Fan, G. Ji, and Q. Zhao, “JL-DCF: joint learning and densely-cooperative fusion framework for RGB-D salient object detection,” in 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 3049–3059.
- J. Zhang, D. Fan, Y. Dai, S. Anwar, F. S. Saleh, T. Zhang, and N. Barnes, “Uc-net: Uncertainty inspired RGB-D saliency detection via conditional variational autoencoders,” in 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 8579–8588.
- N. Liu, N. Zhang, and J. Han, “Learning selective self-mutual attention for RGB-D saliency detection,” in 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 13 753–13 762.
- Z. Tu, T. Xia, C. Li, X. Wang, Y. Ma, and J. Tang, “RGB-T image saliency detection via collaborative graph learning,” IEEE Trans. Multim., vol. 22, no. 1, pp. 160–173, 2020.
- Z. Tu, Y. Ma, Z. Li, C. Li, J. Xu, and Y. Liu, “RGBT salient object detection: A large-scale dataset and benchmark,” CoRR, vol. abs/2007.03262, 2020.
- Z. Tu, Z. Li, C. Li, Y. Lang, and J. Tang, “Multi-interactive dual-decoder for rgb-thermal salient object detection,” IEEE Trans. Image Process., vol. 30, pp. 5678–5691, 2021.
- Z. Liu, Y. Tan, Q. He, and Y. Xiao, “Swinnet: Swin transformer drives edge-aware RGB-D and RGB-T salient object detection,” IEEE Trans. Circuits Syst. Video Technol., vol. 32, no. 7, pp. 4486–4497, 2022.
- X. Zhao, Y. Pang, L. Zhang, and H. Lu, “Joint learning of salient object detection, depth estimation and contour extraction,” IEEE Trans. Image Process., vol. 31, pp. 7350–7362, 2022.
- V. Casser, S. Pirk, R. Mahjourian, and A. Angelova, “Depth prediction without the sensors: Leveraging structure for unsupervised learning from monocular videos,” in Proceedings of the AAAI conference on artificial intelligence, 2019, pp. 8001–8008.
- T. X. B. Nguyen, K. Rosser, and J. S. Chahl, “A review of modern thermal imaging sensor technology and applications for autonomous aerial navigation,” J. Imaging, vol. 7, no. 10, p. 217, 2021.
- A. George, Z. Mostaani, D. Geissenbuhler, O. Nikisins, A. Anjos, and S. Marcel, “Biometric face presentation attack detection with multi-channel convolutional neural network,” IEEE Trans. Inf. Forensics Secur., vol. 15, pp. 42–55, 2020.
- N. Liu, N. Zhang, K. Wan, L. Shao, and J. Han, “Visual saliency transformer,” in 2021 IEEE/CVF International Conference on Computer Vision, 2021, pp. 4702–4712.
- Y. Pang, X. Zhao, L. Zhang, and H. Lu, “CAVER: cross-modal view-mixed transformer for bi-modal salient object detection,” IEEE Trans. Image Process., vol. 32, pp. 892–904, 2023.
- W. Zhou, Y. Zhu, J. Lei, R. Yang, and L. Yu, “Lsnet: Lightweight spatial boosting network for detecting salient objects in rgb-thermal images,” IEEE Trans. Image Process., vol. 32, pp. 1329–1340, 2023.
- B. Tang, Z. Liu, Y. Tan, and Q. He, “Hrtransnet: Hrformer-driven two-modality salient object detection,” IEEE Trans. Circuits Syst. Video Technol., vol. 33, no. 2, pp. 728–742, 2023.
- Y. K. Yun and W. Lin, “Towards a complete and detail-preserved salient object detection,” IEEE Transactions on Multimedia, pp. 1–15, 2023.
- M. Zhuge, D. Fan, N. Liu, D. Zhang, D. Xu, and L. Shao, “Salient object detection via integrity learning,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 45, no. 3, pp. 3738–3752, 2023.
- R. Cong, H. Liu, C. Zhang, W. Zhang, F. Zheng, R. Song, and S. Kwong, “Point-aware interaction and cnn-induced refinement network for RGB-D salient object detection,” in Proceedings of the 31st ACM International Conference on Multimedia, 2023, pp. 406–416.
- F. Sun, P. Ren, B. Yin, F. Wang, and H. Li, “Catnet: A cascaded and aggregated transformer network for rgb-d salient object detection,” IEEE Transactions on Multimedia, pp. 1–14, 2023.
- W. Zhou, F. Sun, Q. Jiang, R. Cong, and J. Hwang, “Wavenet: Wavelet network with knowledge distillation for RGB-T salient object detection,” IEEE Trans. Image Process., vol. 32, pp. 3027–3039, 2023.
- L. Wang, H. Lu, Y. Wang, M. Feng, D. Wang, B. Yin, and X. Ruan, “Learning to detect salient objects with image-level supervision,” in 2017 IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 3796–3805.
- Y. Niu, Y. Geng, X. Li, and F. Liu, “Leveraging stereopsis for saliency analysis,” in 2012 IEEE Conference on Computer Vision and Pattern Recognition, 2012, pp. 454–461.
- B. Lester, R. Al-Rfou, and N. Constant, “The power of scale for parameter-efficient prompt tuning,” in Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, M. Moens, X. Huang, L. Specia, and S. W. Yih, Eds., 2021, pp. 3045–3059.
- P. Liu, W. Yuan, J. Fu, Z. Jiang, H. Hayashi, and G. Neubig, “Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing,” ACM Comput. Surv., vol. 55, no. 9, pp. 195:1–195:35, 2023.
- M. Jia, L. Tang, B. Chen, C. Cardie, S. J. Belongie, B. Hariharan, and S. Lim, “Visual prompt tuning,” in European Conference on Computer Vision, S. Avidan, G. J. Brostow, M. Cissé, G. M. Farinella, and T. Hassner, Eds., vol. 13693, 2022, pp. 709–727.
- Y. Lee, Y. Tsai, W. Chiu, and C. Lee, “Multimodal prompting with missing modalities for visual recognition,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 14 943–14 952.
- J. Zhu, S. Lai, X. Chen, D. Wang, and H. Lu, “Visual prompt multi-modal tracking,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 9516–9526.
- K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” in 3rd International Conference on Learning Representations, Y. Bengio and Y. LeCun, Eds., 2015.
- K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in 2016 IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 770–778.
- S. Gao, M. Cheng, K. Zhao, X. Zhang, M. Yang, and P. H. S. Torr, “Res2net: A new multi-scale backbone architecture,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 43, no. 2, pp. 652–662, 2021.
- W. Wang, Q. Lai, H. Fu, J. Shen, H. Ling, and R. Yang, “Salient object detection in the deep learning era: An in-depth survey,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 44, no. 6, pp. 3239–3259, 2022.
- W. Wang, S. Zhao, J. Shen, S. C. H. Hoi, and A. Borji, “Salient object detection with pyramid attention and salient edges,” in IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 1448–1457.
- J. Zhao, J. Liu, D. Fan, Y. Cao, J. Yang, and M. Cheng, “Egnet: Edge guidance network for salient object detection,” in 2019 IEEE/CVF International Conference on Computer Vision, 2019, pp. 8778–8787.
- X. Qin, Z. Zhang, C. Huang, C. Gao, M. Dehghan, and M. Jägersand, “Basnet: Boundary-aware salient object detection,” in IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 7479–7489.
- Y. K. Yun and T. Tsubono, “Recursive contour-saliency blending network for accurate salient object detection,” in IEEE/CVF Winter Conference on Applications of Computer Vision, 2022, pp. 1360–1370.
- Z. Wu, L. Su, and Q. Huang, “Cascaded partial decoder for fast and accurate salient object detection,” in IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 3907–3916.
- X. Zhao, Y. Pang, L. Zhang, H. Lu, and L. Zhang, “Suppress and balance: A simple gated network for salient object detection,” in European Conference on Computer Vision, A. Vedaldi, H. Bischof, T. Brox, and J. Frahm, Eds., vol. 12347, 2020, pp. 35–51.
- Y. Wu, Y. Liu, L. Zhang, M. Cheng, and B. Ren, “EDN: salient object detection via extremely-downsampled network,” IEEE Trans. Image Process., vol. 31, pp. 3125–3136, 2022.
- J. Liu, Q. Hou, Z. Liu, and M. Cheng, “Poolnet+: Exploring the potential of pooling for salient object detection,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 45, no. 1, pp. 887–904, 2023.
- J. Wei, S. Wang, Z. Wu, C. Su, Q. Huang, and Q. Tian, “Label decoupling framework for salient object detection,” in 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 13 022–13 031.
- J. Wei, S. Wang, and Q. Huang, “F3net: Fusion, feedback and focus for salient object detection,” in Proceedings of the AAAI conference on artificial intelligence, 2020, pp. 12 321–12 328.
- N. Liu, J. Han, and M. Yang, “Picanet: Pixel-wise contextual attention learning for accurate saliency detection,” IEEE Trans. Image Process., vol. 29, pp. 6438–6451, 2020.
- Z. Zhou, Z. Wang, H. Lu, S. Wang, and M. Sun, “Multi-type self-attention guided degraded saliency detection,” in Proceedings of the AAAI conference on artificial intelligence, 2020, pp. 13 082–13 089.
- Y. Wang, R. Wang, X. Fan, T. Wang, and X. He, “Pixels, regions, and objects: Multiple enhancement for salient object detection,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 10 031–10 040.
- A. Li, J. Zhang, Y. Lv, B. Liu, T. Zhang, and Y. Dai, “Uncertainty-aware joint salient object and camouflaged object detection,” in IEEE Conference on Computer Vision and Pattern Recognition, 2021, pp. 10 071–10 081.
- Y. Wang, W. Zhang, L. Wang, T. Liu, and H. Lu, “Multi-source uncertainty mining for deep unsupervised saliency detection,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 11 717–11 726.
- C. Xie, C. Xia, M. Ma, Z. Zhao, X. Chen, and J. Li, “Pyramid grafting network for one-stage high resolution saliency detection,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 11 707–11 716.
- A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, and I. Polosukhin, “Attention is all you need,” in Advances in Neural Information Processing Systems, I. Guyon, U. von Luxburg, S. Bengio, H. M. Wallach, R. Fergus, S. V. N. Vishwanathan, and R. Garnett, Eds., 2017, pp. 5998–6008.
- Y. Piao, W. Ji, J. Li, M. Zhang, and H. Lu, “Depth-induced multi-scale recurrent attention network for saliency detection,” in 2019 IEEE/CVF International Conference on Computer Vision, 2019, pp. 7253–7262.
- W. Ji, J. Li, S. Yu, M. Zhang, Y. Piao, S. Yao, Q. Bi, K. Ma, Y. Zheng, H. Lu, and L. Cheng, “Calibrated RGB-D salient object detection,” in IEEE Conference on Computer Vision and Pattern Recognition, 2021, pp. 9471–9481.
- P. Sun, W. Zhang, H. Wang, S. Li, and X. Li, “Deep RGB-D saliency detection with depth-sensitive attention and automatic multi-modal fusion,” in IEEE Conference on Computer Vision and Pattern Recognition, 2021, pp. 1407–1417.
- W. Ji, J. Li, Q. Bi, C. Guo, J. Liu, and L. Cheng, “Promoting saliency from depth: Deep unsupervised RGB-D saliency detection,” in The Tenth International Conference on Learning Representations, 2022.
- Z. Tu, Z. Li, C. Li, and J. Tang, “Weakly alignment-free RGBT salient object detection with deep correlation network,” IEEE Trans. Image Process., vol. 31, pp. 3752–3764, 2022.
- C. Ge, J. Chen, E. Xie, Z. Wang, L. Hong, H. Lu, Z. Li, and P. Luo, “Metabev: Solving sensor failures for 3d detection and map segmentation,” in Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2023, pp. 8721–8731.
- A. Konwer, X. Hu, J. Bae, X. Xu, C. Chen, and P. Prasanna, “Enhancing modality-agnostic representations via meta-learning for brain tumor segmentation,” in Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2023, pp. 21 415–21 425.
- C. Chen, J. Wei, C. Peng, W. Zhang, and H. Qin, “Improved saliency detection in RGB-D images using two-phase depth estimation and selective deep fusion,” IEEE Trans. Image Process., vol. 29, pp. 4296–4307, 2020.
- Y. Zhang, J. Zheng, W. Jia, W. Huang, L. Li, N. Liu, F. Li, and X. He, “Deep RGB-D saliency detection without depth,” IEEE Trans. Multim., vol. 24, pp. 755–767, 2022.
- C. Tu, Z. Mai, and W. Chao, “Visual query tuning: Towards effective usage of intermediate representations for parameter and memory efficient transfer learning,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 7725–7735.
- B. Dong, P. Zhou, S. Yan, and W. Zuo, “LPT: long-tailed prompt tuning for image classification,” in The Eleventh International Conference on Learning Representations, 2023.
- X. Nie, B. Ni, J. Chang, G. Meng, C. Huo, S. Xiang, and Q. Tian, “Pro-tuning: Unified prompt tuning for vision tasks,” IEEE Transactions on Circuits and Systems for Video Technology, 2023.
- T. Zhou, H. Fu, G. Chen, Y. Zhou, D. Fan, and L. Shao, “Specificity-preserving RGB-D saliency detection,” in 2021 IEEE/CVF International Conference on Computer Vision, 2021, pp. 4661–4671.
- Y. Pang, X. Zhao, L. Zhang, and H. Lu, “Multi-scale interactive network for salient object detection,” in 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 9410–9419.
- Z. Liu, Y. Lin, Y. Cao, H. Hu, Y. Wei, Z. Zhang, S. Lin, and B. Guo, “Swin transformer: Hierarchical vision transformer using shifted windows,” in 2021 IEEE/CVF International Conference on Computer Vision, 2021, pp. 9992–10 002.
- A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, J. Uszkoreit, and N. Houlsby, “An image is worth 16x16 words: Transformers for image recognition at scale,” in 9th International Conference on Learning Representations, 2021.
- Y. Wu, Y. Liu, J. Xu, J. Bian, Y. Gu, and M. Cheng, “Mobilesal: Extremely efficient RGB-D salient object detection,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 44, no. 12, pp. 10 261–10 269, 2022.
- C. Godard, O. M. Aodha, and G. J. Brostow, “Unsupervised monocular depth estimation with left-right consistency,” in 2017 IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 6602–6611.
- F. Milletari, N. Navab, and S. Ahmadi, “V-net: Fully convolutional neural networks for volumetric medical image segmentation,” in Fourth International Conference on 3D Vision, 2016, pp. 565–571.
- M. Feng, H. Lu, and E. Ding, “Attentive feedback network for boundary-aware salient object detection,” in IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 1623–1632.
- J. Liu, Q. Hou, M. Cheng, J. Feng, and J. Jiang, “A simple pooling-based design for real-time salient object detection,” in IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 3917–3926.
- B. Xu, H. Liang, R. Liang, and P. Chen, “Locate globally, segment locally: A progressive architecture with knowledge review network for salient object detection,” in Proceedings of the AAAI conference on artificial intelligence, 2021, pp. 3004–3012.
- M. Ma, C. Xia, and J. Li, “Pyramidal feature shrinking for salient object detection,” in Proceedings of the AAAI conference on artificial intelligence, 2021, pp. 2311–2318.
- G. Li, Z. Liu, M. Chen, Z. Bai, W. Lin, and H. Ling, “Hierarchical alternate interaction network for RGB-D salient object detection,” IEEE Trans. Image Process., vol. 30, pp. 3528–3542, 2021.
- Q. Chen, Z. Liu, Y. Zhang, K. Fu, Q. Zhao, and H. Du, “RGB-D salient object detection via 3d convolutional neural networks,” in Proceedings of the AAAI conference on artificial intelligence, 2021, pp. 1063–1071.
- X. Zhao, Y. Pang, L. Zhang, H. Lu, and X. Ruan, “Self-supervised pretraining for RGB-D salient object detection,” in Proceedings of the AAAI conference on artificial intelligence, 2022, pp. 3463–3471.
- X. Jin, K. Yi, and J. Xu, “Moadnet: Mobile asymmetric dual-stream networks for real-time and lightweight RGB-D salient object detection,” IEEE Trans. Circuits Syst. Video Technol., vol. 32, no. 11, pp. 7632–7645, 2022.
- R. Cong, Q. Lin, C. Zhang, C. Li, X. Cao, Q. Huang, and Y. Zhao, “Cir-net: Cross-modality interaction and refinement for RGB-D salient object detection,” IEEE Trans. Image Process., vol. 31, pp. 6800–6815, 2022.
- X. Cheng, X. Zheng, J. Pei, H. Tang, Z. Lyu, and C. Chen, “Depth-induced gap-reducing network for RGB-D salient object detection: An interaction, guidance and refinement approach,” IEEE Trans. Multim., vol. 25, pp. 4253–4266, 2023.
- C. Yang, L. Zhang, H. Lu, X. Ruan, and M. Yang, “Saliency detection via graph-based manifold ranking,” in 2013 IEEE Conference on Computer Vision and Pattern Recognition, 2013, pp. 3166–3173.
- Q. Yan, L. Xu, J. Shi, and J. Jia, “Hierarchical saliency detection,” in 2013 IEEE Conference on Computer Vision and Pattern Recognition, 2013, pp. 1155–1162.
- G. Li and Y. Yu, “Visual saliency based on multiscale deep features,” in IEEE Conference on Computer Vision and Pattern Recognition, 2015, pp. 5455–5463.
- Y. Li, X. Hou, C. Koch, J. M. Rehg, and A. L. Yuille, “The secrets of salient object segmentation,” in 2014 IEEE Conference on Computer Vision and Pattern Recognition, 2014, pp. 280–287.
- R. Ju, L. Ge, W. Geng, T. Ren, and G. Wu, “Depth saliency based on anisotropic center-surround difference,” in 2014 IEEE International Conference on Image Processing, 2014, pp. 1115–1119.
- H. Peng, B. Li, W. Xiong, W. Hu, and R. Ji, “RGBD salient object detection: A benchmark and algorithms,” in European Conference on Computer Vision, D. J. Fleet, T. Pajdla, B. Schiele, and T. Tuytelaars, Eds., vol. 8691, 2014, pp. 92–109.
- G. Li and C. Zhu, “A three-pathway psychobiological framework of salient object detection using stereoscopic technology,” in 2017 IEEE International Conference on Computer Vision Workshops, 2017, pp. 3008–3014.
- D. Fan, Z. Lin, Z. Zhang, M. Zhu, and M. Cheng, “Rethinking RGB-D salient object detection: Models, data sets, and large-scale benchmarks,” IEEE Trans. Neural Networks Learn. Syst., vol. 32, no. 5, pp. 2075–2089, 2021.
- J. Tang, D. Fan, X. Wang, Z. Tu, and C. Li, “RGBT salient object detection: Benchmark and A novel cooperative ranking approach,” IEEE Trans. Circuits Syst. Video Technol., vol. 30, no. 12, pp. 4421–4433, 2020.
- S. Ma, K. Song, H. Dong, H. Tian, and Y. Yan, “Modal complementary fusion network for RGB-T salient object detection,” Appl. Intell., vol. 53, no. 8, pp. 9038–9055, 2023.
- D. Fan, C. Gong, Y. Cao, B. Ren, M. Cheng, and A. Borji, “Enhanced-alignment measure for binary foreground map evaluation,” in Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, J. Lang, Ed., 2018, pp. 698–704.
- M. Cheng and D. Fan, “Structure-measure: A new way to evaluate foreground maps,” Int. J. Comput. Vis., vol. 129, no. 9, pp. 2622–2638, 2021.
- R. Margolin, L. Zelnik-Manor, and A. Tal, “How to evaluate foreground maps,” in 2014 IEEE Conference on Computer Vision and Pattern Recognition, 2014, pp. 248–255.
- F. Perazzi, P. Krähenbühl, Y. Pritch, and A. Hornung, “Saliency filters: Contrast based filtering for salient region detection,” in 2012 IEEE Conference on Computer Vision and Pattern Recognition, 2012, pp. 733–740.
- A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” Commun. ACM, vol. 60, no. 6, pp. 84–90, 2017.
- I. Loshchilov and F. Hutter, “Decoupled weight decay regularization,” in 7th International Conference on Learning Representations, 2019.