{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,8]],"date-time":"2025-10-08T15:31:37Z","timestamp":1759937497283,"version":"3.37.3"},"reference-count":68,"publisher":"Springer Science and Business Media LLC","issue":"20","license":[{"start":{"date-parts":[[2023,12,19]],"date-time":"2023-12-19T00:00:00Z","timestamp":1702944000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,12,19]],"date-time":"2023-12-19T00:00:00Z","timestamp":1702944000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U21B6001"],"award-info":[{"award-number":["U21B6001"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012593","name":"Scientific Research Foundation of Graduate School of Harbin Medical University: Sino Russian Special Fund","doi-asserted-by":"publisher","award":["2021YJSO2S02"],"award-info":[{"award-number":["2021YJSO2S02"]}],"id":[{"id":"10.13039\/501100012593","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-023-17747-y","type":"journal-article","created":{"date-parts":[[2023,12,19]],"date-time":"2023-12-19T10:02:26Z","timestamp":1702980146000},"page":"59169-59193","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Adaptive interactive network for RGB-T salient object detection with double mapping transformer"],"prefix":"10.1007","volume":"83","author":[{"given":"Feng","family":"Dong","sequence":"first","affiliation":[]},{"given":"Yuxuan","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Jinchao","family":"Zhu","sequence":"additional","affiliation":[]},{"given":"Yuehua","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,12,19]]},"reference":[{"issue":"27","key":"17747_CR1","doi-asserted-by":"publisher","first-page":"38921","DOI":"10.1007\/s11042-022-13083-9","volume":"81","author":"Z Zhou","year":"2022","unstructured":"Zhou Z, Guo Y, Huang J, Dai M, Deng M, Yu Q (2022) Superpixel attention guided network for accurate and real-time salient object detection. Multimedia Tools Appl 81(27):38921\u201338944","journal-title":"Multimedia Tools Appl"},{"issue":"25","key":"17747_CR2","doi-asserted-by":"publisher","first-page":"35831","DOI":"10.1007\/s11042-021-11555-y","volume":"81","author":"N Yang","year":"2022","unstructured":"Yang N, Zhang C, Zhang Y, Yang H, Du L (2022) A benchmark dataset and baseline model for co-salient object detection within RGB-D images. Multimedia Tool Appl 81(25):35831\u201335842","journal-title":"Multimedia Tool Appl"},{"issue":"19","key":"17747_CR3","doi-asserted-by":"publisher","first-page":"27551","DOI":"10.1007\/s11042-022-12839-7","volume":"81","author":"Y Wang","year":"2022","unstructured":"Wang Y, Zhou T, Li Z, Huang H, Qu B (2022) Salient object detection based on multi-feature graphs and improved manifold ranking. Multimedia Tools Appl 81(19):27551\u201327567","journal-title":"Multimedia Tools Appl"},{"key":"17747_CR4","doi-asserted-by":"publisher","first-page":"5678","DOI":"10.1109\/TIP.2021.3087412","volume":"30","author":"Z Tu","year":"2021","unstructured":"Tu Z, Li Z, Li C, Lang Y, Tang J (2021) Multi-interactive dual-decoder for RGB-thermal salient object detection. IEEE Trans Image Process 30:5678\u20135691","journal-title":"IEEE Trans Image Process"},{"key":"17747_CR5","doi-asserted-by":"crossref","unstructured":"Fan D-P, Lin Z, Zhang Z, Zhu M, Cheng M-M (2020) Rethinking RGB-D salient object detection: models, datasets, and large-scale benchmarks. IEEE Trans Neural Netw Learn Syst","DOI":"10.1109\/TNNLS.2020.2996406"},{"key":"17747_CR6","doi-asserted-by":"publisher","first-page":"12023","DOI":"10.1609\/aaai.v34i07.6879","volume":"34","author":"S Song","year":"2020","unstructured":"Song S, Yu H, Miao Z, Fang J, Zheng K, Ma C, Wang S (2020) Multi-spectral salient object detection by adversarial domain adaptation. Proceedings of the AAAI conference on artificial intelligence 34:12023\u201312030","journal-title":"Proceedings of the AAAI conference on artificial intelligence"},{"key":"17747_CR7","doi-asserted-by":"crossref","unstructured":"Liu Y, Zhang Q, Zhang D, Han J (2019) Employing deep part-object relationships for salient object detection, vol 2019-October, pp 1232\u20131241","DOI":"10.1109\/ICCV.2019.00132"},{"issue":"7","key":"17747_CR8","first-page":"3688","volume":"44","author":"Y Liu","year":"2022","unstructured":"Liu Y, Zhang D, Zhang Q, Han J (2022) Part-object relational visual saliency. IEEE Trans Pattern Anal Mach Intell 44(7):3688\u20133704","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"17747_CR9","doi-asserted-by":"publisher","first-page":"6719","DOI":"10.1109\/TIP.2022.3215887","volume":"31","author":"Y Liu","year":"2022","unstructured":"Liu Y, Zhang D, Liu N, Xu S, Han J (2022) Disentangled capsule routing for fast part-object relational saliency. IEEE Trans Image Process 31:6719\u20136732","journal-title":"IEEE Trans Image Process"},{"key":"17747_CR10","doi-asserted-by":"crossref","unstructured":"Cheng M-M, Zhang F-L, Mitra NJ, Huang X, Hu S-M (2010) Repfinder: finding approximately repeated scene elements for image editing. ACM Trans Graph 29(4)","DOI":"10.1145\/1833351.1778820"},{"issue":"5","key":"17747_CR11","first-page":"1","volume":"28","author":"T Chen","year":"2009","unstructured":"Chen T, Cheng M-M, Tan P, Shamir A, Hu S-M (2009) Sketch2photo: internet image montage. ACM Trans Graph 28(5):1\u201310","journal-title":"ACM Trans Graph"},{"key":"17747_CR12","doi-asserted-by":"crossref","unstructured":"Mahadevan V, Vasconcelos N (2009) Saliency-based discriminant tracking. In: Proceeding of the IEEE conference on computer vision and pattern recognition (CVPR), pp 1007\u20131013","DOI":"10.1109\/CVPRW.2009.5206573"},{"key":"17747_CR13","doi-asserted-by":"publisher","first-page":"1680","DOI":"10.1109\/LSP.2020.3025128","volume":"27","author":"S Chen","year":"2020","unstructured":"Chen S, Li Z, Tang Z (2020) Relation R-CNN: a graph based relation-aware network for object detection. IEEE Signal Process Lett 27:1680\u20131684","journal-title":"IEEE Signal Process Lett"},{"issue":"9","key":"17747_CR14","doi-asserted-by":"publisher","first-page":"4299","DOI":"10.1007\/s00521-020-05255-1","volume":"33","author":"Y Quan","year":"2021","unstructured":"Quan Y, Li Z, Chen S, Zhang C, Ma H (2021) Joint deep separable convolution network and border regression reinforcement for object detection. Neural Comput Appl 33(9):4299\u20134314","journal-title":"Neural Comput Appl"},{"key":"17747_CR15","doi-asserted-by":"crossref","unstructured":"Wang H, Zhu J, Dai W, Liu J (2019) A Re-ID and tracking-by-detection framework for multiple wildlife tracking with artiodactyla characteristics in ecological surveillance. In: Proceeding of the IEEE international conference on real-time computing and robotics (RCAR), pp 901\u2013906","DOI":"10.1109\/RCAR47638.2019.9043947"},{"key":"17747_CR16","doi-asserted-by":"crossref","unstructured":"Zhu J, Wang H, Han D, Liu J (2018) Smart surveillance: a nature ecological intelligent surveillance system with robotic observation cameras and environment factors sensors. In: Proceeding of the IEEE international conference on CYBER technology in automation, control, and intelligent systems (CYBER), pp 451\u2013456","DOI":"10.1109\/CYBER.2018.8688130"},{"key":"17747_CR17","doi-asserted-by":"crossref","unstructured":"Wang G, Li C, Ma Y, Zheng A, Tang J, Luo B (2018) RGB-T saliency detection benchmark: dataset, baselines, analysis and a novel approach. In: Image Graph Technol Appl (IGTA), pp 359\u2013369","DOI":"10.1007\/978-981-13-1702-6_36"},{"issue":"12","key":"17747_CR18","doi-asserted-by":"publisher","first-page":"4421","DOI":"10.1109\/TCSVT.2019.2951621","volume":"30","author":"J Tang","year":"2020","unstructured":"Tang J, Fan D, Wang X, Tu Z, Li C (2020) RGBT salient object detection: benchmark and a novel cooperative ranking approach. IEEE Trans Circuits Syst Video Technol 30(12):4421\u20134433","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"17747_CR19","doi-asserted-by":"crossref","unstructured":"Tu Z, Xia T, Li C, Lu Y, Tang J (2019) M3S-NIR: multi-modal multi-scale noise-insensitive ranking for RGB-T saliency detection. In: Proceeding of the IEEE conference on multimedia information processing and retrieval (MIPR), pp 141\u2013146","DOI":"10.1109\/MIPR.2019.00032"},{"issue":"1","key":"17747_CR20","doi-asserted-by":"publisher","first-page":"160","DOI":"10.1109\/TMM.2019.2924578","volume":"22","author":"Z Tu","year":"2020","unstructured":"Tu Z, Xia T, Li C, Wang X, Ma Y, Tang J (2020) rGB-T image saliency detection via collaborative graph learning. Trans Multimedia 22(1):160\u2013173","journal-title":"Trans Multimedia"},{"key":"17747_CR21","unstructured":"Tu Z, Ma Y, Li Z, Li C, Xu J, Liu Y (2020) RGBT salient object detection: a large-scale dataset and benchmark. arXiv:2007.03262"},{"key":"17747_CR22","doi-asserted-by":"publisher","first-page":"3321","DOI":"10.1109\/TIP.2019.2959253","volume":"29","author":"Q Zhang","year":"2020","unstructured":"Zhang Q, Huang N, Yao L, Zhang D, Shan C, Han J (2020) RGB-T salient object detection via fusing multi-level CNN features. IEEE Trans Image Process 29:3321\u20133335","journal-title":"IEEE Trans Image Process"},{"key":"17747_CR23","doi-asserted-by":"crossref","unstructured":"Woo S, Park J, Lee J-Y, Kweon IS (2018) CBAM: convolutional block attention module. In: Proceeding of the Europeon conference on computer vision (ECCV)","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"17747_CR24","doi-asserted-by":"crossref","unstructured":"Zhu C, Cai X, Huang K, Li TH, Li G (2019) PDNet: prior-model guided depth-enhanced network for salient object detection. In: Proceeding of the IEEE international conference on multimedia and expo (ICME), pp 199\u2013204","DOI":"10.1109\/ICME.2019.00042"},{"key":"17747_CR25","doi-asserted-by":"publisher","first-page":"7012","DOI":"10.1109\/TIP.2020.3028289","volume":"30","author":"Z Chen","year":"2021","unstructured":"Chen Z, Cong R, Xu Q, Huang Q (2021) DPANet: depth potentiality-aware gated attention network for RGB-D salient object detection. IEEE Trans Image Process 30:7012\u20137024","journal-title":"IEEE Trans Image Process"},{"key":"17747_CR26","doi-asserted-by":"crossref","unstructured":"Li G, Liu Z, Ye L, Wang Y, Ling H (2020) Cross-modal weighting network for RGB-D salient object detection. In: Proceeding of the Europeon conference on computer vision (ECCV), pp 665\u2013681","DOI":"10.1007\/978-3-030-58520-4_39"},{"key":"17747_CR27","doi-asserted-by":"crossref","unstructured":"Pang Y, Zhang L, Zhao X, Lu H (2020) Hierarchical dynamic filtering network for RGB-D salient object detection. In: Proceeding of the Europeon conference on computer vision (ECCV)","DOI":"10.1007\/978-3-030-58595-2_15"},{"key":"17747_CR28","doi-asserted-by":"publisher","first-page":"1343","DOI":"10.1109\/TMM.2020.2997184","volume":"23","author":"B Jiang","year":"2021","unstructured":"Jiang B, Zhou Z, Wang X, Tang J, Luo B (2021) cmSalGAN: RGB-D salient object detection with cross-view generative adversarial networks. IEEE Trans Multimedia 23:1343\u20131353","journal-title":"IEEE Trans Multimedia"},{"key":"17747_CR29","doi-asserted-by":"crossref","unstructured":"Wang X, Girshick R, Gupta A, He K (2018) Non-local neural networks. In: Proceeding of the IEEE conference on computer vision and pattern recognition (CVPR), pp 7794\u20137803","DOI":"10.1109\/CVPR.2018.00813"},{"key":"17747_CR30","unstructured":"Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser L, Polosukhin I (2017) Attention is all you need. In: Proceeding of the conference on neural information processing systems (NeurIPS)"},{"key":"17747_CR31","doi-asserted-by":"crossref","unstructured":"Mallick R, Benois-Pineau J, Zemmari A (2022) I saw: a self-attention weighted method for explanation of visual transformers. In: 2022 IEEE international conference on image processing (ICIP), pp 3271\u20133275","DOI":"10.1109\/ICIP46576.2022.9897347"},{"key":"17747_CR32","doi-asserted-by":"crossref","unstructured":"Zhao X, Zhang L, Pang Y, Lu H, Zhang L (2020) A single stream network for robust and real-time RGB-D salient object detection. In: Proceeding of the Europeon conference on computer vision (ECCV)","DOI":"10.1007\/978-3-030-58542-6_39"},{"key":"17747_CR33","doi-asserted-by":"crossref","unstructured":"Wang W, Xie E, Li X, Fan D-P, Song K, Liang D, Lu T, Luo P, Shao L (2021) Pyramid vision transformer: a versatile backbone for dense prediction without convolutions. In: Proceeding of the IEEE international conference on computer vision (ICCV)","DOI":"10.1109\/ICCV48922.2021.00061"},{"key":"17747_CR34","doi-asserted-by":"crossref","unstructured":"Liu Z, Lin Y, Cao Y, Hu H, Wei Y, Zhang Z, Lin S, Guo B (2021) Swin transformer: hierarchical vision transformer using shifted windows. In: Proceeding of the international conference on computer vision (ICCV), pp 9992\u201310002","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"17747_CR35","doi-asserted-by":"crossref","unstructured":"Zhang J, Liu H, Yang K, Hu X, Liu R, Stiefelhagen R (2023) CMX: cross-modal fusion for RGB-X semantic segmentation with transformers. IEEE Trans Intell Transp Syst 1\u201316","DOI":"10.1109\/TITS.2023.3300537"},{"key":"17747_CR36","doi-asserted-by":"crossref","unstructured":"Shin U, Lee K, Kweon IS (2023) Complementary random masking for RGB-thermal semantic segmentation","DOI":"10.1109\/ICRA57147.2024.10611200"},{"key":"17747_CR37","doi-asserted-by":"crossref","unstructured":"Liu N, Zhang N, Wan K, Shao L, Han J (2021) Visual saliency transformer. In: Proceeding of the IEEE international conference on computer vision (ICCV), pp 4702\u20134712","DOI":"10.1109\/ICCV48922.2021.00468"},{"key":"17747_CR38","doi-asserted-by":"crossref","unstructured":"Zhu J, Zhang X, Fang X, Dong F, Qiu Y (2021) Modal-adaptive gated recoding network for RGB-D salient object detection. IEEE Signal Process Lett 1\u20131","DOI":"10.1109\/LSP.2021.3125268"},{"key":"17747_CR39","doi-asserted-by":"crossref","unstructured":"Wang W, Xie E, Li X, Fan D-P, Song K, Liang D, Lu T, Luo P, Shao L (2021) PVTv2: improved baselines with pyramid vision transformer. arXiv:2106.13797","DOI":"10.1007\/s41095-022-0274-8"},{"key":"17747_CR40","unstructured":"Park J, Woo S, Lee J, Kweon IS (2018) BAM: bottleneck attention module. In: Proceeding of the British machine vision conference (BMVC), p 147"},{"key":"17747_CR41","unstructured":"Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S, Uszkoreit J, Houlsby N (2021) An image is worth 16x16 words: transformers for image recognition at scale. In: International conference on learning representations (ICLR)"},{"key":"17747_CR42","doi-asserted-by":"crossref","unstructured":"Rahman MA, Wang Y (2016) Optimizing intersection-over-union in deep neural networks for image segmentation. In: Proceeding of the international symposium on visual computing (ISVC)","DOI":"10.1007\/978-3-319-50835-1_22"},{"key":"17747_CR43","doi-asserted-by":"crossref","unstructured":"Wei J, Wang S, Huang Q (2020) F3Net: fusion, feedback and focus for salient object detection. In: Proceeding of the AAAI conference on artificial intelligence (AAAI)","DOI":"10.1609\/aaai.v34i07.6916"},{"key":"17747_CR44","doi-asserted-by":"crossref","unstructured":"Perazzi F, Krahenbuhl P, Pritch Y, Hornung A (2012) Saliency filters: contrast based filtering for salient region detection. In: Proceeding of the IEEE conference on computer vision and pattern recognition (CVPR), pp 733\u2013740","DOI":"10.1109\/CVPR.2012.6247743"},{"key":"17747_CR45","doi-asserted-by":"crossref","unstructured":"Achanta R, Hemami S, Estrada F, Susstrunk S (2009) Frequency-tuned salient region detection. In: Proceeding of the IEEE conference on computer vision and pattern recognition (CVPR), pp 1597\u20131604","DOI":"10.1109\/CVPR.2009.5206596"},{"key":"17747_CR46","doi-asserted-by":"crossref","unstructured":"Fan D-P, Cheng M-M, Liu Y, Li T, Borji A (2017) Structure-measure: a new way to evaluate foreground maps. In: Proceeding of the IEEE international conference on computer vision (ICCV), pp 4558\u20134567","DOI":"10.1109\/ICCV.2017.487"},{"key":"17747_CR47","doi-asserted-by":"crossref","unstructured":"Fan D, Gong C, Cao Y, Ren B, Cheng M, Borji A (2018) Enhanced-alignment measure for binary foreground map evaluation. In: Proceeding of the joint conference on artificial intelligence (IJCAI), pp 698\u2013704","DOI":"10.24963\/ijcai.2018\/97"},{"key":"17747_CR48","doi-asserted-by":"crossref","unstructured":"Piao Y, Ji W, Li J, Zhang M, Lu H (2019) Depth-induced multi-scale recurrent attention network for saliency detection. In: Proceeding of the IEEE international conference on computer vision (ICCV), pp 7253\u20137262","DOI":"10.1109\/ICCV.2019.00735"},{"key":"17747_CR49","doi-asserted-by":"crossref","unstructured":"Liu N, Zhang N, Han J (2020) Learning selective self-mutual attention for RGB-D saliency detection. In: Proceeding of the IEEE conference on computer vision and pattern recognition (CVPR)","DOI":"10.1109\/CVPR42600.2020.01377"},{"key":"17747_CR50","doi-asserted-by":"crossref","unstructured":"Deng Z, Hu X, Zhu L, Xu X, Qin J, Han G (2018) R3Net: recurrent residual refinement network for saliency detection. In: Proceeding of the international joint conference on artificial intelligence (IJCAI), pp 684\u2013690","DOI":"10.24963\/ijcai.2018\/95"},{"key":"17747_CR51","doi-asserted-by":"crossref","unstructured":"Qin X, Zhang Z, Huang C, Gao C, Dehghan M, Jagersand M (2019) BASNet: boundary-aware salient object detection. In: Proceeding of the IEEE conference on computer vision and pattern recognition (CVPR)","DOI":"10.1109\/CVPR.2019.00766"},{"key":"17747_CR52","doi-asserted-by":"crossref","unstructured":"Liu J-J, Hou Q, Cheng M-M, Feng J, Jiang J (2019) A simple pooling-based design for real-time salient object detection. In: Proceeding of the IEEE conference on computer vision and pattern recognition (CVPR)","DOI":"10.1109\/CVPR.2019.00404"},{"key":"17747_CR53","doi-asserted-by":"crossref","unstructured":"Wu Z, Su L, Huang Q (2019) Cascaded partial decoder for fast and accurate salient object detection. In: Proceeding of the IEEE conference on computer vision and pattern recognition (CVPR), pp 3902\u20133911","DOI":"10.1109\/CVPR.2019.00403"},{"key":"17747_CR54","doi-asserted-by":"crossref","unstructured":"Zhao J-X, Liu J-J, Fan D-P, Cao Y, Yang J, Cheng M-M (2019) EGNet:edge guidance network for salient object detection. In: Proceeding of the IEEE international conference on computer vision (ICCV)","DOI":"10.1109\/ICCV.2019.00887"},{"key":"17747_CR55","doi-asserted-by":"publisher","first-page":"2593","DOI":"10.1109\/TIP.2023.3270801","volume":"32","author":"H Zhou","year":"2023","unstructured":"Zhou H, Tian C, Zhang Z, Li C, Ding Y, Xie Y, Li Z (2023) Position-aware relation learning for rgb-thermal salient object detection. IEEE Trans Image Process 32:2593\u20132607","journal-title":"IEEE Trans Image Process"},{"issue":"5","key":"17747_CR56","doi-asserted-by":"publisher","first-page":"3111","DOI":"10.1109\/TCSVT.2021.3102268","volume":"32","author":"F Huo","year":"2022","unstructured":"Huo F, Zhu X, Zhang L, Liu Q, Shu Y (2022) Efficient context-guided stacked refinement network for RGB-T salient object detection. IEEE Trans Circuits Syst Video Technol 32(5):3111\u20133124","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"issue":"4","key":"17747_CR57","doi-asserted-by":"publisher","first-page":"2091","DOI":"10.1109\/TCSVT.2021.3082939","volume":"32","author":"W Gao","year":"2022","unstructured":"Gao W, Liao G, Ma S, Li G, Liang Y, Lin W (2022) Unified information fusion network for multi-modal RGB-D and RGB-T salient object detection. IEEE Trans Circuits Syst Video Technol 32(4):2091\u20132106","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"17747_CR58","doi-asserted-by":"crossref","unstructured":"Zhou T, Fan D-P, Cheng M-M, Shen J, Shao L (2021) RGB-D salient object detection: a survey. Comput Vis Media 7(4)","DOI":"10.1007\/s41095-020-0199-z"},{"issue":"9","key":"17747_CR59","doi-asserted-by":"publisher","first-page":"6308","DOI":"10.1109\/TCSVT.2022.3166914","volume":"32","author":"G Chen","year":"2022","unstructured":"Chen G, Shao F, Chai X, Chen H, Jiang Q, Meng X, Ho Y-S (2022) CGMDRNet: cross-guided modality difference reduction network for RGB-T salient object detection. IEEE Trans Circuits Syst Video Technol 32(9):6308\u20136323","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"issue":"7","key":"17747_CR60","doi-asserted-by":"publisher","first-page":"4486","DOI":"10.1109\/TCSVT.2021.3127149","volume":"32","author":"Z Liu","year":"2022","unstructured":"Liu Z, Tan Y, He Q, Xiao Y (2022) Swinnet: swin transformer drives edge-aware RGB-D and RGB-T salient object detection. IEEE Trans Circuits Syst Video Technol 32(7):4486\u20134497","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"17747_CR61","doi-asserted-by":"publisher","first-page":"892","DOI":"10.1109\/TIP.2023.3234702","volume":"32","author":"Y Pang","year":"2023","unstructured":"Pang Y, Zhao X, Zhang L, Lu H (2023) Caver: cross-modal view-mixed transformer for bi-modal salient object detection. IEEE Trans Image Process 32:892\u2013904","journal-title":"IEEE Trans Image Process"},{"key":"17747_CR62","doi-asserted-by":"publisher","first-page":"115","DOI":"10.1016\/j.image.2015.07.002","volume":"38","author":"R Ju","year":"2015","unstructured":"Ju R, Liu Y, Ren T, Ge L, Wu G (2015) Depth-aware salient object detection using anisotropic center-surround difference. Signal Process Image Commun 38:115\u2013126","journal-title":"Signal Process Image Commun"},{"key":"17747_CR63","doi-asserted-by":"crossref","unstructured":"Peng H, Li B, Xiong W, Hu W, Ji R (2014) RGBD salient object detection: a benchmark and algorithms. In: Proceeding of the Europeon conference on computer vision (ECCV)","DOI":"10.1007\/978-3-319-10578-9_7"},{"key":"17747_CR64","doi-asserted-by":"crossref","unstructured":"Li G, Zhu C (2017) A three-pathway psychobiological framework of salient object detection using stereoscopic technology. In: Proceeding of the IEEE international conference on computer vision workshops (ICCVW), pp 3008\u20133014","DOI":"10.1109\/ICCVW.2017.355"},{"key":"17747_CR65","unstructured":"Niu Y, Geng Y, Li X, Liu F (2012) Leveraging stereopsis for saliency analysis. In: Proceeding of the IEEE conference on computer vision and pattern recognition, pp 454\u2013461"},{"key":"17747_CR66","doi-asserted-by":"crossref","unstructured":"Ji W, Li J, Zhang M, Piao Y, Lu H (2020) Accurate rgb-d salient object detection via collaborative learning. In: Proceeding of the Europeon conference on computer vision (ECCV), pp 52\u201369","DOI":"10.1007\/978-3-030-58523-5_4"},{"key":"17747_CR67","doi-asserted-by":"publisher","first-page":"1343","DOI":"10.1109\/TMM.2020.2997184","volume":"23","author":"B Jiang","year":"2021","unstructured":"Jiang B, Zhou Z, Wang X, Tang J, Luo B (2021) cmsalgan: Rgb-d salient object detection with cross-view generative adversarial networks. IEEE Trans Multimedia 23:1343\u20131353","journal-title":"IEEE Trans Multimedia"},{"key":"17747_CR68","doi-asserted-by":"publisher","unstructured":"Zhu J, Zhang X, Dong F, Yan S, Meng X, Li Y, Tan P (2022) Transformer-based Adaptive Interactive Promotion Network for RGB-T Salient Object Detection. In: 2022 34th Chinese Control and Decision Conference (CCDC), pp. 1989\u20131994. https:\/\/doi.org\/10.1109\/CCDC55256.2022.10034159","DOI":"10.1109\/CCDC55256.2022.10034159"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-17747-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-023-17747-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-17747-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,6]],"date-time":"2024-11-06T09:21:28Z","timestamp":1730884888000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-023-17747-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,19]]},"references-count":68,"journal-issue":{"issue":"20","published-online":{"date-parts":[[2024,6]]}},"alternative-id":["17747"],"URL":"https:\/\/doi.org\/10.1007\/s11042-023-17747-y","relation":{},"ISSN":["1573-7721"],"issn-type":[{"type":"electronic","value":"1573-7721"}],"subject":[],"published":{"date-parts":[[2023,12,19]]},"assertion":[{"value":"5 April 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"1 November 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"25 November 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 December 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of Interest"}}]}}