{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,7]],"date-time":"2026-02-07T10:36:01Z","timestamp":1770460561635,"version":"3.49.0"},"reference-count":55,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,11,26]],"date-time":"2025-11-26T00:00:00Z","timestamp":1764115200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,11,26]],"date-time":"2025-11-26T00:00:00Z","timestamp":1764115200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100012226","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"crossref","award":["3132025113"],"award-info":[{"award-number":["3132025113"]}],"id":[{"id":"10.13039\/501100012226","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Basic Scientific Research Project of the Educational Department of Liaoning Province","award":["LJ212410151006"],"award-info":[{"award-number":["LJ212410151006"]}]},{"name":"Open Fund of Hubei Key Laboratory of Intelligent Robot","award":["HBIR202411"],"award-info":[{"award-number":["HBIR202411"]}]},{"name":"Open Fund of National Center for International Research of Subsea Engineering Technology and Equipment","award":["HG20240202"],"award-info":[{"award-number":["HG20240202"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["52501429"],"award-info":[{"award-number":["52501429"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["52371373"],"award-info":[{"award-number":["52371373"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100005047","name":"Natural Science Foundation of Liaoning Province","doi-asserted-by":"publisher","award":["2025-BS-0210"],"award-info":[{"award-number":["2025-BS-0210"]}],"id":[{"id":"10.13039\/501100005047","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Real-Time Image Proc"],"published-print":{"date-parts":[[2026,2]]},"DOI":"10.1007\/s11554-025-01814-8","type":"journal-article","created":{"date-parts":[[2025,11,26]],"date-time":"2025-11-26T09:59:30Z","timestamp":1764151170000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Multi-stage differential-aware attention network for real-time underwater salient object detection"],"prefix":"10.1007","volume":"23","author":[{"given":"Xun","family":"Ji","sequence":"first","affiliation":[]},{"given":"Long","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Xiaolin","family":"Li","sequence":"additional","affiliation":[]},{"given":"Wei","family":"Liu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,11,26]]},"reference":[{"key":"1814_CR1","doi-asserted-by":"publisher","first-page":"473","DOI":"10.1016\/j.inffus.2022.08.022","volume":"89","author":"D Zhu","year":"2023","unstructured":"Zhu, D., Wang, X., Li, G., Zhang, X.-P.: Vessel detection via multi-order saliency-based fuzzy fusion of spaceborne and airborne sar images. Inf. Fusion 89, 473\u2013485 (2023)","journal-title":"Inf. Fusion"},{"key":"1814_CR2","doi-asserted-by":"publisher","first-page":"108414","DOI":"10.1016\/j.patcog.2021.108414","volume":"123","author":"G-P Ji","year":"2022","unstructured":"Ji, G.-P., Zhu, L., Zhuge, M., Fu, K.: Fast camouflaged object detection via edge-based reversible re-calibration network. Pattern Recognit. 123, 108414 (2022)","journal-title":"Pattern Recognit."},{"key":"1814_CR3","doi-asserted-by":"publisher","first-page":"108782","DOI":"10.1016\/j.engappai.2024.108782","volume":"135","author":"X Ji","year":"2024","unstructured":"Ji, X., Tang, L., Chen, L., Hao, L.-Y., Guo, H.: Toward efficient and lightweight sea-land segmentation for remote sensing images. Eng. Appl. Artif. Intell. 135, 108782 (2024)","journal-title":"Eng. Appl. Artif. Intell."},{"key":"1814_CR4","doi-asserted-by":"crossref","unstructured":"Yang, J., Gao, S., Li, Z., Zheng, F., Leonardis, A.: Resource-efficient RGB-D aerial tracking. In: 2023 IEEE\/CVF conference on computer vision and pattern recognition (CVPR)","DOI":"10.1109\/CVPR52729.2023.01285"},{"issue":"10","key":"1814_CR5","doi-asserted-by":"publisher","first-page":"1723","DOI":"10.1587\/transinf.2023EDP7055","volume":"E106.D","author":"T Yamana","year":"2023","unstructured":"Yamana, T., Suzuki, T., Nobustsune, T., Wu, C.: Multi-scale estimation for omni-directional saliency maps using learnable equator bias. IEICE Trans. Inf. Syst. E106.D(10), 1723\u20131731 (2023)","journal-title":"IEICE Trans. Inf. Syst."},{"issue":"12","key":"1814_CR6","doi-asserted-by":"publisher","first-page":"1723","DOI":"10.3390\/atmos14121723","volume":"14","author":"K Wang","year":"2023","unstructured":"Wang, K., Huang, Y., Yu, T., Chen, Y., Li, Z., Kuang, Q.: MAFormer: a new method for radar reflectivity reconstructing using satellite data. Atmosphere 14(12), 1723 (2023)","journal-title":"Atmosphere"},{"key":"1814_CR7","doi-asserted-by":"crossref","unstructured":"Fu, H., Xu, D., Lin, S., Liu, J.: Object-based RGB-D image co-segmentation with mutex constraint. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), pp. 4428\u20134436, (2015)","DOI":"10.1109\/CVPR.2015.7299072"},{"issue":"6","key":"1814_CR8","doi-asserted-by":"publisher","first-page":"3239","DOI":"10.1109\/TPAMI.2021.3051099","volume":"44","author":"W Wang","year":"2022","unstructured":"Wang, W., Lai, Q., Fu, H., Shen, J., Ling, H., Yang, R.: Salient object detection in the deep learning era: an in-depth survey. IEEE Trans. Pattern Anal. Mach. Intell. 44(6), 3239\u20133259 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1814_CR9","doi-asserted-by":"publisher","first-page":"1602","DOI":"10.1109\/TIP.2023.3266163","volume":"34","author":"L Hong","year":"2025","unstructured":"Hong, L., Wang, X., Zhang, G., Zhao, M.: USOD10K: A new benchmark dataset for underwater salient object detection. IEEE Trans. Image Process. 34, 1602\u20131615 (2025)","journal-title":"IEEE Trans. Image Process."},{"issue":"9","key":"1814_CR10","doi-asserted-by":"publisher","first-page":"1757","DOI":"10.3390\/jmse11091757","volume":"11","author":"W Huang","year":"2023","unstructured":"Huang, W., Zhu, D., Chen, M.: A fusion underwater salient object detection based on multi-scale saliency and spatial optimization. J. Mar. Sci. Eng. 11(9), 1757 (2023)","journal-title":"J. Mar. Sci. Eng."},{"key":"1814_CR11","doi-asserted-by":"publisher","first-page":"102806","DOI":"10.1016\/j.inffus.2024.102806","volume":"117","author":"G Yuan","year":"2025","unstructured":"Yuan, G., Song, J., Li, J.: IF-USOD: Multimodal information fusion interactive feature enhancement architecture for underwater salient object detection. Inf. Fusion 117, 102806 (2025)","journal-title":"Inf. Fusion"},{"key":"1814_CR12","doi-asserted-by":"publisher","DOI":"10.1109\/JSEN.2025.3606908","author":"X Ji","year":"2025","unstructured":"Ji, X., Dai, C., Hao, L.-Y., Cai, C., Liu, S.: PAC-Net: Physics-aware constraint network for underwater image enhancement. IEEE Sens J (2025). https:\/\/doi.org\/10.1109\/JSEN.2025.3606908","journal-title":"IEEE Sens J"},{"issue":"3","key":"1814_CR13","doi-asserted-by":"publisher","first-page":"2147","DOI":"10.1109\/TCSVT.2024.3491907","volume":"35","author":"J Jin","year":"2025","unstructured":"Jin, J., Jiang, Q., Wu, Q., Xu, B., Cong, R.: Underwater salient object detection via dual-stage self-paced learning and depth emphasis. IEEE Trans. Circuits Syst. Video Technol. 35(3), 2147\u20132160 (2025)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"1814_CR14","doi-asserted-by":"publisher","DOI":"10.1109\/TBC.2025.3611669","author":"X Ji","year":"2025","unstructured":"Ji, X., Wang, X., Hao, L.-Y., Cai, C., Dai, C., Liu, R.W.: Fusing learning and non-learning: hybrid CNN-Transformer cooperative-competitive network for underwater image enhancement. IEEE Trans. Broadcast. (2025). https:\/\/doi.org\/10.1109\/TBC.2025.3611669","journal-title":"IEEE Trans. Broadcast."},{"issue":"11","key":"1814_CR15","doi-asserted-by":"publisher","first-page":"6404","DOI":"10.1109\/TCSVT.2023.3269951","volume":"33","author":"Z Wang","year":"2023","unstructured":"Wang, Z., Zhang, Y., Liu, Y., Zhu, D., Coleman, S.A., Kerr, D.: ELWNet: An extremely lightweight approach for real-time salient object detection. IEEE Trans. Circuits Syst. Video Technol. 33(11), 6404\u20136417 (2023)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"1814_CR16","doi-asserted-by":"publisher","first-page":"3125","DOI":"10.1109\/TIP.2022.3164550","volume":"31","author":"Y-H Wu","year":"2022","unstructured":"Wu, Y.-H., Liu, Y., Zhang, L., Cheng, M.-M., Ren, B.: EDN: Salient object detection via extremely-downsampled network. IEEE Trans. Image Process. 31, 3125\u20133136 (2022)","journal-title":"IEEE Trans. Image Process."},{"key":"1814_CR17","doi-asserted-by":"crossref","unstructured":"Zhang, P., Wang, D., Lu, H., Wang, H., Ruan, X.: Amulet: Aggregating multi-level convolutional features for salient object detection. In: 2017 IEEE international conference on computer vision (ICCV), pp.\u00a0202\u2013211, (2017)","DOI":"10.1109\/ICCV.2017.31"},{"key":"1814_CR18","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2020.107303","volume":"103","author":"S Mohammadi","year":"2020","unstructured":"Mohammadi, S., Noori, M., Bahri, A., Majelan, S.G., Havaei, M.: CAGNet: Content-aware guidance for salient object detection. Pattern Recognit. 103, 107303 (2020)","journal-title":"Pattern Recognit."},{"key":"1814_CR19","doi-asserted-by":"publisher","first-page":"10599","DOI":"10.1609\/aaai.v34i07.6633","volume":"34","author":"Z Chen","year":"2020","unstructured":"Chen, Z., Xu, Q., Cong, R., Huang, Q.: Global context-aware progressive aggregation network for salient object detection. AAAI 34, 10599\u201310606 (2020)","journal-title":"AAAI"},{"key":"1814_CR20","doi-asserted-by":"publisher","first-page":"3763","DOI":"10.1109\/TIP.2020.2965989","volume":"29","author":"S Chen","year":"2020","unstructured":"Chen, S., Tan, X., Wang, B., Lu, H., Hu, X., Fu, Y.: Reverse attention-based residual network for salient object detection. IEEE Trans. Image Process. 29, 3763\u20133776 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"1814_CR21","doi-asserted-by":"crossref","unstructured":"Pang, Y., Zhao, X., Zhang, L., Lu, H.: Multi-scale interactive network for salient object detection. In: IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp.\u00a09410\u20139419, (2020)","DOI":"10.1109\/CVPR42600.2020.00943"},{"issue":"8","key":"1814_CR22","doi-asserted-by":"publisher","first-page":"1742","DOI":"10.1109\/TMM.2017.2693022","volume":"19","author":"L Ye","year":"2017","unstructured":"Ye, L., Liu, Z., Li, L., Shen, L., Bai, C., Wang, Y.: Salient object segmentation via effective integration of saliency and objectness. IEEE Trans. Multimedia 19(8), 1742\u20131756 (2017)","journal-title":"IEEE Trans. Multimedia"},{"issue":"1","key":"1814_CR23","doi-asserted-by":"publisher","first-page":"74","DOI":"10.1109\/TMM.2018.2845667","volume":"21","author":"Y Zhou","year":"2019","unstructured":"Zhou, Y., Mao, A., Huo, S., Lei, J., Kung, S.-Y.: Salient object detection via fuzzy theory and object-level enhancement. IEEE Trans. Multimedia 21(1), 74\u201385 (2019)","journal-title":"IEEE Trans. Multimedia"},{"issue":"9","key":"1814_CR24","doi-asserted-by":"publisher","first-page":"4439","DOI":"10.1109\/TCYB.2020.3035613","volume":"51","author":"Y Liu","year":"2021","unstructured":"Liu, Y., Gu, Y.-C., Zhang, X.-Y., Wang, W., Cheng, M.-M.: Lightweight salient object detection via hierarchical visual perception learning. IEEE T. Cybern. 51(9), 4439\u20134449 (2021)","journal-title":"IEEE T. Cybern."},{"issue":"2","key":"1814_CR25","doi-asserted-by":"publisher","first-page":"728","DOI":"10.1109\/TCSVT.2020.2988768","volume":"31","author":"L Wang","year":"2021","unstructured":"Wang, L., Chen, R., Zhu, L., Xie, H., Li, X.: Deep sub-region network for salient object detection. IEEE Trans. Circuits Syst. Video Technol. 31(2), 728\u2013741 (2021)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"1814_CR26","doi-asserted-by":"crossref","unstructured":"Luo, Z., Mishra, A., Achkar, A., Eichel, J., Li, S., Jodoin, P.-M.: Non-local deep features for salient object detection. In: IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp.\u00a06593\u20136601, (2017)","DOI":"10.1109\/CVPR.2017.698"},{"key":"1814_CR27","doi-asserted-by":"crossref","unstructured":"Liu, N., Han, J., Yang, M.-H.: PiCANet: Learning pixel-wise contextual attention for saliency detection. In: IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp.\u00a03089\u20133098, (2018)","DOI":"10.1109\/CVPR.2018.00326"},{"key":"1814_CR28","doi-asserted-by":"crossref","unstructured":"Zhang, P., Wang, D., Lu, H., Wang, H., Yin, B.: Learning uncertain convolutional features for accurate saliency detection. In: IEEE international conference on computer vision (ICCV), pp.\u00a0212\u2013221, (2017)","DOI":"10.1109\/ICCV.2017.32"},{"key":"1814_CR29","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2024.3502424","author":"Z Jin","year":"2024","unstructured":"Jin, Z., Wang, C., Luo, X.: Colorization-inspired customized low-light image enhancement by a decoupled network. IEEE Trans. Neural. Netw. Learn Syst. (2024). https:\/\/doi.org\/10.1109\/TNNLS.2024.3502424","journal-title":"IEEE Trans. Neural. Netw. Learn Syst."},{"issue":"3","key":"1814_CR30","first-page":"1","volume":"21","author":"S Wang","year":"2025","unstructured":"Wang, S., Wang, Y., Qiao, R., Wu, K., Lin, C.-W., Cai, C.: Multi-scale dynamic fusion for visible-infrared person re-identification. ACM Trans. Multimedia Comput. Commun. Appl. 21(3), 1\u201324 (2025)","journal-title":"ACM Trans. Multimedia Comput. Commun. Appl."},{"issue":"3","key":"1814_CR31","doi-asserted-by":"publisher","first-page":"131","DOI":"10.1007\/s11554-025-01715-w","volume":"22","author":"Z Wang","year":"2025","unstructured":"Wang, Z., Zhang, Y., Zhang, S.: Real-time personal protective equipment detection and classification with YOLOv8 multi-scale fusion. J. Real-Time Image Proc. 22(3), 131 (2025)","journal-title":"J. Real-Time Image Proc."},{"issue":"7","key":"1814_CR32","doi-asserted-by":"publisher","first-page":"12321","DOI":"10.1609\/aaai.v34i07.6916","volume":"34","author":"J Wei","year":"2020","unstructured":"Wei, J., Wang, S., Huang, Q.: F3Net: Fusion, feedback and focus for salient object detection. AAAI 34(7), 12321\u201312328 (2020)","journal-title":"AAAI"},{"key":"1814_CR33","doi-asserted-by":"crossref","unstructured":"Wu, Z., Su, L., Huang, Q.: Stacked cross refinement network for edge-aware salient object detection. In: IEEE\/CVF international conference on computer vision (ICCV), pp.\u00a07263\u20137272, (2019)","DOI":"10.1109\/ICCV.2019.00736"},{"key":"1814_CR34","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TGRS.2021.3123984","volume":"60","author":"R Cong","year":"2022","unstructured":"Cong, R., Zhang, Y., Fang, L., Li, J., Zhao, Y., Kwong, S.: RRNet: Relational reasoning network with parallel multiscale attention for salient object detection in optical remote sensing images. IEEE Trans. Geosci. Remote Sensing 60, 1\u201311 (2022)","journal-title":"IEEE Trans. Geosci. Remote Sensing"},{"key":"1814_CR35","doi-asserted-by":"publisher","first-page":"107404","DOI":"10.1016\/j.patcog.2020.107404","volume":"106","author":"X Qin","year":"2020","unstructured":"Qin, X., Zhang, Z., Huang, C., Dehghan, M., Zaiane, O.R., Jagersand, M.: U2-Net: Going deeper with nested u-structure for salient object detection. Pattern Recognit. 106, 107404 (2020)","journal-title":"Pattern Recognit."},{"key":"1814_CR36","doi-asserted-by":"crossref","unstructured":"Li, L., Han, J., Zhang, N., Liu, N., Khan, S., Cholakkal, H., Anwer, R.M., Khan, F.S: Discriminative co-saliency and background mining transformer for co-salient object detection. In: IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp.\u00a07247\u20137256, (2023)","DOI":"10.1109\/CVPR52729.2023.00700"},{"key":"1814_CR37","first-page":"1","volume":"73","author":"Y Luo","year":"2024","unstructured":"Luo, Y., Shao, F., Xie, Z., Wang, H., Chen, H., Mu, B., Jiang, Q.: Hfmdnet: Hierarchical fusion and multilevel decoder network for rgb-d salient object detection. IEEE Trans. Instrum. Meas. 73, 1\u20135 (2024)","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"2","key":"1814_CR38","doi-asserted-by":"publisher","first-page":"898","DOI":"10.1109\/JOE.2024.3503664","volume":"50","author":"Z Wang","year":"2025","unstructured":"Wang, Z., Yu, M., Cao, L., Liu, P., Wang, L., Li, X., Hong, Y., He, C., Shen, L.: UVTD: A large-scale multilabel data set for underwater vision tasks. IEEE J. Ocean. Eng. 50(2), 898\u2013918 (2025)","journal-title":"IEEE J. Ocean. Eng."},{"issue":"5","key":"1814_CR39","doi-asserted-by":"publisher","first-page":"3795","DOI":"10.1109\/TCSVT.2023.3321190","volume":"34","author":"J Yuan","year":"2024","unstructured":"Yuan, J., Zhu, A., Xu, Q., Wattanachote, K., Gong, Y.: CTIF-Net: A CNN-Transformer iterative fusion network for salient object detection. IEEE Trans. Circuits Syst. Video Technol. 34(5), 3795\u20133805 (2024)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"1814_CR40","doi-asserted-by":"crossref","unstructured":"Ma, N., Zhang, X., Zheng, H., Sun, J.: ShuffleNet V2: Practical guidelines for efficient CNN architecture design. In: ECCV, vol.\u00a011218, (2018)","DOI":"10.1007\/978-3-030-01264-9_8"},{"key":"1814_CR41","doi-asserted-by":"publisher","first-page":"121778","DOI":"10.1016\/j.eswa.2023.121778","volume":"238","author":"B Liang","year":"2024","unstructured":"Liang, B., Luo, H.: MEANet: An effective and lightweight solution for salient object detection in optical remote sensing images. Expert Syst. Appl. 238, 121778 (2024)","journal-title":"Expert Syst. Appl."},{"key":"1814_CR42","doi-asserted-by":"crossref","unstructured":"Gao, S., qiang Tan, Y., Cheng, M.-M., Lu, C., Chen, Y., Yan, S.: Highly efficient salient object detection with 100k parameters. In: ECCV, pp.\u00a0702\u2013721, (2020)","DOI":"10.1007\/978-3-030-58539-6_42"},{"key":"1814_CR43","first-page":"1","volume":"61","author":"G Li","year":"2023","unstructured":"Li, G., Liu, Z., Zhang, X., Lin, W.: Lightweight salient object detection in optical remote-sensing images via semantic matching and edge alignment. IEEE Trans. Geosci. Remote Sensing 61, 1\u201311 (2023)","journal-title":"IEEE Trans. Geosci. Remote Sensing"},{"key":"1814_CR44","doi-asserted-by":"publisher","first-page":"3804","DOI":"10.1109\/TIP.2021.3065239","volume":"30","author":"Y Liu","year":"2021","unstructured":"Liu, Y., Zhang, X.-Y., Bian, J.-W., Zhang, L., Cheng, M.-M.: SAMNet: Stereoscopically attentive multi-scale network for lightweight salient object detection. IEEE Trans. Image Process. 30, 3804\u20133814 (2021)","journal-title":"IEEE Trans. Image Process."},{"key":"1814_CR45","doi-asserted-by":"crossref","unstructured":"Choi, J.I., Tian, Q.: Visual-saliency-guided channel pruning for deep visual detectors in autonomous driving, In: IEEE intelligent vehicles symposium (IV), pp.\u00a01\u20136, (2023)","DOI":"10.1109\/IV55152.2023.10186819"},{"key":"1814_CR46","doi-asserted-by":"publisher","first-page":"10828","DOI":"10.1109\/TMM.2024.3413529","volume":"26","author":"X Zhou","year":"2024","unstructured":"Zhou, X., Shen, K., Liu, Z.: ADMNet: Attention-guided densely multi-scale network for lightweight salient object detection. IEEE Trans. Multimedia 26, 10828\u201310841 (2024)","journal-title":"IEEE Trans. Multimedia"},{"key":"1814_CR47","doi-asserted-by":"publisher","DOI":"10.1109\/TGRS.2025.3558189","author":"P Liu","year":"2025","unstructured":"Liu, P., Bai, H., Xu, T., Wang, J., Chen, H., Li, J.: Hyperspectral remote sensing images salient object detection: the first benchmark dataset and baseline. IEEE Trans. Geosci. Remote Sensing (2025). https:\/\/doi.org\/10.1109\/TGRS.2025.3558189","journal-title":"IEEE Trans. Geosci. Remote Sensing"},{"key":"1814_CR48","doi-asserted-by":"crossref","unstructured":"Wu, Q., Fu, Z., Lin, H., Ma, C., Tu, X., Ding, X.: EffiSeaNet: Pioneering lightweight network for underwater salient object detection. In: Computer vision-ACCV 2024, PT IV, vol.\u00a015475, pp.\u00a089\u2013104, (2025)","DOI":"10.1007\/978-981-96-0911-6_6"},{"key":"1814_CR49","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.-C.: MobileNetV2: Inverted residuals and linear bottlenecks. In: IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp.\u00a04510\u20134520, (2018)","DOI":"10.1109\/CVPR.2018.00474"},{"key":"1814_CR50","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp.\u00a07132\u20137141, (2018)","DOI":"10.1109\/CVPR.2018.00745"},{"key":"1814_CR51","doi-asserted-by":"crossref","unstructured":"Islam, M.J., Edge, C., Xiao, Y., Luo, P., Mehtaz, M., Morse, C., Enan, S.S., Sattar, J.: Semantic segmentation of underwater imagery: Dataset and benchmark. In: IEEE\/RSJ international conference on intelligent robots and systems (IROS), pp.\u00a01769\u20131776, (2020)","DOI":"10.1109\/IROS45743.2020.9340821"},{"key":"1814_CR52","doi-asserted-by":"crossref","unstructured":"Achanta, R., Hemami, S., Estrada, F., Susstrunk, S.: Frequency-tuned salient region detection. In: IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp.\u00a01597\u20131604, (2009)","DOI":"10.1109\/CVPR.2009.5206596"},{"key":"1814_CR53","doi-asserted-by":"publisher","first-page":"2622","DOI":"10.1007\/s11263-021-01490-8","volume":"129","author":"M Cheng","year":"2021","unstructured":"Cheng, M., Fan, D.: Structure-measure: a new way to evaluate foreground maps. Int. J. Comput. Vis. 129, 2622\u20132638 (2021)","journal-title":"Int. J. Comput. Vis."},{"key":"1814_CR54","doi-asserted-by":"crossref","unstructured":"Perazzi, F., Kr\u00e4henb\u00fchl, P., Pritch, Y., Hornung, A.: Saliency filters: contrast based filtering for salient region detection. In: IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp.\u00a0733\u2013740, (2012)","DOI":"10.1109\/CVPR.2012.6247743"},{"key":"1814_CR55","doi-asserted-by":"crossref","unstructured":"Fan, D.-P., Gong, C., Cao, Y., Ren, B., Cheng, M.-M., Borji, A.: Enhanced-alignment measure for binary foreground map evaluation. In: Proceedings of the twenty-seventh international joint conference on artificial intelligence, IJCAI-18, pp.\u00a0698\u2013704, (2018)","DOI":"10.24963\/ijcai.2018\/97"}],"container-title":["Journal of Real-Time Image Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11554-025-01814-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11554-025-01814-8","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11554-025-01814-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,6]],"date-time":"2026-02-06T16:52:07Z","timestamp":1770396727000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11554-025-01814-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,26]]},"references-count":55,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,2]]}},"alternative-id":["1814"],"URL":"https:\/\/doi.org\/10.1007\/s11554-025-01814-8","relation":{},"ISSN":["1861-8200","1861-8219"],"issn-type":[{"value":"1861-8200","type":"print"},{"value":"1861-8219","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,11,26]]},"assertion":[{"value":"20 July 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 November 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 November 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interest or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"18"}}