{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,27]],"date-time":"2026-01-27T22:05:07Z","timestamp":1769551507225,"version":"3.49.0"},"reference-count":33,"publisher":"MDPI AG","issue":"3","license":[{"start":{"date-parts":[[2024,3,6]],"date-time":"2024-03-06T00:00:00Z","timestamp":1709683200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"Natural Science Foundation of Hubei Province","award":["2020CFB546"],"award-info":[{"award-number":["2020CFB546"]}]},{"name":"Natural Science Foundation of Hubei Province","award":["12001411"],"award-info":[{"award-number":["12001411"]}]},{"name":"Natural Science Foundation of Hubei Province","award":["12201479"],"award-info":[{"award-number":["12201479"]}]},{"name":"Natural Science Foundation of Hubei Province","award":["2021IVB024"],"award-info":[{"award-number":["2021IVB024"]}]},{"name":"Natural Science Foundation of Hubei Province","award":["2020-IB-003"],"award-info":[{"award-number":["2020-IB-003"]}]},{"name":"National Natural Science Foundation of China","award":["2020CFB546"],"award-info":[{"award-number":["2020CFB546"]}]},{"name":"National Natural Science Foundation of China","award":["12001411"],"award-info":[{"award-number":["12001411"]}]},{"name":"National Natural Science Foundation of China","award":["12201479"],"award-info":[{"award-number":["12201479"]}]},{"name":"National Natural Science Foundation of China","award":["2021IVB024"],"award-info":[{"award-number":["2021IVB024"]}]},{"name":"National Natural Science Foundation of China","award":["2020-IB-003"],"award-info":[{"award-number":["2020-IB-003"]}]},{"name":"Fundamental Research Funds for the Central Universities","award":["2020CFB546"],"award-info":[{"award-number":["2020CFB546"]}]},{"name":"Fundamental Research Funds for the Central Universities","award":["12001411"],"award-info":[{"award-number":["12001411"]}]},{"name":"Fundamental Research Funds for the Central Universities","award":["12201479"],"award-info":[{"award-number":["12201479"]}]},{"name":"Fundamental Research Funds for the Central Universities","award":["2021IVB024"],"award-info":[{"award-number":["2021IVB024"]}]},{"name":"Fundamental Research Funds for the Central Universities","award":["2020-IB-003"],"award-info":[{"award-number":["2020-IB-003"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Algorithms"],"abstract":"<jats:p>Salient object detection (SOD) aims to identify the most visually striking objects in a scene, simulating the function of the biological visual attention system. The attention mechanism in deep learning is commonly used as an enhancement strategy which enables the neural network to concentrate on the relevant parts when processing input data, effectively improving the model\u2019s learning and prediction abilities. Existing saliency object detection methods based on RGB deep learning typically treat all regions equally by using the extracted features, overlooking the fact that different regions have varying contributions to the final predictions. Based on the U2Net algorithm, this paper incorporates the split coordinate channel attention (SCCA) mechanism into the feature extraction stage. SCCA conducts spatial transformation in width and height dimensions to efficiently extract the location information of the target to be detected. While pixel-level semantic segmentation based on annotation has been successful, it assigns the same weight to each pixel which leads to poor performance in detecting the boundary of objects. In this paper, the Canny edge detection loss is incorporated into the loss calculation stage to improve the model\u2019s ability to detect object edges. Based on the DUTS and HKU-IS datasets, experiments confirm that the proposed strategies effectively enhance the model\u2019s detection performance, resulting in a 0.8% and 0.7% increase in the F1-score of U2Net. This paper also compares the traditional attention modules with the newly proposed attention, and the SCCA attention module achieves a top-three performance in prediction time, mean absolute error (MAE), F1-score, and model size on both experimental datasets.<\/jats:p>","DOI":"10.3390\/a17030109","type":"journal-article","created":{"date-parts":[[2024,3,6]],"date-time":"2024-03-06T08:43:14Z","timestamp":1709714594000},"page":"109","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Application of Split Coordinate Channel Attention Embedding U2Net in Salient Object Detection"],"prefix":"10.3390","volume":"17","author":[{"given":"Yuhuan","family":"Wu","sequence":"first","affiliation":[{"name":"School of Science, Wuhan University of Technology, Wuhan 430074, China"}]},{"given":"Yonghong","family":"Wu","sequence":"additional","affiliation":[{"name":"School of Science, Wuhan University of Technology, Wuhan 430074, China"}]}],"member":"1968","published-online":{"date-parts":[[2024,3,6]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"24","DOI":"10.1016\/j.gltp.2021.01.004","article-title":"Machine learning and deep learning applications\u2014A vision","volume":"2","author":"Sharma","year":"2021","journal-title":"Glob. Transit. Proc."},{"key":"ref_2","doi-asserted-by":"crossref","unstructured":"Shinde, P.P., and Shah, S. (2018, January 16\u201318). A review of machine learning and deep learning applications. Proceedings of the 2018 Fourth International Conference on Computing Communication Control and Automation (ICCUBEA), Pune, India.","DOI":"10.1109\/ICCUBEA.2018.8697857"},{"key":"ref_3","first-page":"100134","article-title":"Deep learning in computer vision: A critical review of emerging techniques and application scenarios","volume":"6","author":"Chai","year":"2021","journal-title":"Mach. Learn. Appl."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"3383","DOI":"10.1007\/s11831-020-09504-3","article-title":"Computer vision techniques in construction: A critical review","volume":"28","author":"Xu","year":"2021","journal-title":"Arch. Comput. Methods Eng."},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"7068349","DOI":"10.1155\/2018\/7068349","article-title":"Deep learning for computer vision: A brief review","volume":"2018","author":"Voulodimos","year":"2018","journal-title":"Comput. Intell. Neurosci."},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"48","DOI":"10.1016\/j.neucom.2021.03.091","article-title":"A review on the attention mechanism of deep learning","volume":"452","author":"Niu","year":"2021","journal-title":"Neurocomputing"},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"3279","DOI":"10.1109\/TKDE.2021.3126456","article-title":"A general survey on attention mechanisms in deep learning","volume":"35","author":"Brauwers","year":"2021","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"331","DOI":"10.1007\/s41095-022-0271-y","article-title":"Attention mechanisms in computer vision: A survey","volume":"8","author":"Guo","year":"2022","journal-title":"Comput. Vis. Media"},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"4121","DOI":"10.1109\/JSTARS.2020.3009352","article-title":"Channel-attention-based DenseNet network for remote sensing image scene classification","volume":"13","author":"Tong","year":"2020","journal-title":"IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"37","DOI":"10.1007\/s41095-020-0199-z","article-title":"RGB-D salient object detection: A survey","volume":"7","author":"Zhou","year":"2021","journal-title":"Comput. Vis. Media"},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"835","DOI":"10.1016\/j.ins.2020.09.003","article-title":"CNN-based encoder-decoder networks for salient object detection: A comprehensive review and recent advances","volume":"546","author":"Ji","year":"2021","journal-title":"Inf. Sci."},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Wang, L., Lu, H., Wang, Y., Feng, M., Wang, D., Yin, B., and Ruan, X. (2017, January 21). Learning to detect salient objects with image-level supervision. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.404"},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Rong, W., Li, Z., Zhang, W., and Sun, L. (2014, January 3\u20136). An improved CANNY edge detection algorithm. Proceedings of the 2014 IEEE International Conference on Mechatronics and Automation, Tianjin, China.","DOI":"10.1109\/ICMA.2014.6885761"},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"107404","DOI":"10.1016\/j.patcog.2020.107404","article-title":"U2-Net: Going deeper with nested U-structure for salient object detection","volume":"106","author":"Qin","year":"2020","journal-title":"Pattern Recognit."},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Wang, W., Zhao, S., Shen, J., Hoi, S.C., and Borji, A. (2019, January 15\u201320). Salient object detection with pyramid attention and salient edges. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.00154"},{"key":"ref_16","unstructured":"Wei, J., Wang, S., and Huang, Q. (2020, January 7\u201312). F\u00b3Net: Fusion, feedback and focus for salient object detection. Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA."},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Qin, X., Zhang, Z., Huang, C., Gao, C., Dehghan, M., and Jagersand, M. (2019, January 15\u201320). Basnet: Boundary-aware salient object detection. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.00766"},{"key":"ref_18","doi-asserted-by":"crossref","unstructured":"Zhao, X., Pang, Y., Zhang, L., Lu, H., and Zhang, L. (2020, January 23\u201328). Suppress and balance: A simple gated network for salient object detection. Proceedings of the Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK. Proceedings, Part II 16.","DOI":"10.1007\/978-3-030-58536-5_3"},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"7547","DOI":"10.1007\/s00521-021-06845-3","article-title":"CFIDNet: Cascaded feature interaction decoder for RGB-D salient object detection","volume":"34","author":"Chen","year":"2022","journal-title":"Neural Comput. Appl."},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"7646","DOI":"10.1109\/TCSVT.2022.3184840","article-title":"Cross-collaborative fusion-encoder network for robust RGB-thermal salient object detection","volume":"32","author":"Liao","year":"2022","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"5624915","DOI":"10.1109\/TGRS.2022.3224815","article-title":"Hybrid feature aligned network for salient object detection in optical remote sensing imagery","volume":"60","author":"Wang","year":"2022","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"539","DOI":"10.1109\/TCYB.2022.3163152","article-title":"Edge-guided recurrent positioning network for salient object detection in optical remote sensing images","volume":"53","author":"Zhou","year":"2022","journal-title":"IEEE Trans. Cybern."},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"3125","DOI":"10.1109\/TIP.2022.3164550","article-title":"EDN: Salient object detection via extremely-downsampled network","volume":"31","author":"Wu","year":"2022","journal-title":"IEEE Trans. Image Process."},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"2676","DOI":"10.1109\/TNNLS.2020.3007534","article-title":"CASNet: A cross-attention siamese network for video salient object detection","volume":"32","author":"Ji","year":"2020","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"ref_25","unstructured":"Canny, J.F. (1983). Finding Edges and Lines in Images, Artificial lntelligence Laboratory. Theory of Computing Systems\/Mathematical Systems Theory."},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Hou, Q., Zhou, D., and Feng, J. (2021, January 19\u201325). Coordinate attention for efficient mobile network design. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA.","DOI":"10.1109\/CVPR46437.2021.01350"},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Woo, S., Park, J., Lee, J.Y., and Kweon, I.S. (2018, January 8\u201314). Cbam: Convolutional block attention module. Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany.","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"ref_28","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., and Sun, G. (2018, January 18\u201323). Squeeze-and-excitation networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00745"},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Misra, D., Nalamada, T., Arasanipalai, A.U., and Hou, Q. (2021, January 5\u20139). Rotate to attend: Convolutional triplet attention module. Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, Virtual.","DOI":"10.1109\/WACV48630.2021.00318"},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Yang, Y., and Deng, H. (2020). GC-YOLOv3: You only look once with global context block. Electronics, 9.","DOI":"10.3390\/electronics9081235"},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Li, X., Wang, W., Hu, X., and Yang, J. (2019, January 15\u201320). Selective kernel networks. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.00060"},{"key":"ref_32","doi-asserted-by":"crossref","unstructured":"Wang, X., Girshick, R., Gupta, A., and He, K. (2018, January 18\u201323). Non-local neural networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00813"},{"key":"ref_33","doi-asserted-by":"crossref","unstructured":"Zhu, Z., Xu, M., Bai, S., Huang, T., and Bai, X. (2019, January 27\u201328). Asymmetric non-local neural networks for semantic segmentation. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Seoul, Republic of Korea.","DOI":"10.1109\/ICCV.2019.00068"}],"container-title":["Algorithms"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1999-4893\/17\/3\/109\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T14:09:55Z","timestamp":1760105395000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1999-4893\/17\/3\/109"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,3,6]]},"references-count":33,"journal-issue":{"issue":"3","published-online":{"date-parts":[[2024,3]]}},"alternative-id":["a17030109"],"URL":"https:\/\/doi.org\/10.3390\/a17030109","relation":{},"ISSN":["1999-4893"],"issn-type":[{"value":"1999-4893","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,3,6]]}}}