{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,2]],"date-time":"2026-04-02T03:00:02Z","timestamp":1775098802153,"version":"3.50.1"},"reference-count":35,"publisher":"MDPI AG","issue":"2","license":[{"start":{"date-parts":[[2026,2,13]],"date-time":"2026-02-13T00:00:00Z","timestamp":1770940800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"Innovation Program of the Chinese Academy of Agricultural Sciences","award":["CAAS-CAE-202302"],"award-info":[{"award-number":["CAAS-CAE-202302"]}]},{"name":"Innovation Program of the Chinese Academy of Agricultural Sciences","award":["CAAS-CAE-202301"],"award-info":[{"award-number":["CAAS-CAE-202301"]}]},{"name":"Open Research Project of the Key Laboratory of Applied Research on Tropical Crop Information Technology of Hainan Province","award":["ZDSYS-KFJJ-202503"],"award-info":[{"award-number":["ZDSYS-KFJJ-202503"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["MAKE"],"abstract":"<jats:p>Perception in trellised orchards is often challenged by dense canopy occlusion and overhead plastic coverings, which cause pronounced variations in sky visibility at row terminals. Accurately recognizing row terminals, including both row head and row tail positions, is therefore essential for understanding orchard row structures. This study presents SkySeg-Net, a sky segmentation-based framework for row-terminal recognition in trellised orchards. SkySeg-Net is built on an enhanced multi-scale U-Net architecture and employs ResNeSt residual split-attention blocks as the backbone. To improve feature discrimination under complex illumination and occlusion conditions, the Convolutional Block Attention Module (CBAM) is integrated into the downsampling path, while a Pyramid Pooling Module (PPM) is introduced during upsampling to strengthen multi-scale contextual representation. Sky regions are segmented from both front-view and rear-view camera images, and a hierarchical threshold-based pixel-sum analysis is applied to infer row-terminal locations based on sky-region distribution patterns. To support a comprehensive evaluation, a dedicated trellised vineyard dataset was constructed, featuring front-view and rear-view images and covering three representative grapevine growth stages (BBCH 69\u201371, 73\u201377, and 79\u201389). Experimental results show that SkySeg-Net achieves an mIoU of 91.21% and an mPA of 94.82% for sky segmentation, with a row-terminal recognition accuracy exceeding 98.17% across all growth stages. These results demonstrate that SkySeg-Net provides a robust and reliable visual perception approach for row-terminal recognition in trellised orchard environments.<\/jats:p>","DOI":"10.3390\/make8020046","type":"journal-article","created":{"date-parts":[[2026,2,13]],"date-time":"2026-02-13T16:09:32Z","timestamp":1770998972000},"page":"46","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["SkySeg-Net: Sky Segmentation-Based Row-Terminal Recognition in Trellised Orchards"],"prefix":"10.3390","volume":"8","author":[{"given":"Haiyang","family":"Gu","sequence":"first","affiliation":[{"name":"State Key Laboratory of Efficient Utilization of Arid and Semi-Arid Arable Land in Northern China, Institute of Agricultural Resources and Regional Planning, Chinese Academy of Agricultural Sciences, Beijing 100086, China"},{"name":"School of Mechanical and Electrical Engineering, Soochow University, Suzhou 215000, China"}]},{"given":"Yong","family":"Wang","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Efficient Utilization of Arid and Semi-Arid Arable Land in Northern China, Institute of Agricultural Resources and Regional Planning, Chinese Academy of Agricultural Sciences, Beijing 100086, China"},{"name":"School of Mechanical and Electrical Engineering, Soochow University, Suzhou 215000, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-4643-5539","authenticated-orcid":false,"given":"Huaiyang","family":"Liu","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Efficient Utilization of Arid and Semi-Arid Arable Land in Northern China, Institute of Agricultural Resources and Regional Planning, Chinese Academy of Agricultural Sciences, Beijing 100086, China"}]},{"given":"Tong","family":"Tian","sequence":"additional","affiliation":[{"name":"Hainan Tang Huajun Academician Workstation, Key Laboratory of Applied Research on Tropical Crop Information Technology of Hainan Province, Institute of Scientific and Technical Information, Chinese Academy of Tropical Agricultural Sciences, Haikou 570000, China"}]},{"given":"Changxing","family":"Geng","sequence":"additional","affiliation":[{"name":"School of Mechanical and Electrical Engineering, Soochow University, Suzhou 215000, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6294-0124","authenticated-orcid":false,"given":"Yun","family":"Shi","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Efficient Utilization of Arid and Semi-Arid Arable Land in Northern China, Institute of Agricultural Resources and Regional Planning, Chinese Academy of Agricultural Sciences, Beijing 100086, China"}]}],"member":"1968","published-online":{"date-parts":[[2026,2,13]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","unstructured":"Li, Y., Feng, Q., Ji, C., Sun, J., and Sun, Y. (2024). GNSS and LiDAR Integrated Navigation Method in Orchards with Intermittent GNSS Dropout. Appl. Sci., 14.","DOI":"10.3390\/app14083231"},{"key":"ref_2","doi-asserted-by":"crossref","unstructured":"Liu, H., Du, Z., Shen, Y., Du, W., and Zhang, X. (2024). Development and evaluation of an intelligent multivariable spraying robot for orchards and nurseries. Comput. Electron. Agric., 222.","DOI":"10.1016\/j.compag.2024.109056"},{"key":"ref_3","doi-asserted-by":"crossref","unstructured":"Su, Z., Zou, W., Zhai, C., Tan, H., Yang, S., and Qin, X. (2024). Design of an Autonomous Orchard Navigation System Based on Multi-Sensor Fusion. Agronomy, 14.","DOI":"10.3390\/agronomy14122825"},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"429","DOI":"10.1002\/rob.21995","article-title":"Localization for precision navigation in agricultural fields\u2014Beyond crop row following","volume":"38","author":"Winterhalter","year":"2021","journal-title":"J. Field Robot."},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"De Silva, R., Cielniak, G., and Gao, J. (2024). Vision based crop row navigation under varying field conditions in arable fields. Comput. Electron. Agric., 217.","DOI":"10.1016\/j.compag.2023.108581"},{"key":"ref_6","first-page":"1","article-title":"Vision Based Row Guidance Approach for Navigation of Agricultural Mobile Robots in Orchards","volume":"18","author":"Shah","year":"2024","journal-title":"Mod. Appl. Sci."},{"key":"ref_7","doi-asserted-by":"crossref","unstructured":"Peng, C., Fei, Z., and Vougioukas, S.G. (July, January 28). Depth camera based row-end detection and headland manuvering in orchard navigation without GNSS. Proceedings of the 2022 30th Mediterranean Conference on control and Automation (MED), Vouliagmeni, Greece.","DOI":"10.1109\/MED54222.2022.9837270"},{"key":"ref_8","doi-asserted-by":"crossref","unstructured":"Peng, C., Fei, Z., and Vougioukas, S.G. (2023). GNSS-free end-of-row detection and headland maneuvering for orchard navigation using a depth camera. Machines, 11.","DOI":"10.3390\/machines11010084"},{"key":"ref_9","doi-asserted-by":"crossref","unstructured":"Jiang, S., Qi, P., Han, L., Liu, L., Li, Y., Huang, Z., Liu, Y., and He, X. (2024). Navigation system for orchard spraying robot based on 3D LiDAR SLAM with NDT_ICP point cloud registration. Comput. Electron. Agric., 220.","DOI":"10.1016\/j.compag.2024.108870"},{"key":"ref_10","doi-asserted-by":"crossref","unstructured":"Xia, Y., Lei, X., Pan, J., Chen, L., Zhang, Z., and Lyu, X. (2023). Research on orchard navigation method based on fusion of 3D SLAM and point cloud positioning. Front. Plant Sci., 14.","DOI":"10.3389\/fpls.2023.1207742"},{"key":"ref_11","doi-asserted-by":"crossref","unstructured":"Anagnostis, A., Tagarakis, A.C., Kateris, D., Moysiadis, V., S\u00f8rensen, C.G., Pearson, S., and Bochtis, D. (2021). Orchard mapping with deep learning semantic segmentation. Sensors, 21.","DOI":"10.3390\/s21113813"},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Gao, Y., Tian, G., Gu, B., Zhao, J., Liu, Q., Qiu, C., and Xue, J. (2023). A Study on the Rapid Detection of Steering Markers in Orchard Management Robots Based on Improved YOLOv7. Electronics, 12.","DOI":"10.3390\/electronics12173614"},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Gu, B., Liu, Q., Gao, Y., Tian, G., Zhang, B., Wang, H., and Li, H. (2023). Research on the Relative Position Detection Method between Orchard Robots and Fruit Tree Rows. Sensors, 23.","DOI":"10.20944\/preprints202309.1150.v1"},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"Kang, H., and Chen, C.J.C. (2020). Fruit detection, segmentation and 3D visualisation of environments in apple orchards. Comput. Electron. Agric., 171.","DOI":"10.1016\/j.compag.2020.105302"},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"22941","DOI":"10.1109\/JSEN.2025.3564622","article-title":"Real-Time Navigation Line Extraction for Safflower Harvesting Robots Using an Improved Detection Transformer","volume":"25","author":"Wu","year":"2025","journal-title":"IEEE Sens. J."},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Xu, Y., Xue, X., Sun, Z., Gu, W., Cui, L., Jin, Y., and Lan, Y. (2025). Global path planning for navigating orchard vehicle based on fruit tree positioning and planting rows detection from UAV imagery. Comput. Electron. Agric., 236.","DOI":"10.1016\/j.compag.2025.110446"},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Xue, X., Luo, Q., Bu, M., Li, Z., Lyu, S., and Song, S. (2023). Citrus tree canopy segmentation of orchard spraying robot based on RGB-D image and the improved DeepLabv3+. Agronomy, 13.","DOI":"10.3390\/agronomy13082059"},{"key":"ref_18","doi-asserted-by":"crossref","unstructured":"Zhang, J., Tian, M., Yang, Z., Li, J., and Zhao, L. (2024). An improved target detection method based on YOLOv5 in natural orchard environments. Comput. Electron. Agric., 219.","DOI":"10.1016\/j.compag.2024.108780"},{"key":"ref_19","doi-asserted-by":"crossref","unstructured":"Zhang, W., Chen, X., Qi, J., and Yang, S. (2022). Automatic instance segmentation of orchard canopy in unmanned aerial vehicle imagery using deep learning. Front. Plant Sci., 13.","DOI":"10.3389\/fpls.2022.1041791"},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"852","DOI":"10.1109\/LRA.2020.2965061","article-title":"MinneApple: A benchmark dataset for apple detection and segmentation","volume":"5","author":"Roy","year":"2020","journal-title":"IEEE Robot. Autom. Lett."},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"1564","DOI":"10.1007\/s11263-024-02230-4","article-title":"4seasons: Benchmarking visual slam and long-term localization for autonomous driving in challenging conditions","volume":"133","author":"Wenzel","year":"2025","journal-title":"Int. J. Comput. Vis."},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Kusrini, K., Suputa, S., Setyanto, A., Agastya, I.M.A., Priantoro, H., Chandramouli, K., and Izquierdo, E. (2020). Data augmentation for automated pest classification in Mango farms. Comput. Electron. Agric., 179.","DOI":"10.1016\/j.compag.2020.105842"},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"60","DOI":"10.1186\/s40537-019-0197-0","article-title":"A survey on image data augmentation for deep learning","volume":"6","author":"Shorten","year":"2019","journal-title":"J. Big Data"},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., and Brox, T. (2015, January 5\u20139). U-net: Convolutional networks for biomedical image segmentation. Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany.","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Zhang, H., Wu, C., Zhang, Z., Zhu, Y., Lin, H., Zhang, Z., Sun, Y., He, T., Mueller, J., and Manmatha, R. (2022, January 18\u201324). Resnest: Split-attention networks. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA.","DOI":"10.1109\/CVPRW56347.2022.00309"},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Woo, S., Park, J., Lee, J.-Y., and Kweon, I.S. (2018, January 8\u201314). Cbam: Convolutional block attention module. Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany.","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Zhao, H., Shi, J., Qi, X., Wang, X., and Jia, J. (2017, January 21\u201326). Pyramid scene parsing network. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.660"},{"key":"ref_28","doi-asserted-by":"crossref","first-page":"303","DOI":"10.1007\/s11263-009-0275-4","article-title":"The pascal visual object classes (voc) challenge","volume":"88","author":"Everingham","year":"2010","journal-title":"Int. J. Comput. Vis."},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Long, J., Shelhamer, E., and Darrell, T. (2015, January 7\u201312). Fully convolutional networks for semantic segmentation. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA.","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Alom, M.Z., Hasan, M., Yakopcic, C., Taha, T.M., and Asari, V.K. (2018). Recurrent residual convolutional neural network based on u-net (r2u-net) for medical image segmentation. arXiv.","DOI":"10.1109\/NAECON.2018.8556686"},{"key":"ref_31","unstructured":"Oktay, O., Schlemper, J., Folgoc, L.L., Lee, M., Heinrich, M., Misawa, K., Mori, K., McDonagh, S., Hammerla, N.Y., and Kainz, B. (2018). Attention u-net: Learning where to look for the pancreas. arXiv."},{"key":"ref_32","doi-asserted-by":"crossref","unstructured":"Zuo, Q., Chen, S., and Wang, Z. (2021). R2AU-Net: Attention recurrent residual convolutional neural network for multimodal medical image segmentation. Secur. Commun. Netw., 2021.","DOI":"10.1155\/2021\/6625688"},{"key":"ref_33","doi-asserted-by":"crossref","unstructured":"Zhou, Z., Rahman Siddiquee, M.M., Tajbakhsh, N., and Liang, J. (2018, January 20). Unet++: A nested u-net architecture for medical image segmentation. Proceedings of the International Workshop on Deep Learning in Medical Image Analysis, Granada, Spain.","DOI":"10.1007\/978-3-030-00889-5_1"},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"3349","DOI":"10.1109\/TPAMI.2020.2983686","article-title":"Deep high-resolution representation learning for visual recognition","volume":"43","author":"Wang","year":"2020","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_35","doi-asserted-by":"crossref","unstructured":"Chen, L.-C., Zhu, Y., Papandreou, G., Schroff, F., and Adam, H. (2018, January 8\u201314). Encoder-decoder with atrous separable convolution for semantic image segmentation. Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany.","DOI":"10.1007\/978-3-030-01234-2_49"}],"container-title":["Machine Learning and Knowledge Extraction"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2504-4990\/8\/2\/46\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,13]],"date-time":"2026-02-13T16:39:37Z","timestamp":1771000777000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2504-4990\/8\/2\/46"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2,13]]},"references-count":35,"journal-issue":{"issue":"2","published-online":{"date-parts":[[2026,2]]}},"alternative-id":["make8020046"],"URL":"https:\/\/doi.org\/10.3390\/make8020046","relation":{},"ISSN":["2504-4990"],"issn-type":[{"value":"2504-4990","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,2,13]]}}}