{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,5]],"date-time":"2026-01-05T07:33:07Z","timestamp":1767598387975,"version":"3.38.0"},"reference-count":40,"publisher":"SAGE Publications","issue":"5","license":[{"start":{"date-parts":[[2023,9,1]],"date-time":"2023-09-01T00:00:00Z","timestamp":1693526400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/journals.sagepub.com\/page\/policies\/text-and-data-mining-license"}],"content-domain":{"domain":["journals.sagepub.com"],"crossmark-restriction":true},"short-container-title":["Intelligent Data Analysis: An International Journal"],"published-print":{"date-parts":[[2023,9]]},"abstract":"<jats:p> The first major contribution of the paper is the proposal of using an improved DEtection Transformer network (named R2N-DETR) and Kinect-V2 camera for detecting multiple-size peaches under orchards with varied illumination and fruit occlusion. R2N-DETR model first employed Res2Net-50 to extract a fused low-high level feature map containing fine spatial features and precise semantic information of multi-size peaches from Red-Green-Blue-Depth (RGB-D) images. Second, the encoder-decoder was performed on the feature map to obtain the global context. Finally, all detected objects were detected according to each object\u2019s global context. For the detection of 1101 RGB-D images (imaged from two orchards over three years), the R2N-DETR model achieves an average precision of 0.944 and an average detecting time of 53 ms for each image. The developed system could provide precise visual guidance for robotic picking and contribute to improving yield prediction by providing accurate fruit counting. <\/jats:p>","DOI":"10.3233\/ida-220449","type":"journal-article","created":{"date-parts":[[2025,1,27]],"date-time":"2025-01-27T19:39:51Z","timestamp":1738006791000},"page":"1539-1554","update-policy":"https:\/\/doi.org\/10.1177\/sage-journals-update-policy","source":"Crossref","is-referenced-by-count":2,"title":["Detection of multi-size peach in orchard using RGB-D camera combined with an improved DEtection Transformer model"],"prefix":"10.1177","volume":"27","author":[{"given":"Yu","family":"Yang","sequence":"first","affiliation":[{"name":"Key Laboratory of Advanced Process Control for Light Industry (Ministry of Education), Jiangnan University, Wuxi, Jiangsu, China"},{"name":"Department of Bioresource Engineering, Macdonald Campus, McGill University, Sainte-Anne-de-Bellevue, QC, Canada"}]},{"given":"Xin","family":"Wang","sequence":"additional","affiliation":[{"name":"Key Laboratory of Advanced Process Control for Light Industry (Ministry of Education), Jiangnan University, Wuxi, Jiangsu, China"}]},{"given":"Zhenfang","family":"Liu","sequence":"additional","affiliation":[{"name":"Key Laboratory of Advanced Process Control for Light Industry (Ministry of Education), Jiangnan University, Wuxi, Jiangsu, China"}]},{"given":"Min","family":"Huang","sequence":"additional","affiliation":[{"name":"Key Laboratory of Advanced Process Control for Light Industry (Ministry of Education), Jiangnan University, Wuxi, Jiangsu, China"}]},{"given":"Shangpeng","family":"Sun","sequence":"additional","affiliation":[{"name":"Department of Bioresource Engineering, Macdonald Campus, McGill University, Sainte-Anne-de-Bellevue, QC, Canada"}]},{"given":"Qibing","family":"Zhu","sequence":"additional","affiliation":[{"name":"Key Laboratory of Advanced Process Control for Light Industry (Ministry of Education), Jiangnan University, Wuxi, Jiangsu, China"}]}],"member":"179","published-online":{"date-parts":[[2023,9,1]]},"reference":[{"key":"bibr1-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1016\/j.hpj.2020.06.001"},{"key":"bibr2-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2020.113594"},{"key":"bibr3-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1080\/20964129.2019.1571443"},{"key":"bibr4-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1504\/IJCVR."},{"key":"bibr5-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1016\/j.compag.2022.107085"},{"key":"bibr6-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1016\/j.asoc.2021.107805"},{"key":"bibr7-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1016\/j.biosystemseng.2020.07.007"},{"key":"bibr8-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1016\/j.biosystemseng.2016.01.013"},{"key":"bibr9-IDA-220449","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1905.05055"},{"key":"bibr10-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2018.2876865"},{"key":"bibr11-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-022-13644-y"},{"key":"bibr12-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2022.07.003"},{"key":"bibr13-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1016\/j.biosystemseng.2018.09.004"},{"key":"bibr14-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1016\/j.compag.2020.105475"},{"key":"bibr15-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1016\/j.ifacol.2019.12.500"},{"key":"bibr16-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1016\/j.patrec.2021.04.022"},{"key":"bibr17-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1016\/j.compag.2019.01.012"},{"key":"bibr18-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1016\/j.compag.2021.106533"},{"key":"bibr19-IDA-220449","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2011.09315"},{"key":"bibr20-IDA-220449","doi-asserted-by":"crossref","unstructured":"HeoB. YunS. HanD. ChunS. ChoeJ. and OhS.J., Rethinking spatial dimensions of vision transformers, in: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2021,  11936\u201311945.","DOI":"10.1109\/ICCV48922.2021.01172"},{"key":"bibr21-IDA-220449","doi-asserted-by":"crossref","unstructured":"CarionN. MassaF. SynnaeveG. UsunierN. KirillovA. and ZagoruykoS., End-to-end object detection with transformers, in: European Conference on Computer Vision, Springer, 2020,  213\u2013229.","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"bibr22-IDA-220449","doi-asserted-by":"crossref","unstructured":"LinT.Y. MaireM. BelongieS. HaysJ. PeronaP. RamananD. Doll\u00e1rP. and ZitnickC.L., Microsoft coco: Common objects in context, in: European Conference on Computer Vision, Springer, 2014,  740\u2013755.","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"bibr23-IDA-220449","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2010.04159"},{"key":"bibr24-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1007\/s11119-020-09754-y"},{"key":"bibr25-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2019.2938758"},{"key":"bibr26-IDA-220449","unstructured":"VaswaniA. ShazeerN. ParmarN. UszkoreitJ. JonesL. GomezA.N. Kaiser\u0141. and PolosukhinI., Attention is all you need, Advances in Neural Information Processing Systems 30 (2017), 5998\u20136008."},{"key":"bibr27-IDA-220449","unstructured":"BelloI. ZophB. VaswaniA. ShlensJ. and LeQ.V., Attention augmented convolutional networks, in: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2019, pp. 3286\u20133295."},{"key":"bibr28-IDA-220449","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1711.05101"},{"key":"bibr29-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1007\/s00521-021-06651-x"},{"key":"bibr30-IDA-220449","doi-asserted-by":"crossref","unstructured":"QiuH. MaY. LiZ. LiuS. and SunJ., Borderdet: Border feature for dense object detection, in: European Conference on Computer Vision, Springer, 2020, pp. 549\u2013564.","DOI":"10.1007\/978-3-030-58452-8_32"},{"key":"bibr31-IDA-220449","doi-asserted-by":"crossref","unstructured":"SongG. LiuY. and WangX., Revisiting the sibling head in object detector, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 11563\u201311572.","DOI":"10.1109\/CVPR42600.2020.01158"},{"key":"bibr32-IDA-220449","unstructured":"YangJ. LiC. ZhangP. DaiX. XiaoB. YuanL. and GaoJ., Focal attention for long-range interactions in vision transformers, Advances in Neural Information Processing Systems 34 (2021), 30008\u201330022."},{"key":"bibr33-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1007\/s10462-021-09967-1"},{"key":"bibr34-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1016\/j.biosystemseng.2021.03.012"},{"key":"bibr35-IDA-220449","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2004.10934"},{"key":"bibr36-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1109\/TASE.2020.2964289"},{"key":"bibr37-IDA-220449","doi-asserted-by":"publisher","DOI":"10.1016\/j.comnet.2019.107036"},{"key":"bibr38-IDA-220449","doi-asserted-by":"crossref","unstructured":"GongC. WangD. LiM. ChandraV. and LiuQ., Keepaugment: A simple information-preserving data augmentation approach, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 1055\u20131064.","DOI":"10.1109\/CVPR46437.2021.00111"},{"key":"bibr39-IDA-220449","doi-asserted-by":"crossref","unstructured":"YangC. WangY. ZhangJ. ZhangH. WeiZ. LinZ. and YuilleA., Lite vision transformer with enhanced self-attention, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 11998\u201312008.","DOI":"10.1109\/CVPR52688.2022.01169"},{"key":"bibr40-IDA-220449","doi-asserted-by":"crossref","unstructured":"DaiX. ChenY. YangJ. ZhangP. YuanL. and ZhangL., Dynamic detr: End-to-end object detection with dynamic attention, in: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2021, pp. 2988\u20132997.","DOI":"10.1109\/ICCV48922.2021.00298"}],"container-title":["Intelligent Data Analysis: An International Journal"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/journals.sagepub.com\/doi\/pdf\/10.3233\/IDA-220449","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/journals.sagepub.com\/doi\/full-xml\/10.3233\/IDA-220449","content-type":"application\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/journals.sagepub.com\/doi\/pdf\/10.3233\/IDA-220449","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,3,11]],"date-time":"2025-03-11T06:12:22Z","timestamp":1741673542000},"score":1,"resource":{"primary":{"URL":"https:\/\/journals.sagepub.com\/doi\/10.3233\/IDA-220449"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,9]]},"references-count":40,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2023,9]]}},"alternative-id":["10.3233\/IDA-220449"],"URL":"https:\/\/doi.org\/10.3233\/ida-220449","relation":{},"ISSN":["1088-467X","1571-4128"],"issn-type":[{"type":"print","value":"1088-467X"},{"type":"electronic","value":"1571-4128"}],"subject":[],"published":{"date-parts":[[2023,9]]}}}