{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,29]],"date-time":"2025-12-29T18:37:23Z","timestamp":1767033443967,"version":"build-2065373602"},"reference-count":34,"publisher":"MDPI AG","issue":"22","license":[{"start":{"date-parts":[[2022,11,17]],"date-time":"2022-11-17T00:00:00Z","timestamp":1668643200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"BK21 FOUR project funded by the Ministry of Education of Korea","award":["4199990113966","NRF-2018R1A6A1A03025109","NRF-2022R1I1A3069260","2020M3H2A1078119","2021-0-00944","2022-0-00816","2022-0-01170"],"award-info":[{"award-number":["4199990113966","NRF-2018R1A6A1A03025109","NRF-2022R1I1A3069260","2020M3H2A1078119","2021-0-00944","2022-0-00816","2022-0-01170"]}]},{"name":"Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education","award":["4199990113966","NRF-2018R1A6A1A03025109","NRF-2022R1I1A3069260","2020M3H2A1078119","2021-0-00944","2022-0-00816","2022-0-01170"],"award-info":[{"award-number":["4199990113966","NRF-2018R1A6A1A03025109","NRF-2022R1I1A3069260","2020M3H2A1078119","2021-0-00944","2022-0-00816","2022-0-01170"]}]},{"name":"Ministry of Science and ICT","award":["4199990113966","NRF-2018R1A6A1A03025109","NRF-2022R1I1A3069260","2020M3H2A1078119","2021-0-00944","2022-0-00816","2022-0-01170"],"award-info":[{"award-number":["4199990113966","NRF-2018R1A6A1A03025109","NRF-2022R1I1A3069260","2020M3H2A1078119","2021-0-00944","2022-0-00816","2022-0-01170"]}]},{"name":"Institute of Information and communications Technology Planning and Evaluation (IITP) grant funded by the Korean government (MSIT)","award":["4199990113966","NRF-2018R1A6A1A03025109","NRF-2022R1I1A3069260","2020M3H2A1078119","2021-0-00944","2022-0-00816","2022-0-01170"],"award-info":[{"award-number":["4199990113966","NRF-2018R1A6A1A03025109","NRF-2022R1I1A3069260","2020M3H2A1078119","2021-0-00944","2022-0-00816","2022-0-01170"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Sensors"],"abstract":"<jats:p>Because of the development of image processing using cameras and the subsequent development of artificial intelligence technology, various fields have begun to develop. However, it is difficult to implement an image processing algorithm that requires a lot of calculations on a light board. This paper proposes a method using real-time deep learning object recognition algorithms in lightweight embedded boards. We have developed an algorithm suitable for lightweight embedded boards by appropriately using two deep neural network architectures. The first architecture requires small computational volumes, although it provides low accuracy. The second architecture uses large computational volumes and provides high accuracy. The area is determined using the first architecture, which processes semantic segmentation with relatively little computation. After masking the area using the more accurate deep learning architecture, object detection is implemented with improved accuracy, as the image is filtered by segmentation and the cases that have not been recognized by various variables, such as differentiation from the background, are excluded. OpenCV (Open source Computer Vision) is used to process input images in Python, and images are processed using an efficient neural network (ENet) and You Only Look Once (YOLO). By running this algorithm, the average error can be reduced by approximately 2.4 times, allowing for more accurate object detection. In addition, object recognition can be performed in real time for lightweight embedded boards, as a rate of about 4 FPS (frames per second) is achieved.<\/jats:p>","DOI":"10.3390\/s22228890","type":"journal-article","created":{"date-parts":[[2022,11,18]],"date-time":"2022-11-18T06:11:34Z","timestamp":1668751894000},"page":"8890","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":10,"title":["Efficient Object Detection Based on Masking Semantic Segmentation Region for Lightweight Embedded Processors"],"prefix":"10.3390","volume":"22","author":[{"given":"Heuijee","family":"Yun","sequence":"first","affiliation":[{"name":"School of Electronic and Electrical Engineering, Kyungpook National University, Daegu 41566, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5560-873X","authenticated-orcid":false,"given":"Daejin","family":"Park","sequence":"additional","affiliation":[{"name":"School of Electronic and Electrical Engineering, Kyungpook National University, Daegu 41566, Republic of Korea"}]}],"member":"1968","published-online":{"date-parts":[[2022,11,17]]},"reference":[{"key":"ref_1","first-page":"215","article-title":"Efficient Power Reduction Technique of LiDAR Sensor for Controlling Detection Accuracy Based on Vehicle Speed","volume":"15","author":"Lee","year":"2020","journal-title":"IEMEK J. Embed. Syst. Appl."},{"key":"ref_2","first-page":"205","article-title":"Communication-power overhead reduction method using template-based linear approximation in lightweight ecg measurement embedded device","volume":"15","author":"Lee","year":"2020","journal-title":"IEMEK J. Embed. Syst. Appl."},{"key":"ref_3","first-page":"193","article-title":"Autonomous-flight Drone Algorithm use Computer vision and GPS","volume":"11","author":"Kim","year":"2016","journal-title":"IEMEK J. Embed. Syst. Appl."},{"key":"ref_4","doi-asserted-by":"crossref","unstructured":"Yogamani, S., Hughes, C., Horgan, J., Sistu, G., Varley, P., O\u2019Dea, D., Uric\u00e1r, M., Milz, S., Simon, M., and Amende, K. (2019\u20132, January 27). Woodscape: A multi-task, multi-camera fisheye dataset for autonomous driving. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Seoul, Republic of Korea.","DOI":"10.1109\/ICCV.2019.00940"},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"3212","DOI":"10.1109\/TNNLS.2018.2876865","article-title":"Object detection with deep learning: A review","volume":"30","author":"Zhao","year":"2019","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"ref_6","first-page":"643","article-title":"Lane detection based on inverse perspective transformation and Kalman filter","volume":"12","author":"Huang","year":"2018","journal-title":"KSII Trans. Internet Inf. Syst. (TIIS)"},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"112848","DOI":"10.1016\/j.eswa.2019.112848","article-title":"Map-Reduce based tipping point scheduler for parallel image processing","volume":"139","author":"Akhtar","year":"2020","journal-title":"Expert Syst. Appl."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"834","DOI":"10.1109\/TNNLS.2018.2852783","article-title":"Cost-Effective Object Detection: Active Sample Mining with Switchable Selection Criteria","volume":"30","author":"Wang","year":"2019","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"ref_9","doi-asserted-by":"crossref","unstructured":"He, K., Gkioxari, G., Doll\u00e1r, P., and Girshick, R. (2017, January 22\u201329). Mask R-CNN. Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy.","DOI":"10.1109\/ICCV.2017.322"},{"key":"ref_10","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., and Farhadi, A. (2016, January 27\u201330). You only look once: Unified, real-time object detection. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.91"},{"key":"ref_11","doi-asserted-by":"crossref","unstructured":"Wang, P., Chen, P., Yuan, Y., Liu, D., Huang, Z., Hou, X., and Cottrell, G. (2018, January 12\u201315). Understanding convolution for semantic segmentation. Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA.","DOI":"10.1109\/WACV.2018.00163"},{"key":"ref_12","unstructured":"Dai, J., Li, Y., He, K., and Sun, J. (2016, January 5\u201311). R-fcn: Object detection via region-based fully convolutional networks. Proceedings of the Advances in Neural Information Processing Systems 29 (NIPS 2016), Barcelona, Spain."},{"key":"ref_13","unstructured":"Paszke, A., Chaurasia, A., Kim, S., and Culurciello, E. (2016). Enet: A deep neural network architecture for real-time semantic segmentation. arXiv."},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., and Sun, J. (2016, January 27\u201330). Deep residual learning for image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Yun, H., and Park, D. (2022, January 21\u201324). Mitigating Overflow of Object Detection Tasks Based on Masking Semantic Difference Region of Vision Snapshot for High Efficiency. Proceedings of the 2022 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), Jeju Island, Republic of Korea.","DOI":"10.1109\/ICAIIC54071.2022.9722651"},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Kim, S., Ji, Y., and Lee, K.B. (2018\u20132, January 31). An effective sign language learning with object detection based ROI segmentation. Proceedings of the 2018 Second IEEE International Conference on Robotic Computing (IRC), Laguna Hills, CA, USA.","DOI":"10.1109\/IRC.2018.00069"},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Andriluka, M., Pishchulin, L., Gehler, P., and Schiele, B. (2014, January 23\u201328). 2d human pose estimation: New benchmark and state of the art analysis. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA.","DOI":"10.1109\/CVPR.2014.471"},{"key":"ref_18","doi-asserted-by":"crossref","unstructured":"Zhao, H., Qi, X., Shen, X., Shi, J., and Jia, J. (2018, January 8\u201314). Icnet for real-time semantic segmentation on high-resolution images. Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany.","DOI":"10.1007\/978-3-030-01219-9_25"},{"key":"ref_19","unstructured":"Abadi, M., Barham, P., Chen, J., Chen, Z., Davis, A., Dean, J., Devin, M., Ghemawat, S., Irving, G., and Isard, M. (2016, January 2\u20134). TensorFlow: A System for Large-Scale Machine Learning. Proceedings of the 12th USENIX Symposium on Operating Systems Design and Implementation (OSDI 16), Savannah, GA, USA."},{"key":"ref_20","first-page":"207","article-title":"Developing of new a tensorflow tutorial model on machine learning: Focusing on the Kaggle titanic dataset","volume":"14","author":"Kim","year":"2019","journal-title":"IEMEK J. Embed. Syst. Appl."},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.Y., and Berg, A.C. (2016, January 11\u201314). Ssd: Single shot multibox detector. Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands.","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"ref_22","unstructured":"Howard, A.G., Zhu, M., Chen, B., Kalenichenko, D., Wang, W., Weyand, T., Andreetto, M., and Adam, H. (2017). Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv."},{"key":"ref_23","unstructured":"Ren, S., He, K., Girshick, R., and Sun, J. (2015, January 7\u201312). Faster R-CNN: Towards real-time object detection with region proposal networks. Proceedings of the Advances in Neural Information Processing Systems 28 (NIPS 2015), Montreal, Canada."},{"key":"ref_24","unstructured":"Trieu, T.H., and GitHub Repository (2022, June 10). Darkflow. Available online: https:\/\/github.com\/thtrieu\/darkflow."},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Yun, H., and Park, D. (2021, January 16\u201319). Yolo-based Realtime Object Detection using Interleaved Redirection of Time-Multiplexed Streamline of Vision Snapshot for Lightweighted Embedded Processors. Proceedings of the 2021 International Symposium on Intelligent Signal Processing and Communication Systems (ISPACS), Hualien City, Taiwan.","DOI":"10.1109\/ISPACS51563.2021.9651042"},{"key":"ref_26","unstructured":"NXP (2021, July 15). Layerscape LS1028A Family of Industrial Applications Processors. Available online: https:\/\/www.nxp.com\/docs\/en\/fact-sheet\/LS1028AFS.pdf."},{"key":"ref_27","unstructured":"NXP (2021, July 15). Layerscape LX2160A Communications Processor. Available online: https:\/\/www.nxp.com\/docs\/en\/fact-sheet\/LX2160AFS.pdf."},{"key":"ref_28","doi-asserted-by":"crossref","first-page":"40","DOI":"10.3758\/BF03213026","article-title":"Theoretical analysis of an alphabetic confusion matrix","volume":"9","author":"Townsend","year":"1971","journal-title":"Percept. Psychophys."},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Redmon, J., and Farhadi, A. (2017, January 21\u201326). YOLO9000: Better, faster, stronger. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.690"},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Ma, J., Chen, L., and Gao, Z. (2017, January 8\u20139). Hardware implementation and optimization of tiny-YOLO network. Proceedings of the International Forum on Digital TV and Wireless Multimedia Communications, Shanghai, China.","DOI":"10.1007\/978-981-10-8108-8_21"},{"key":"ref_31","unstructured":"Bochkovskiy, A., Wang, C.Y., and Liao, H.Y.M. (2020). Yolov4: Optimal speed and accuracy of object detection. arXiv."},{"key":"ref_32","unstructured":"Jocher, G. (2022, May 05). ultralytics\/yolov5: V6.0\u2014YOLOv5n \u2019Nano\u2019 models, Roboflow integration; TensorFlow Export; OpenCV DNN Support. Available online: https:\/\/zenodo.org\/record\/5563715#.Y3LvP3ZBxdh."},{"key":"ref_33","unstructured":"Li, C., Li, L., Jiang, H., Weng, K., Geng, Y., Li, L., Ke, Z., Li, Q., Cheng, M., and Nie, W. (2022). YOLOv6: A single-stage object detection framework for industrial applications. arXiv."},{"key":"ref_34","doi-asserted-by":"crossref","unstructured":"Wang, C.Y., Bochkovskiy, A., and Liao, H.Y.M. (2022). YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv.","DOI":"10.1109\/CVPR52729.2023.00721"}],"container-title":["Sensors"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1424-8220\/22\/22\/8890\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T01:20:12Z","timestamp":1760145612000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1424-8220\/22\/22\/8890"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,11,17]]},"references-count":34,"journal-issue":{"issue":"22","published-online":{"date-parts":[[2022,11]]}},"alternative-id":["s22228890"],"URL":"https:\/\/doi.org\/10.3390\/s22228890","relation":{},"ISSN":["1424-8220"],"issn-type":[{"type":"electronic","value":"1424-8220"}],"subject":[],"published":{"date-parts":[[2022,11,17]]}}}