{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,14]],"date-time":"2025-10-14T00:39:13Z","timestamp":1760402353278,"version":"build-2065373602"},"reference-count":40,"publisher":"MDPI AG","issue":"4","license":[{"start":{"date-parts":[[2021,4,1]],"date-time":"2021-04-01T00:00:00Z","timestamp":1617235200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["J. Imaging"],"abstract":"<jats:p>Object recognition represents the ability of a system to identify objects, humans or animals in images. Within this domain, this work presents a comparative analysis among different classification methods aiming at Tactode tile recognition. The covered methods include: (i) machine learning with HOG and SVM; (ii) deep learning with CNNs such as VGG16, VGG19, ResNet152, MobileNetV2, SSD and YOLOv4; (iii) matching of handcrafted features with SIFT, SURF, BRISK and ORB; and (iv) template matching. A dataset was created to train learning-based methods (i and ii), and with respect to the other methods (iii and iv), a template dataset was used. To evaluate the performance of the recognition methods, two test datasets were built: tactode_small and tactode_big, which consisted of 288 and 12,000 images, holding 2784 and 96,000 regions of interest for classification, respectively. SSD and YOLOv4 were the worst methods for their domain, whereas ResNet152 and MobileNetV2 showed that they were strong recognition methods. SURF, ORB and BRISK demonstrated great recognition performance, while SIFT was the worst of this type of method. The methods based on template matching attained reasonable recognition results, falling behind most other methods. The top three methods of this study were: VGG16 with an accuracy of 99.96% and 99.95% for tactode_small and tactode_big, respectively; VGG19 with an accuracy of 99.96% and 99.68% for the same datasets; and HOG and SVM, which reached an accuracy of 99.93% for tactode_small and 99.86% for tactode_big, while at the same time presenting average execution times of 0.323 s and 0.232 s on the respective datasets, being the fastest method overall. This work demonstrated that VGG16 was the best choice for this case study, since it minimised the misclassifications for both test datasets.<\/jats:p>","DOI":"10.3390\/jimaging7040065","type":"journal-article","created":{"date-parts":[[2021,4,1]],"date-time":"2021-04-01T10:44:01Z","timestamp":1617273841000},"page":"65","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["A Comparative Analysis for 2D Object Recognition: A Case Study with Tactode Puzzle-Like Tiles"],"prefix":"10.3390","volume":"7","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-9999-1550","authenticated-orcid":false,"given":"Daniel","family":"Silva","sequence":"first","affiliation":[{"name":"Institute for Systems and Computer Engineering, Technology and Science (INESC TEC), 4200-465 Porto, Portugal"},{"name":"Department of Engineering, University of Tr\u00e1s-os-Montes e Alto Douro (UTAD), 5000-801 Vila Real, Portugal"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0317-4714","authenticated-orcid":false,"given":"Armando","family":"Sousa","sequence":"additional","affiliation":[{"name":"Institute for Systems and Computer Engineering, Technology and Science (INESC TEC), 4200-465 Porto, Portugal"},{"name":"Faculty of Engineering, University of Porto (FEUP), 4200-465 Porto, Portugal"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4769-8843","authenticated-orcid":false,"given":"Valter","family":"Costa","sequence":"additional","affiliation":[{"name":"Institute of Science and Innovation in Mechanical and Industrial Engineering (INEGI), 4200-465 Porto, Portugal"}]}],"member":"1968","published-online":{"date-parts":[[2021,4,1]]},"reference":[{"key":"ref_1","unstructured":"Cardoso, A. (2019). Tangible Language for Educational Programming of Robots and Other Targets. [Master\u2019s Thesis, Faculdade de Engenharia da Universidade do Porto]."},{"key":"ref_2","doi-asserted-by":"crossref","unstructured":"Alves, M. (2019). Robotic Simulator for the Tactode Tangible Block Programming System. [Master\u2019s Thesis, Faculdade de Engenharia da Universidade do Porto].","DOI":"10.1007\/978-3-030-35990-4_40"},{"key":"ref_3","unstructured":"Silva, D. (2020). Computer Vision System for Tactode Programming. [Master\u2019s Thesis, Faculdade de Engenharia da Universidade do Porto]."},{"key":"ref_4","unstructured":"Sousa, A. (2021, March 30). Tactode. Available online: https:\/\/fe.up.pt\/asousa\/tactode."},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"Alves, M., Sousa, A., and Cardoso, \u00c2. (2020). Web Based Robotic Simulator for Tactode Tangible Block Programming System. Robot 2019: Fourth Iberian Robotics Conference, Springer.","DOI":"10.1007\/978-3-030-35990-4_40"},{"key":"ref_6","unstructured":"Dana, H., and Ballard, C.M.B. (1982). Computer Vision, Prentice-Hall."},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"641","DOI":"10.1016\/0031-3203(95)00118-2","article-title":"Feature extraction methods for character recognition\u2014A survey","volume":"29","author":"Trier","year":"1996","journal-title":"Pattern Recognit."},{"key":"ref_8","unstructured":"Dalal, N., and Triggs, B. (2005, January 20\u201325). Histograms of oriented gradients for human detection. Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR\u201905), San Diego, CA, USA."},{"key":"ref_9","first-page":"1150","article-title":"Object recognition from local scale-invariant features","volume":"2","author":"Lowe","year":"1999","journal-title":"Proc. IEEE Int. Conf. Comput. Vis."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"91","DOI":"10.1023\/B:VISI.0000029664.99615.94","article-title":"Distinctive image features from scale-invariant keypoints","volume":"60","author":"Lowe","year":"2004","journal-title":"Int. J. Comput. Vis."},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"346","DOI":"10.1016\/j.cviu.2007.09.014","article-title":"Speeded-Up Robust Features (SURF)","volume":"110","author":"Bay","year":"2008","journal-title":"Comput. Vis. Image Underst."},{"key":"ref_12","first-page":"183","article-title":"Adaptive and generic corner detection based on the accelerated segment test","volume":"Volume 6312","author":"Daniilidis","year":"2010","journal-title":"ECCV 2010: Computer Vision"},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Leutenegger, S., Chli, M., and Siegwart, R. (2011, January 6\u201313). BRISK: Binary Robust invariant scalable keypoints. Proceedings of the IEEE International Conference on Computer Vision, Barcelona, Spain.","DOI":"10.1109\/ICCV.2011.6126542"},{"key":"ref_14","first-page":"430","article-title":"Machine learning for high-speed corner detection","volume":"Volume 3951","author":"Leonardis","year":"2006","journal-title":"ECCV 2006: Computer Vision"},{"key":"ref_15","first-page":"778","article-title":"BRIEF: Binary robust independent elementary features","volume":"Volume 6314","author":"Daniilidis","year":"2010","journal-title":"ECCV 2010: Computer Vision"},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Rublee, E., Rabaud, V., Konolige, K., and Bradski, G. (2011, January 6\u201313). ORB: An efficient alternative to SIFT or SURF. Proceedings of the IEEE International Conference on Computer Vision, Barcelona, Spain.","DOI":"10.1109\/ICCV.2011.6126544"},{"key":"ref_17","first-page":"1174","article-title":"Machine Learning Algorithms: A Review","volume":"7","author":"Dey","year":"2016","journal-title":"Int. J. Comput. Sci. Inf. Technol."},{"key":"ref_18","first-page":"63","article-title":"Support-Vector Networks","volume":"7","author":"Farhat","year":"1992","journal-title":"IEEE Expert Intell. Syst. Their Appl."},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"1140","DOI":"10.1007\/11424925_119","article-title":"A comparison of model selection methods for multi-class support vector machines","volume":"3483","author":"Li","year":"2005","journal-title":"Lect. Notes Comput. Sci."},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., and Farhadi, A. (2016, January 27\u201330). You only look once: Unified, real-time object detection. Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.91"},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"Redmon, J., and Farhadi, A. (2017, January 21\u201326). YOLO9000: Better, faster, stronger. Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.690"},{"key":"ref_22","unstructured":"Redmon, J., and Farhadi, A. (2020, June 09). YOLO v.3. Technical report, University of Washington. Available online: https:\/\/pjreddie.com\/media\/files\/papers\/YOLOv3.pdf."},{"key":"ref_23","unstructured":"Bochkovskiy, A., Wang, C.Y., and Liao, H.Y.M. (2020). YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv."},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Wang, C.Y., Liao, H.Y.M., Yeh, I.H., Wu, Y.H., Chen, P.Y., and Hsieh, J.W. (2019). CSPNet: A New Backbone that can Enhance Learning Capability of CNN. arXiv.","DOI":"10.1109\/CVPRW50498.2020.00203"},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., and Sun, J. (2014). Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. Lect. Notes Comput. Sci., 346\u2013361.","DOI":"10.1007\/978-3-319-10578-9_23"},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Liu, S., Qi, L., Qin, H., Shi, J., and Jia, J. (2018). Path Aggregation Network for Instance Segmentation. arXiv.","DOI":"10.1109\/CVPR.2018.00913"},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.Y., and Berg, A.C. (2016). SSD: Single Shot MultiBox Detector. Lect. Notes Comput. Sci., 21\u201337.","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"ref_28","unstructured":"Simonyan, K., and Zisserman, A. (2014). Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv."},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., and Sun, J. (2016, January 27\u201330). Deep Residual Learning for Image Recognition. Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref_30","unstructured":"Howard, A.G., Zhu, M., Chen, B., Kalenichenko, D., Wang, W., Weyand, T., Andreetto, M., and Adam, H. (2017). MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv."},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., and Chen, L.C. (2018). MobileNetV2: Inverted Residuals and Linear Bottlenecks. arXiv.","DOI":"10.1109\/CVPR.2018.00474"},{"key":"ref_32","doi-asserted-by":"crossref","unstructured":"Howard, A., Sandler, M., Chu, G., Chen, L.C., Chen, B., Tan, M., Wang, W., Zhu, Y., Pang, R., and Vasudevan, V. (2019). Searching for MobileNetV3. arXiv.","DOI":"10.1109\/ICCV.2019.00140"},{"key":"ref_33","unstructured":"Hough, P. (1962). A Method and Means for Recognition Complex Patterns. (US3069654A), US Patent, Available online: https:\/\/patents.google.com\/patent\/US3069654A\/en."},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"32","DOI":"10.1016\/0734-189X(85)90016-7","article-title":"Topological structural analysis of digitized binary images by border following","volume":"30","author":"Suzuki","year":"1985","journal-title":"Comput. Vis. Graph. Image Process."},{"key":"ref_35","unstructured":"OpenCV (2020, July 04). OpenCV. Available online: https:\/\/docs.opencv.org\/4.4.0\/index.html."},{"key":"ref_36","unstructured":"M\u00fchler, V. (2020, June 20). opencv4nodejs. Available online: https:\/\/www.npmjs.com\/package\/opencv4nodejs."},{"key":"ref_37","doi-asserted-by":"crossref","first-page":"381","DOI":"10.1145\/358669.358692","article-title":"Random Sample Consensus: A Paradigm for Model Fitting with Applications to Image Analysis and Automated Cartography","volume":"24","author":"Fischler","year":"1981","journal-title":"Commun. ACM"},{"key":"ref_38","unstructured":"Google (2020, October 21). Google Colaboratory. Available online: https:\/\/colab.research.google.com\/notebooks\/intro.ipynb."},{"key":"ref_39","unstructured":"TensorFlow (2020, October 20). TensorFlow. Available online: https:\/\/www.tensorflow.org."},{"key":"ref_40","doi-asserted-by":"crossref","unstructured":"Tareen, S.A.K., and Saleem, Z. (2018, January 3\u20134). A comparative analysis of SIFT, SURF, KAZE, AKAZE, ORB, and BRISK. Proceedings of the 2018 International Conference on Computing, Mathematics and Engineering Technologies (iCoMET), Sukkur, Pakistan.","DOI":"10.1109\/ICOMET.2018.8346440"}],"container-title":["Journal of Imaging"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2313-433X\/7\/4\/65\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,13]],"date-time":"2025-10-13T13:33:27Z","timestamp":1760362407000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2313-433X\/7\/4\/65"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,4,1]]},"references-count":40,"journal-issue":{"issue":"4","published-online":{"date-parts":[[2021,4]]}},"alternative-id":["jimaging7040065"],"URL":"https:\/\/doi.org\/10.3390\/jimaging7040065","relation":{},"ISSN":["2313-433X"],"issn-type":[{"type":"electronic","value":"2313-433X"}],"subject":[],"published":{"date-parts":[[2021,4,1]]}}}