{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,27]],"date-time":"2025-10-27T20:52:44Z","timestamp":1761598364848,"version":"3.37.3"},"reference-count":30,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2021,8,9]],"date-time":"2021-08-09T00:00:00Z","timestamp":1628467200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,8,9]],"date-time":"2021-08-09T00:00:00Z","timestamp":1628467200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"funder":[{"DOI":"10.13039\/100000006","name":"Office of Naval Research","doi-asserted-by":"publisher","award":["N00014-16-1-2312"],"award-info":[{"award-number":["N00014-16-1-2312"]}],"id":[{"id":"10.13039\/100000006","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000183","name":"Army Research Office","doi-asserted-by":"publisher","award":["W911NF-20-2-0084"],"award-info":[{"award-number":["W911NF-20-2-0084"]}],"id":[{"id":"10.13039\/100000183","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Appl Intell"],"published-print":{"date-parts":[[2022,3]]},"DOI":"10.1007\/s10489-021-02657-z","type":"journal-article","created":{"date-parts":[[2021,8,9]],"date-time":"2021-08-09T05:02:32Z","timestamp":1628485352000},"page":"5290-5309","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["A one-shot next best view system for active object recognition"],"prefix":"10.1007","volume":"52","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3473-9906","authenticated-orcid":false,"given":"Pourya","family":"Hoseini","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1791-3925","authenticated-orcid":false,"given":"Shuvo Kumar","family":"Paul","sequence":"additional","affiliation":[]},{"given":"Mircea","family":"Nicolescu","sequence":"additional","affiliation":[]},{"given":"Monica","family":"Nicolescu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,8,9]]},"reference":[{"issue":"20","key":"2657_CR1","doi-asserted-by":"publisher","first-page":"2440","DOI":"10.3390\/rs11202440","volume":"11","author":"R Almadhoun","year":"2019","unstructured":"Almadhoun R, Abduldayem A, Taha T, Seneviratne L, Zweiri Y (2019) Guided next best view for 3d reconstruction of large complex structures. Remote Sens 11(20):2440","journal-title":"Remote Sens"},{"issue":"5","key":"2657_CR2","doi-asserted-by":"publisher","first-page":"1078","DOI":"10.1109\/TRO.2014.2320795","volume":"30","author":"N Atanasov","year":"2014","unstructured":"Atanasov N, Sankaran B, Le Ny J, Pappas GJ, Daniilidis K (2014) Nonmyopic view planning for active object classification and pose estimation. IEEE Trans Robot 30(5):1078\u20131090","journal-title":"IEEE Trans Robot"},{"issue":"2","key":"2657_CR3","doi-asserted-by":"publisher","first-page":"177","DOI":"10.1007\/s10514-017-9615-3","volume":"42","author":"R Bajcsy","year":"2018","unstructured":"Bajcsy R, Aloimonos Y, Tsotsos JK (2018) Revisiting active perception. Auton Robot 42 (2):177\u2013196","journal-title":"Auton Robot"},{"issue":"5","key":"2657_CR4","doi-asserted-by":"publisher","first-page":"056004","DOI":"10.1088\/1748-3190\/aa7728","volume":"12","author":"O Barzilay","year":"2017","unstructured":"Barzilay O, Zelnik-Manor L, Gutfreund Y, Wagner H, Wolf A (2017) From biokinematics to a robotic active vision system. Bioinspir Biomim 12(5):056004","journal-title":"Bioinspir Biomim"},{"key":"2657_CR5","doi-asserted-by":"crossref","unstructured":"Bircher A, Kamel M, Alexis K, Oleynikova H, Siegwart R (2016) Receding horizon\u201d next-best-view\u201d planner for 3d exploration. In: 2016 IEEE international conference on robotics and automation (ICRA), IEEE, pp 1462\u20131468","DOI":"10.1109\/ICRA.2016.7487281"},{"key":"2657_CR6","doi-asserted-by":"crossref","unstructured":"Cui J, Wen JT, Trinkle J (2019) A multi-sensor next-best-view framework for geometric model-based robotics applications. In: 2019 International conference on robotics and automation (ICRA), IEEE, pp 8769\u20138775","DOI":"10.1109\/ICRA.2019.8794423"},{"key":"2657_CR7","doi-asserted-by":"publisher","first-page":"3336","DOI":"10.1109\/TIP.2019.2959254","volume":"29","author":"D Das","year":"2019","unstructured":"Das D, Lee CG (2019) A two-stage approach to few-shot learning for image recognition. IEEE Trans Image Process 29:3336\u20133350","journal-title":"IEEE Trans Image Process"},{"key":"2657_CR8","doi-asserted-by":"crossref","unstructured":"Doumanoglou A, Kouskouridas R, Malassiotis S, Kim TK (2016) Recovering 6d object pose and predicting next-best-view in the crowd. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3583\u20133592","DOI":"10.1109\/CVPR.2016.390"},{"key":"2657_CR9","doi-asserted-by":"crossref","unstructured":"Edmonds M, Yigit T, Yi J (2020) Auto-calibrated 3d hyperspectral scanning using a heterogeneous set of cameras and lights with spectrally-optimal next-best-view planning. In: 2020 IEEE 16th International conference on automation science and engineering (CASE), pp 863\u2013868. IEEE","DOI":"10.1109\/CASE48305.2020.9216990"},{"key":"2657_CR10","doi-asserted-by":"publisher","first-page":"105448","DOI":"10.1016\/j.knosys.2019.105448","volume":"193","author":"P Gao","year":"2020","unstructured":"Gao P, Yuan R, Wang F, Xiao L, Fujita H, Zhang Y (2020) Siamese attentional keypoint network for high performance visual tracking. Knowl Based Syst 193:105448","journal-title":"Knowl Based Syst"},{"key":"2657_CR11","doi-asserted-by":"publisher","first-page":"52","DOI":"10.1016\/j.ins.2019.12.084","volume":"517","author":"P Gao","year":"2020","unstructured":"Gao P, Zhang Q, Wang F, Xiao L, Fujita H, Zhang Y (2020) Learning reinforced attentional representation for end-to-end visual tracking. Inform Sci 517:52\u201367","journal-title":"Inform Sci"},{"key":"2657_CR12","unstructured":"Gonzalez RC, Richard E (2018) Woods digital image processing, Pearson Prentice Hall"},{"key":"2657_CR13","first-page":"1","volume":"12","author":"T Hayashi","year":"2020","unstructured":"Hayashi T, Fujita H (2020) Cluster-based zero-shot learning for multivariate data. J Ambient Intell Humaniz Comput 12:1\u201315","journal-title":"J Ambient Intell Humaniz Comput"},{"key":"2657_CR14","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"issue":"4","key":"2657_CR15","doi-asserted-by":"publisher","first-page":"71","DOI":"10.3390\/computers8040071","volume":"8","author":"P Hoseini","year":"2019","unstructured":"Hoseini P, Blankenburg J, Nicolescu M, Nicolescu M, Feil-Seifer D (2019) Active eye-in-hand data management to improve the robotic object detection performance. Computers 8(4):71","journal-title":"Computers"},{"key":"2657_CR16","doi-asserted-by":"crossref","unstructured":"Hoseini P, Blankenburg J, Nicolescu M, Nicolescu M, Feil-Seifer D (2019) An active robotic vision system with a pair of moving and stationary cameras. In: International symposium on visual computing, Springer, pp 184\u2013195","DOI":"10.1007\/978-3-030-33723-0_15"},{"key":"2657_CR17","doi-asserted-by":"crossref","unstructured":"Jia Z, Chang YJ, Chen T (2010) A general boosting-based framework for active object recognition. In: British machine vision conference (BMVC), Citeseer, pp 1\u201311","DOI":"10.5244\/C.24.46"},{"issue":"4","key":"2657_CR18","doi-asserted-by":"publisher","first-page":"5323","DOI":"10.1109\/LRA.2020.3007445","volume":"5","author":"M Lauri","year":"2020","unstructured":"Lauri M, Pajarinen J, Peters J, Frintrop S (2020) Multi-sensor next-best-view planning as matroid-constrained submodular maximization. IEEE Robot Autom Lett 5(4):5323\u20135330","journal-title":"IEEE Robot Autom Lett"},{"key":"2657_CR19","doi-asserted-by":"crossref","unstructured":"Lehnert C, Tsai D, Eriksson A, McCool C (2019) 3d move to see: Multi-perspective visual servoing towards the next best view within unstructured and occluded environments. In: 2019 IEEE\/RSJ International conference on intelligent robots and systems (IROS), IEEE, pp 3890\u20133897","DOI":"10.1109\/IROS40897.2019.8967918"},{"key":"2657_CR20","doi-asserted-by":"crossref","unstructured":"Morrison D, Corke P, Leitner J (2019) Multi-view picking: Next-best-view reaching for improved grasping in clutter. In: 2019 International conference on robotics and automation (ICRA), IEEE, pp 8762\u20138768","DOI":"10.1109\/ICRA.2019.8793805"},{"issue":"2","key":"2657_CR21","doi-asserted-by":"publisher","first-page":"1619","DOI":"10.1109\/LRA.2019.2896759","volume":"4","author":"N Palomeras","year":"2019","unstructured":"Palomeras N, Hurt\u00f3s N, Vidal E, Carreras M (2019) Autonomous exploration of complex underwater environments using a probabilistic next-best-view planner. IEEE Robot Autom Lett 4(2):1619\u20131625","journal-title":"IEEE Robot Autom Lett"},{"key":"2657_CR22","doi-asserted-by":"publisher","first-page":"105590","DOI":"10.1016\/j.knosys.2020.105590","volume":"194","author":"F P\u00e9rez-Hern\u00e1ndez","year":"2020","unstructured":"P\u00e9rez-Hern\u00e1ndez F, Tabik S, Lamas A, Olmos R, Fujita H, Herrera F (2020) Object detection binary classifiers methodology based on deep learning to identify small objects handled similarly: Application in video surveillance. Knowl Based Syst 194 :105590","journal-title":"Knowl Based Syst"},{"issue":"1","key":"2657_CR23","doi-asserted-by":"publisher","first-page":"148","DOI":"10.1016\/j.jvcir.2013.07.006","volume":"25","author":"C Potthast","year":"2014","unstructured":"Potthast C, Sukhatme GS (2014) A probabilistic framework for next best view estimation in a cluttered environment. J Vis Commun Image Represent 25(1):148\u2013164","journal-title":"J Vis Commun Image Represent"},{"key":"2657_CR24","unstructured":"Rebull Mestres J (2017) Implementation of an automated eye-in hand scanning system using best-path planning, Master\u2019s thesis, Universitat Polit\u00e8cnica de Catalunya"},{"issue":"9","key":"2657_CR25","doi-asserted-by":"publisher","first-page":"2220","DOI":"10.1109\/TCSVT.2017.2719122","volume":"28","author":"Z Wang","year":"2017","unstructured":"Wang Z, Xiong J, Yang Y, Li H (2017) A flexible and robust threshold selection method. IEEE Trans Circuits Syst Video Technol 28(9):2220\u20132232","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"2657_CR26","doi-asserted-by":"publisher","first-page":"107405","DOI":"10.1016\/j.asoc.2021.107405","volume":"108","author":"Y Wu","year":"2021","unstructured":"Wu Y, Jiang X, Fang Z, Gao Y, Fujita H (2021) Multi-modal 3d object detection by 2d-guided precision anchor proposal and multi-layer fusion. Appl Soft Comput 108:107405","journal-title":"Appl Soft Comput"},{"key":"2657_CR27","unstructured":"Wu Z, Song S, Khosla A, Yu F, Zhang L, Tang X, Xiao J (2015) 3d shapenets: A deep representation for volumetric shapes. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1912\u20131920"},{"issue":"11","key":"2657_CR28","first-page":"2950","volume":"22","author":"Y Xu","year":"2020","unstructured":"Xu Y, Hu J, Wattanachote K, Zeng K, Gong Y (2020) Sketch-based shape retrieval via best view selection and a cross-domain similarity measure. IEEE Trans Multimed 22(11):2950\u20132962","journal-title":"IEEE Trans Multimed"},{"key":"2657_CR29","doi-asserted-by":"crossref","unstructured":"Zeng R, Zhao W, Liu YJ (2020) Pc-nbv: A point cloud based deep network for efficient next best view planning. In: 2020 IEEE\/RSJ international conference on intelligent robots and systems (IROS), IEEE, pp 7050\u20137057","DOI":"10.1109\/IROS45743.2020.9340916"},{"key":"2657_CR30","doi-asserted-by":"publisher","first-page":"106671","DOI":"10.1016\/j.knosys.2020.106671","volume":"213","author":"K Zhu","year":"2021","unstructured":"Zhu K, Jiang X, Fang Z, Gao Y, Fujita H, Hwang JN (2021) Photometric transfer for direct visual odometry. Knowl Based Syst 213:106671","journal-title":"Knowl Based Syst"}],"container-title":["Applied Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-021-02657-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10489-021-02657-z\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-021-02657-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,3,5]],"date-time":"2022-03-05T05:21:33Z","timestamp":1646457693000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10489-021-02657-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,8,9]]},"references-count":30,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2022,3]]}},"alternative-id":["2657"],"URL":"https:\/\/doi.org\/10.1007\/s10489-021-02657-z","relation":{},"ISSN":["0924-669X","1573-7497"],"issn-type":[{"type":"print","value":"0924-669X"},{"type":"electronic","value":"1573-7497"}],"subject":[],"published":{"date-parts":[[2021,8,9]]},"assertion":[{"value":"29 June 2021","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 August 2021","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}