{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T03:20:56Z","timestamp":1740108056428,"version":"3.37.3"},"reference-count":29,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2023,4,18]],"date-time":"2023-04-18T00:00:00Z","timestamp":1681776000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,4,18]],"date-time":"2023-04-18T00:00:00Z","timestamp":1681776000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"CONACYT","award":["707984"],"award-info":[{"award-number":["707984"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Pattern Anal Applic"],"published-print":{"date-parts":[[2023,8]]},"DOI":"10.1007\/s10044-023-01154-y","type":"journal-article","created":{"date-parts":[[2023,4,18]],"date-time":"2023-04-18T10:02:58Z","timestamp":1681812178000},"page":"1045-1057","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Shape completion using orthogonal views through a multi-input\u2013output network"],"prefix":"10.1007","volume":"26","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-7805-5891","authenticated-orcid":false,"given":"Leonardo","family":"Delgado","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7618-8762","authenticated-orcid":false,"given":"Eduardo F.","family":"Morales","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,4,18]]},"reference":[{"key":"1154_CR1","doi-asserted-by":"publisher","unstructured":"Varley J, DeChant C, Richardson A, Ruales J, Allen P (2017) Shape completion enabled robotic grasping. In: 2017 IEEE\/RSJ international conference on intelligent robots and systems (IROS), pp 2442\u20132447. https:\/\/doi.org\/10.1109\/iros.2017.8206060","DOI":"10.1109\/iros.2017.8206060"},{"issue":"12","key":"1154_CR2","doi-asserted-by":"publisher","first-page":"2820","DOI":"10.1109\/TPAMI.2018.2868195","volume":"41","author":"B Yang","year":"2019","unstructured":"Yang B, Rosa S, Markham A, Trigoni N, Wen H (2019) Dense 3D object reconstruction from a single depth view. IEEE Trans Pattern Anal Mach Intell 41(12):2820\u20132834. https:\/\/doi.org\/10.1109\/TPAMI.2018.2868195","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"1154_CR3","doi-asserted-by":"crossref","unstructured":"Yuan W, Khot T, Held D, Mertz C, Hebert M (2018) Pcn: point completion network. In: 2018 international conference on 3D vision (3DV), pp 728\u2013737","DOI":"10.1109\/3DV.2018.00088"},{"key":"1154_CR4","doi-asserted-by":"crossref","unstructured":"Liu M, Sheng L, Yang S, Shao J, Hu S-M (2019) Morphing and sampling network for dense point cloud completion. In: The thirty-fourth AAAI conference on artificial intelligence","DOI":"10.1609\/aaai.v34i07.6827"},{"key":"1154_CR5","doi-asserted-by":"publisher","first-page":"30969","DOI":"10.1109\/ACCESS.2020.2973003","volume":"8","author":"Y Peng","year":"2020","unstructured":"Peng Y, Chang M, Wang Q, Qian Y, Zhang Y, Wei M, Liao X (2020) Sparse-to-dense multi-encoder shape completion of unstructured point cloud. IEEE Access 8:30969\u201330978","journal-title":"IEEE Access"},{"key":"1154_CR6","doi-asserted-by":"publisher","unstructured":"Yu X, Rao Y, Wang Z, Liu Z, Lu J, Zhou J (2021) Pointr: diverse point cloud completion with geometry-aware transformers. In: 2021 IEEE\/CVF international conference on computer vision (ICCV), pp 12478\u201312487. https:\/\/doi.org\/10.1109\/ICCV48922.2021.01227","DOI":"10.1109\/ICCV48922.2021.01227"},{"key":"1154_CR7","doi-asserted-by":"crossref","unstructured":"Xiang P, Wen X, Liu Y-S, Cao Y-P, Wan P, Zheng W, Han Z (2021) SnowflakeNet: point cloud completion by snowflake point deconvolution with skip-transformer. In: Proceedings of the IEEE international conference on computer vision (ICCV)","DOI":"10.1109\/ICCV48922.2021.00545"},{"key":"1154_CR8","doi-asserted-by":"publisher","unstructured":"Charles RQ, Su H, Kaichun M, Guibas LJ (2017) Pointnet: deep learning on point sets for 3D classification and segmentation. In: 2017 IEEE conference on computer vision and pattern recognition (CVPR), pp 77\u201385. https:\/\/doi.org\/10.1109\/CVPR.2017.16","DOI":"10.1109\/CVPR.2017.16"},{"key":"1154_CR9","volume-title":"Advances in neural information processing systems","author":"CR Qi","year":"2017","unstructured":"Qi CR, Yi L, Su H, Guibas LJ (2017) Pointnet++: deep hierarchical feature learning on point sets in a metric space. In: Guyon I, Luxburg UV, Bengio S, Wallach H, Fergus R, Vishwanathan S, Garnett R (eds) Advances in neural information processing systems, vol 30. Curran Associates Inc, USA"},{"key":"1154_CR10","doi-asserted-by":"publisher","unstructured":"Hu T, Han Z, Shrivastava A, Zwicker M (2019) Render4completion: synthesizing multi-view depth maps for 3D shape completion. In: 2019 IEEE\/CVF international conference on computer vision workshop (ICCVW), pp 4114\u20134122. https:\/\/doi.org\/10.1109\/ICCVW.2019.00506","DOI":"10.1109\/ICCVW.2019.00506"},{"key":"1154_CR11","doi-asserted-by":"crossref","unstructured":"Hu T, Han Z, Zwicker M (2020) 3D shape completion with multi-view consistent inference. In: The Thirty-Fourth AAAI conference on artificial intelligence, AAAI 2020, The Thirty-Second innovative applications of artificial intelligence conference, IAAI 2020, The tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, 7\u201312 Feb 2020, pp 10997\u201311004","DOI":"10.1609\/aaai.v34i07.6734"},{"key":"1154_CR12","unstructured":"Chang AX, Funkhouser T, Guibas L, Hanrahan P, Huang Q, Li Z, Savarese S, Savva M, Song S, Su H, Xiao J, Yi L, Yu F (2015) ShapeNet: an information-rich 3D model repository. Technical Report arXiv:1512.03012 [cs.GR], Toyota Technological Institute, Chicago"},{"key":"1154_CR13","doi-asserted-by":"publisher","unstructured":"Calli B, Singh A, Walsman A, Srinivasa S, Abbeel P, Dollar AM (2015) The ycb object and model set: towards common benchmarks for manipulation research. In: 2015 international conference on advanced robotics (ICAR), pp 510\u2013517. https:\/\/doi.org\/10.1109\/ICAR.2015.7251504","DOI":"10.1109\/ICAR.2015.7251504"},{"key":"1154_CR14","doi-asserted-by":"publisher","unstructured":"Kappler D, Bohg J, Schaal S (2015) Leveraging big data for grasp planning. In: 2015 IEEE international conference on robotics and automation (ICRA), pp 4304\u20134311. https:\/\/doi.org\/10.1109\/ICRA.2015.7139793","DOI":"10.1109\/ICRA.2015.7139793"},{"key":"1154_CR15","doi-asserted-by":"publisher","unstructured":"Koenig N, Howard A (2004) Design and use paradigms for gazebo, an open-source multi-robot simulator. In: 2004 IEEE\/RSJ international conference on intelligent robots and systems (IROS) (IEEE Cat. No.04CH37566), vol 3, pp 2149\u201321543 . https:\/\/doi.org\/10.1109\/IROS.2004.1389727","DOI":"10.1109\/IROS.2004.1389727"},{"key":"1154_CR16","unstructured":"Min P (2019) Binvox. http:\/\/www.patrickmin.com\/binvox or https:\/\/www.google.com\/search?q=binvox. Accessed on 05 Oct 2019"},{"key":"1154_CR17","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1371\/journal.pone.0263916","volume":"17","author":"M Saha","year":"2022","unstructured":"Saha M, Amin SB, Sharma A, Kumar TKS, Kalia RK (2022) AI-driven quantification of ground glass opacities in lungs of COVID-19 patients using 3D computed tomography imaging. PLoS ONE 17:1\u201314. https:\/\/doi.org\/10.1371\/journal.pone.0263916","journal-title":"PLoS ONE"},{"key":"1154_CR18","doi-asserted-by":"crossref","unstructured":"Szegedy C, Ioffe S, Vanhoucke V, Alemi AA (2017) Inception-v4, inception-resnet and the impact of residual connections on learning. In: Proceedings of the Thirty-First AAAI conference on artificial intelligence. AAAI\u201917, pp 4278\u20134284","DOI":"10.1609\/aaai.v31i1.11231"},{"key":"1154_CR19","first-page":"234","volume-title":"Medical image computing and computer-assisted intervention\u2014MICCAI 2015","author":"O Ronneberger","year":"2015","unstructured":"Ronneberger O, Fischer P, Brox T (2015) U-net: convolutional networks for biomedical image segmentation. In: Navab N, Hornegger J, Wells WM, Frangi AF (eds) Medical image computing and computer-assisted intervention\u2014MICCAI 2015. Springer, Cham, pp 234\u2013241"},{"key":"1154_CR20","doi-asserted-by":"publisher","unstructured":"Riegler G, Ulusoy AO, Geiger A (2017) Octnet: learning deep 3D representations at high resolutions. In: 2017 IEEE conference on computer vision and pattern recognition (CVPR), pp 6620\u20136629. https:\/\/doi.org\/10.1109\/CVPR.2017.701","DOI":"10.1109\/CVPR.2017.701"},{"key":"1154_CR21","doi-asserted-by":"publisher","first-page":"628","DOI":"10.1007\/978-3-319-46484-8_38","volume-title":"Computer vision\u2014ECCV 2016","author":"CB Choy","year":"2016","unstructured":"Choy CB, Xu D, Gwak J, Chen K, Savarese S (2016) 3D\u2013r2n2: a unified approach for single and multi-view 3D object reconstruction. In: Leibe B, Matas J, Sebe N, Welling M (eds) Computer vision\u2014ECCV 2016. Springer, Cham, pp 628\u2013644"},{"key":"1154_CR22","doi-asserted-by":"crossref","unstructured":"Rusu RB, Cousins S (2011) 3D is here: point cloud library (PCL). In: IEEE international conference on robotics and automation (ICRA). Shanghai, China, pp 1\u20134","DOI":"10.1109\/ICRA.2011.5980567"},{"key":"1154_CR23","doi-asserted-by":"publisher","unstructured":"Han X, Li Z, Huang H, Kalogerakis E, Yu Y (2017) High-resolution shape completion using deep neural networks for global structure and local geometry inference. In: 2017 IEEE international conference on computer vision (ICCV), pp 85\u201393. https:\/\/doi.org\/10.1109\/ICCV.2017.19","DOI":"10.1109\/ICCV.2017.19"},{"key":"1154_CR24","volume-title":"A guide to NumPy","author":"T Oliphant","year":"2006","unstructured":"Oliphant T (2006) A guide to NumPy. Trelgol Publishing, USA"},{"key":"1154_CR25","unstructured":"Kingma D, Ba J (2014) Adam: a method for stochastic optimization. In: International conference on learning representations. arxiv:1412.6980"},{"key":"1154_CR26","unstructured":"Chollet F (2021) Deep Learning with Python, Second Edition. ISBN 9781617296864"},{"key":"1154_CR27","unstructured":"Abadi M, Agarwal A et al (2015) TensorFlow: large-scale machine learning on heterogeneous systems. Software available from https:\/\/www.tensorflow.org\/"},{"key":"1154_CR28","volume-title":"Deep learning with Python","author":"F Chollet","year":"2017","unstructured":"Chollet F (2017) Deep learning with Python, 1st edn. Manning Publications Co., Greenwich","edition":"1"},{"key":"1154_CR29","doi-asserted-by":"publisher","unstructured":"Do T-T, Nguyen A, Reid I (2018) AffordanceNet: an end-to-end deep learning approach for object affordance detection. In: 2018 IEEE international conference on robotics and automation (ICRA), pp 5882\u20135889. https:\/\/doi.org\/10.1109\/icra.2018.8460902","DOI":"10.1109\/icra.2018.8460902"}],"container-title":["Pattern Analysis and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10044-023-01154-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10044-023-01154-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10044-023-01154-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,7,22]],"date-time":"2023-07-22T14:05:54Z","timestamp":1690034754000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10044-023-01154-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,4,18]]},"references-count":29,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2023,8]]}},"alternative-id":["1154"],"URL":"https:\/\/doi.org\/10.1007\/s10044-023-01154-y","relation":{},"ISSN":["1433-7541","1433-755X"],"issn-type":[{"type":"print","value":"1433-7541"},{"type":"electronic","value":"1433-755X"}],"subject":[],"published":{"date-parts":[[2023,4,18]]},"assertion":[{"value":"3 July 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"24 January 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 April 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare they have no financial interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}