{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,28]],"date-time":"2026-03-28T18:00:05Z","timestamp":1774720805669,"version":"3.50.1"},"reference-count":23,"publisher":"MDPI AG","issue":"19","license":[{"start":{"date-parts":[[2024,10,9]],"date-time":"2024-10-09T00:00:00Z","timestamp":1728432000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100001691","name":"JSPS KAKENHI","doi-asserted-by":"publisher","award":["21H04903"],"award-info":[{"award-number":["21H04903"]}],"id":[{"id":"10.13039\/501100001691","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001691","name":"JSPS KAKENHI","doi-asserted-by":"publisher","award":["FRF-IDRY-23-001"],"award-info":[{"award-number":["FRF-IDRY-23-001"]}],"id":[{"id":"10.13039\/501100001691","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100008778","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"publisher","award":["21H04903"],"award-info":[{"award-number":["21H04903"]}],"id":[{"id":"10.13039\/501100008778","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100008778","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"publisher","award":["FRF-IDRY-23-001"],"award-info":[{"award-number":["FRF-IDRY-23-001"]}],"id":[{"id":"10.13039\/501100008778","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Remote Sensing"],"abstract":"<jats:p>Recent advances in 3D scanning have enabled the digital recording of complex objects as large-scale point clouds, which require clear visualization to convey their 3D shapes effectively. Edge-highlighting visualization is used to improve the comprehensibility of complex 3D structures by enhancing the 3D edges and high-curvature regions of the scanned objects. However, traditional methods often struggle with real-world objects due to inadequate representation of soft edges (i.e., rounded edges) and excessive line clutter, impairing resolution and depth perception. To address these challenges, we propose a novel visualization method for 3D scanned point clouds based on dual 3D edge extraction and opacity\u2013color gradation. Dual 3D edge extraction separately identifies sharp and soft edges, integrating both into the visualization. Opacity\u2013color gradation enhances the clarity of fine structures within soft edges through variations in color and opacity, while also creating a halo effect that improves both resolution and depth perception of the visualized edges. Computation times required for dual 3D edge extraction are comparable to conventional binary statistical edge-extraction methods. Visualizations with opacity\u2013color gradation are executable at interactive rendering speeds. The effectiveness of the proposed method is demonstrated using 3D scanned point cloud data from high-value cultural heritage objects.<\/jats:p>","DOI":"10.3390\/rs16193750","type":"journal-article","created":{"date-parts":[[2024,10,9]],"date-time":"2024-10-09T12:22:48Z","timestamp":1728476568000},"page":"3750","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["High-Visibility Edge-Highlighting Visualization of 3D Scanned Point Clouds Based on Dual 3D Edge Extraction"],"prefix":"10.3390","volume":"16","author":[{"given":"Yuri","family":"Yamada","sequence":"first","affiliation":[{"name":"College of Information Science and Engineering, Ritsumeikan University, Ibaraki 567-8570, Osaka, Japan"}]},{"given":"Satoshi","family":"Takatori","sequence":"additional","affiliation":[{"name":"College of Information Science and Engineering, Ritsumeikan University, Ibaraki 567-8570, Osaka, Japan"}]},{"given":"Motoaki","family":"Adachi","sequence":"additional","affiliation":[{"name":"Shrewd Design Co., Ltd., Fushimi-ku, Kyoto 612-8362, Kyoto, Japan"}]},{"family":"Brahmantara","sequence":"additional","affiliation":[{"name":"Indonesian Heritage Agency, The Ministry of Education, Culture, Research, and Technology, Central Jakarta 10110, Indonesia"}]},{"given":"Kyoko","family":"Hasegawa","sequence":"additional","affiliation":[{"name":"School of Information and Telecommunication Engineering, Tokai University, Minato-ku, Tokyo 108-8619, Japan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9452-9965","authenticated-orcid":false,"given":"Liang","family":"Li","sequence":"additional","affiliation":[{"name":"College of Information Science and Engineering, Ritsumeikan University, Ibaraki 567-8570, Osaka, Japan"}]},{"given":"Jiao","family":"Pan","sequence":"additional","affiliation":[{"name":"School of Intelligence Science and Technology, University of Science and Technology Beijing, 30 Xueyuan Road, Haidian District, Beijing 100083, China"}]},{"given":"Fadjar I.","family":"Thufail","sequence":"additional","affiliation":[{"name":"Research Center for Area Studies (PRW), National Research and Innovation Agency (BRIN), Jakarta 12170, Indonesia"}]},{"given":"Hiroshi","family":"Yamaguchi","sequence":"additional","affiliation":[{"name":"Nara National Research Institute for Cultural Properties, Nijo-cho, Nara 630-8577, Nara, Japan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8033-6564","authenticated-orcid":false,"given":"Satoshi","family":"Tanaka","sequence":"additional","affiliation":[{"name":"College of Information Science and Engineering, Ritsumeikan University, Ibaraki 567-8570, Osaka, Japan"}]}],"member":"1968","published-online":{"date-parts":[[2024,10,9]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"73","DOI":"10.5194\/isprs-annals-III-5-73-2016","article-title":"See-Through Imaging of Laser-scanned 3D Cultural Heritage Objects based on Stochastic Rendering of Large-Scale Point Clouds","volume":"3","author":"Tanaka","year":"2016","journal-title":"ISPRS Ann. Photogramm. Remote. Sens. Spat. Inf. Sci."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"124","DOI":"10.1016\/j.isprsjprs.2020.01.004","article-title":"Noise-robust transparent visualization of large-scale point clouds acquired by laser scanning","volume":"161","author":"Uchida","year":"2020","journal-title":"ISPRS J. Photogramm. Remote. Sens."},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"373","DOI":"10.5194\/isprs-annals-V-2-2020-373-2020","article-title":"Opacity-based edge highlighting for transparent visualization of 3D scanned point clouds","volume":"V-2-2020","author":"Kawakami","year":"2020","journal-title":"ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci."},{"key":"ref_4","doi-asserted-by":"crossref","unstructured":"Rusu, R.B. (2013). Semantic 3D Object Maps for Everyday Robot Manipulation (Springer Tracts in Advanced Robotics 85), Springer.","DOI":"10.1007\/978-3-642-35479-3"},{"key":"ref_5","unstructured":"West, K.F., Webb, B.N., Lersch, J.R., Pothier, S., Triscari, J.M., and Iverson, A.E. (2004, January 21). Context-driven automated target detection in 3D data. Proceedings of the Proceedings Volume 5426, Automatic Target Recognition XIV, SPIE Digital Library, Orlando, FL, USA."},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"345","DOI":"10.1007\/s13218-010-0059-6","article-title":"Semantic 3D Object Maps for Everyday Manipulation in Human Living Environments","volume":"24","author":"Rusu","year":"2010","journal-title":"K\u00fcnstliche Intelligenz"},{"key":"ref_7","doi-asserted-by":"crossref","unstructured":"Toshev, A., Mordohai, P., and Taskar, B. (2010, January 13\u201318). Detecting and parsing architecture at city scale from range data. Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA.","DOI":"10.1109\/CVPR.2010.5540187"},{"key":"ref_8","first-page":"97","article-title":"Dimensioality based scale selection in 3D lidar point clouds","volume":"38","author":"Mallet","year":"2011","journal-title":"ISPRS Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci."},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"S71","DOI":"10.1016\/j.isprsjprs.2011.09.008","article-title":"Relevance assessment of full-waveform lidar data for urban area classification","volume":"66","author":"Mallet","year":"2011","journal-title":"ISPRS J. Photogramm. Remote. Sens."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"313","DOI":"10.5194\/isprsannals-II-5-W2-313-2013","article-title":"Feature Relevance Assessment for the semantic interpretation of 3D point cloud data","volume":"2","author":"Weinmann","year":"2013","journal-title":"ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci."},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"181","DOI":"10.5194\/isprsannals-II-3-181-2014","article-title":"Semantic 3D scene interpretation: A framework combining optimal neighborhood size selection with relevant features","volume":"2","author":"Weinmann","year":"2014","journal-title":"ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci."},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"195","DOI":"10.1016\/j.isprsjprs.2017.02.012","article-title":"Analytical and numerical investigations on the accuracy and robustness of geometric features extracted from 3D point cloud data","volume":"126","author":"Dittrich","year":"2017","journal-title":"ISPRS J. Photogramm. Remote. Sens."},{"key":"ref_13","first-page":"219","article-title":"A Curvature Based Adaptive Neighborhood for Individual Point Cloud Classification","volume":"42","author":"He","year":"2017","journal-title":"ISPRS Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci."},{"key":"ref_14","first-page":"4","article-title":"Nearest neighbour classification on laser point clouds to gain object structures from buildings","volume":"38","author":"Jutzi","year":"2009","journal-title":"ISPRS Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci."},{"key":"ref_15","unstructured":"Appel, A., Rohlf, F.J., and Stein, A.J. (1979, January 8\u201310). The haloed line effect for hidden line elimination. Proceedings of the ACM SIGGRAPH \u201979, Chicago, IL, USA."},{"key":"ref_16","doi-asserted-by":"crossref","first-page":"1299","DOI":"10.1109\/TVCG.2009.138","article-title":"Depth-dependent halos: Illustrative rendering of dense line data","volume":"15","author":"Everts","year":"2009","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"key":"ref_17","unstructured":"Interrante, V., and Grosch, C. (1997, January 24). Strategies for effectively visualizing 3D flow with volume LIC. Proceedings of the Proceedings. Visualization \u201997 (Cat. No. 97CB36155), Phoenix, AZ, USA."},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"664","DOI":"10.1109\/TVCG.2004.46","article-title":"Interactive volume rendering of thin thread structures within multivalued scientific datasets","volume":"10","author":"Wenger","year":"2004","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"253","DOI":"10.1109\/2945.942693","article-title":"Volume illustration: Nonphotorealistic rendering of volume models","volume":"7","author":"Rheingans","year":"2001","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"key":"ref_20","unstructured":"Svakhine, N.A., and Ebert, D.S. (2003, January 8\u201310). Interactive volume illustration and feature halos. Proceedings of the 11th Pacific Conference on Computer Graphics and Applications, Canmore, AB, Canada."},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"1344","DOI":"10.1109\/TVCG.2007.70555","article-title":"Enhancing depth-perception with flexible volumetric halos","volume":"13","author":"Bruckner","year":"2007","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Tao, Y., Lin, H., Dong, F., and Clapworthy, G. (2011, January 15\u201317). Opacity Volume based Halo Generation for Enhancing Depth Perception. Proceedings of the 12th International Conference on Computer-Aided Design and Computer Graphics, Jinan, China.","DOI":"10.1109\/CAD\/Graphics.2011.81"},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"439","DOI":"10.1016\/j.isprsjprs.2021.11.007","article-title":"3D reconstruction of Borobudur reliefs from 2D monocular photographs based on soft-edge enhanced deep learning","volume":"183","author":"Pan","year":"2022","journal-title":"ISPRS J. Photogramm. Remote. Sens."}],"container-title":["Remote Sensing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2072-4292\/16\/19\/3750\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T16:09:56Z","timestamp":1760112596000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2072-4292\/16\/19\/3750"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,9]]},"references-count":23,"journal-issue":{"issue":"19","published-online":{"date-parts":[[2024,10]]}},"alternative-id":["rs16193750"],"URL":"https:\/\/doi.org\/10.3390\/rs16193750","relation":{},"ISSN":["2072-4292"],"issn-type":[{"value":"2072-4292","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,10,9]]}}}