{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,3]],"date-time":"2026-04-03T14:19:00Z","timestamp":1775225940986,"version":"3.50.1"},"reference-count":36,"publisher":"MDPI AG","issue":"8","license":[{"start":{"date-parts":[[2017,8,16]],"date-time":"2017-08-16T00:00:00Z","timestamp":1502841600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Remote Sensing"],"abstract":"<jats:p>This paper presents a novel method for fully automatic and convenient extrinsic calibration of a 3D LiDAR and a panoramic camera with a normally printed chessboard. The proposed method is based on the 3D corner estimation of the chessboard from the sparse point cloud generated by one frame scan of the LiDAR. To estimate the corners, we formulate a full-scale model of the chessboard and fit it to the segmented 3D points of the chessboard. The model is fitted by optimizing the cost function under constraints of correlation between the reflectance intensity of laser and the color of the chessboard\u2019s patterns. Powell\u2019s method is introduced for resolving the discontinuity problem in optimization. The corners of the fitted model are considered as the 3D corners of the chessboard. Once the corners of the chessboard in the 3D point cloud are estimated, the extrinsic calibration of the two sensors is converted to a 3D-2D matching problem. The corresponding 3D-2D points are used to calculate the absolute pose of the two sensors with Unified Perspective-n-Point (UPnP). Further, the calculated parameters are regarded as initial values and are refined using the Levenberg-Marquardt method. The performance of the proposed corner detection method from the 3D point cloud is evaluated using simulations. The results of experiments, conducted on a Velodyne HDL-32e LiDAR and a Ladybug3 camera under the proposed re-projection error metric, qualitatively and quantitatively demonstrate the accuracy and stability of the final extrinsic calibration parameters.<\/jats:p>","DOI":"10.3390\/rs9080851","type":"journal-article","created":{"date-parts":[[2017,8,16]],"date-time":"2017-08-16T10:43:51Z","timestamp":1502880231000},"page":"851","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":96,"title":["Reflectance Intensity Assisted Automatic and Accurate Extrinsic Calibration of 3D LiDAR and Panoramic Camera Using a Printed Chessboard"],"prefix":"10.3390","volume":"9","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-6557-7175","authenticated-orcid":false,"given":"Weimin","family":"Wang","sequence":"first","affiliation":[{"name":"Graduate School of Engineering, Nagoya University, Furo-cho, Chikusa-ku, Nagoya 464-8601, Japan"}]},{"given":"Ken","family":"Sakurada","sequence":"additional","affiliation":[{"name":"Graduate School of Engineering, Nagoya University, Furo-cho, Chikusa-ku, Nagoya 464-8601, Japan"}]},{"given":"Nobuo","family":"Kawaguchi","sequence":"additional","affiliation":[{"name":"Graduate School of Engineering, Nagoya University, Furo-cho, Chikusa-ku, Nagoya 464-8601, Japan"},{"name":"Institute of Innovation for Future Society, Nagoya University, Furo-cho, Chikusa-ku, Nagoya 464-8601, Japan"}]}],"member":"1968","published-online":{"date-parts":[[2017,8,16]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","unstructured":"Mastin, A., Kepner, J., and Fisher, J. (2009, January 20\u201325). Automatic registration of LIDAR and optical images of urban scenes. Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Miami, FL, USA.","DOI":"10.1109\/CVPR.2009.5206539"},{"key":"ref_2","first-page":"69","article-title":"Stereopolis II: A multi-purpose and multi-sensor 3d mobile mapping system for street visualisation and 3d metrology","volume":"200","author":"Paparoditis","year":"2012","journal-title":"Rev. Fr. Photogramm. T\u00e9l\u00e9d\u00e9tec."},{"key":"ref_3","unstructured":"Szarvas, M., Sakai, U., and Ogata, J. (2006, January 13\u201315). Real-time pedestrian detection using LIDAR and convolutional neural networks. Proceedings of the 2006 IEEE Intelligent Vehicles Symposium (IV), Tokyo, Japan."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"371","DOI":"10.1177\/0278364912470012","article-title":"Fusing LIDAR, camera and semantic information: A context-based approach for pedestrian detection","volume":"32","author":"Premebida","year":"2013","journal-title":"Int. J. Robot. Res. (IJRR)"},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"Schlosser, J., Chow, C.K., and Kira, Z. (2016, January 16\u201321). Fusing LIDAR and images for pedestrian detection using convolutional neural networks. Proceedings of the 2016 IEEE International Conference on Robotics and Automation (ICRA), Stockholm, Sweden.","DOI":"10.1109\/ICRA.2016.7487370"},{"key":"ref_6","unstructured":"Levinson, J., and Thrun, S. (2017, April 27). Automatic Online Calibration of Cameras and Lasers. Available online: http:\/\/roboticsproceedings.org\/rss09\/p29.pdf."},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"675","DOI":"10.1002\/rob.21523","article-title":"Multi-modal sensor calibration using a gradient orientation measure","volume":"32","author":"Taylor","year":"2014","journal-title":"J. Field Robot. (JFR)"},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"696","DOI":"10.1002\/rob.21542","article-title":"Automatic extrinsic calibration of vision and lidar by maximizing mutual information","volume":"32","author":"Pandey","year":"2014","journal-title":"J. Field Robot. (JFR)"},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"452","DOI":"10.1177\/0278364911435689","article-title":"3D LIDAR\u2013camera intrinsic and extrinsic calibration: Identifiability and analytical least-squares-based initialization","volume":"31","author":"Mirzaei","year":"2012","journal-title":"Int. J. Robot. Res. (IJRR)"},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"5333","DOI":"10.3390\/s140305333","article-title":"Calibration between color camera and 3D LIDAR instruments with a polygonal planar board","volume":"14","author":"Park","year":"2014","journal-title":"Sensors"},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"782","DOI":"10.1016\/j.robot.2014.02.004","article-title":"Error propagation and uncertainty analysis between 3D laser scanner and camera","volume":"62","year":"2014","journal-title":"Robot. Auton. Syst."},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"155","DOI":"10.1093\/comjnl\/7.2.155","article-title":"An efficient method for finding the minimum of a function of several variables without calculating derivatives","volume":"7","author":"Powell","year":"1964","journal-title":"Comput. J."},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Kneip, L., Li, H., and Seo, Y. (2014, January 6\u201312). UPnP: An optimal O(n) solution to the absolute pose problem with universal applicability. Proceedings of the European Conference on Computer Vision (ECCV), Zurich, Switzerland.","DOI":"10.1007\/978-3-319-10590-1_9"},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"164","DOI":"10.1090\/qam\/10666","article-title":"A method for the solution of certain non-linear problems in least squares","volume":"2","author":"Levenberg","year":"1944","journal-title":"Q. Appl. Math."},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"431","DOI":"10.1137\/0111030","article-title":"An algorithm for least-squares estimation of nonlinear parameters","volume":"11","author":"Marquardt","year":"1963","journal-title":"J. Soc. Ind. Appl. Math."},{"key":"ref_16","unstructured":"Zhang, Q., and Pless, R. (October, January 28). Extrinsic calibration of a camera and laser range finder (improves camera calibration). Proceedings of the 2004 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), Sendai, Japan."},{"key":"ref_17","unstructured":"Unnikrishnan, R., and Hebert, M. (2017, April 27). Fast Extrinsic Calibration of a Laser Rangefinder to a Camera. Available online: http:\/\/repository.cmu.edu\/robotics\/339\/."},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"336","DOI":"10.3182\/20100906-3-IT-2019.00059","article-title":"Extrinsic calibration of a 3D laser scanner and an omnidirectional camera","volume":"43","author":"Pandey","year":"2010","journal-title":"IFAC Proc. Vol."},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"1543","DOI":"10.1177\/0278364911400640","article-title":"Ford Campus vision and lidar data set","volume":"30","author":"Pandey","year":"2011","journal-title":"Int. J. Robot. Res. (IJRR)"},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Scaramuzza, D., Harati, A., and Siegwart, R. (November, January 29). Extrinsic self calibration of a camera and a 3D laser range finder from natural scenes. Proceedings of the 2007 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), San Diego, CA, USA.","DOI":"10.1109\/IROS.2007.4399276"},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"Moghadam, P., Bosse, M., and Zlot, R. (2013, January 6\u201310). Line-based extrinsic calibration of range and image sensors. Proceedings of the 2013 IEEE International Conference on Robotics and Automation, Karlsruhe, Germany.","DOI":"10.1109\/ICRA.2013.6631095"},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"1902","DOI":"10.3390\/s130201902","article-title":"3D LIDAR-camera extrinsic calibration using an arbitrary trihedron","volume":"13","author":"Gong","year":"2013","journal-title":"Sensors"},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Geiger, A., Moosmann, F., Car, O., and Schuster, B. (2012, January 14\u201318). Automatic camera and range sensor calibration using a single shot. Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Saint Paul, MN, USA.","DOI":"10.1109\/ICRA.2012.6224570"},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"1231","DOI":"10.1177\/0278364913491297","article-title":"Vision meets robotics: The KITTI dataset","volume":"32","author":"Geiger","year":"2013","journal-title":"Int. J. Robot. Res. (IJRR)"},{"key":"ref_25","doi-asserted-by":"crossref","first-page":"59","DOI":"10.5772\/50900","article-title":"LIDAR velodyne HDL-64E calibration using pattern planes","volume":"8","year":"2011","journal-title":"Int. J. Adv. Robot. Syst."},{"key":"ref_26","doi-asserted-by":"crossref","first-page":"1330","DOI":"10.1109\/34.888718","article-title":"A flexible new technique for camera calibration","volume":"22","author":"Zhang","year":"2000","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI)"},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Scaramuzza, D., Martinelli, A., and Siegwart, R. (2006, January 9\u201315). A toolbox for easily calibrating omnidirectional cameras. Proceedings of the 2006 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), Beijing, China.","DOI":"10.1109\/IROS.2006.282372"},{"key":"ref_28","first-page":"248","article-title":"Segmentation of point clouds using smoothness constraint","volume":"36","author":"Rabbani","year":"2006","journal-title":"Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. (ISPRS)"},{"key":"ref_29","doi-asserted-by":"crossref","first-page":"381","DOI":"10.1145\/358669.358692","article-title":"Random Sample Consensus: A paradigm for model fitting with applications to image analysis and automated cartography","volume":"24","author":"Fischler","year":"1981","journal-title":"Commun. ACM"},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Wang, W., Sakurada, K., and Kawaguchi, N. (2016). Incremental and enhanced scanline-based segmentation method for surface reconstruction of sparse LiDAR data. Remote Sens., 8.","DOI":"10.3390\/rs8110967"},{"key":"ref_31","unstructured":"Velodyne LiDAR, Inc. (2012). HDL-32E User\u2019s Manual, Velodyne LiDAR, Inc."},{"key":"ref_32","unstructured":"Jolliffe, I. (2002). Principal Component Analysis, Wiley Online Library."},{"key":"ref_33","unstructured":"Jones, E., Oliphant, T., and Peterson, P. (2017, April 27). SciPy: Open Source Scientific Tools for Python. Available online: http:\/\/www.scipy.org."},{"key":"ref_34","doi-asserted-by":"crossref","unstructured":"Rufli, M., Scaramuzza, D., and Siegwart, R. (2008, January 22\u201326). Automatic detection of checkerboards on blurred and distorted images. Proceedings of the IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), Nice, France.","DOI":"10.1109\/IROS.2008.4650703"},{"key":"ref_35","doi-asserted-by":"crossref","unstructured":"Kneip, L., and Furgale, P. (June, January 31). OpenGV: A unified and generalized approach to real-time calibrated geometric vision. Proceedings of the 2014 IEEE International Conference on Robotics and Automation (ICRA), Hong Kong, China.","DOI":"10.1109\/ICRA.2014.6906582"},{"key":"ref_36","unstructured":"Point Grey Research, Inc. (2012). Technical Application Note (TAN2012009): Geometric Vision Using Ladybug\u00ae Cameras, Point Grey Research, Inc."}],"container-title":["Remote Sensing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2072-4292\/9\/8\/851\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T18:42:29Z","timestamp":1760208149000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2072-4292\/9\/8\/851"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2017,8,16]]},"references-count":36,"journal-issue":{"issue":"8","published-online":{"date-parts":[[2017,8]]}},"alternative-id":["rs9080851"],"URL":"https:\/\/doi.org\/10.3390\/rs9080851","relation":{},"ISSN":["2072-4292"],"issn-type":[{"value":"2072-4292","type":"electronic"}],"subject":[],"published":{"date-parts":[[2017,8,16]]}}}