{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,25]],"date-time":"2026-03-25T18:19:24Z","timestamp":1774462764640,"version":"3.50.1"},"reference-count":23,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2026,3,1]],"date-time":"2026-03-01T00:00:00Z","timestamp":1772323200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T00:00:00Z","timestamp":1772841600000},"content-version":"vor","delay-in-days":6,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100000780","name":"European Commission","doi-asserted-by":"crossref","award":["886977"],"award-info":[{"award-number":["886977"]}],"id":[{"id":"10.13039\/501100000780","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2026,3]]},"abstract":"<jats:title>Abstract<\/jats:title>\n                  <jats:p>Foreign object detection (FOD) is a crucial concern in various industries, including aerospace, automotive, and manufacturing, where undetected foreign objects can cause severe damage and safety risks. Traditional FOD detection systems rely on visual inspection, radar, X-ray imaging, or acoustic sensing, but face challenges in complex and dynamic environments. In this work, we propose a novel methodology that utilizes neural radiance fields (NeRF) as a comparative baseline for foreign object detection in industrial settings. Our approach consists of two main phases: First, a photorealistic 3D reconstruction of the environment is generated using NeRF when no foreign objects are present. Then, in the detection phase, new images of the environment are captured, and their viewpoints are matched with the NeRF-rendered scene to identify discrepancies using a deep learning-based comparative model. The system utilizes a modified SNUNet-CD architecture, enhancing attention mechanisms for improved accuracy in detecting semantic changes. Experimental validation in a controlled environment demonstrates high detection accuracy, successfully identifying foreign objects of varying sizes, materials, and placements.<\/jats:p>","DOI":"10.1007\/s00371-026-04402-4","type":"journal-article","created":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T12:43:42Z","timestamp":1772887422000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Foreign object detection using neural radiance field as comparative baseline"],"prefix":"10.1007","volume":"42","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3505-6517","authenticated-orcid":false,"given":"Claudio","family":"Lopez","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9476-2804","authenticated-orcid":false,"given":"Martxel","family":"Eizaguirre","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0530-0275","authenticated-orcid":false,"given":"Aiert","family":"Amundarain","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,3,7]]},"reference":[{"issue":"7","key":"4402_CR1","doi-asserted-by":"publisher","first-page":"2463","DOI":"10.3390\/s22072463","volume":"22","author":"Y Jing","year":"2022","unstructured":"Jing, Y., Zheng, H., Lin, C., Zheng, W., Dong, K., Li, X.: Foreign object debris detection for optical imaging sensors based on random forest. Sensors 22(7), 2463 (2022). https:\/\/doi.org\/10.3390\/s22072463","journal-title":"Sensors"},{"key":"4402_CR2","doi-asserted-by":"publisher","first-page":"997","DOI":"10.3390\/s21030997","volume":"21","author":"J Zhong","year":"2021","unstructured":"Zhong, J., Gou, X., Shu, Q., Liu, X., Zeng, Q.: A FOD detection approach on millimeter-wave radar sensors based on optimal VMD and SVDD. Sensors 21, 997 (2021). https:\/\/doi.org\/10.3390\/s21030997","journal-title":"Sensors"},{"key":"4402_CR3","doi-asserted-by":"publisher","unstructured":"Kwon, J., Lee, J., Kim, W.: Real-time detection of foreign objects using X-ray imaging for dry food manufacturing line. In: IEEE International Symposium on Consumer Electronics (ISCE). IEEE, Albufeira, Portugal (2008). https:\/\/doi.org\/10.1109\/ISCE.2008.4559552","DOI":"10.1109\/ISCE.2008.4559552"},{"key":"4402_CR4","doi-asserted-by":"crossref","unstructured":"Cardwell, D., Chana, K., Gilboy, M.: The development and testing of a gas turbine engine foreign object damage (FOD) detection system. In: ASME Turbo Expo: Power for Land. Sea, and Air, vol. 43987, pp. 331\u2013342. ASME, Glasgow, UK (2010)","DOI":"10.1115\/GT2010-23478"},{"key":"4402_CR5","doi-asserted-by":"publisher","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: NeRF: representing scenes as neural radiance fields for view synthesis. Commun. ACM 65(1), 99\u2013106 (2021). https:\/\/doi.org\/10.1145\/3503250","DOI":"10.1145\/3503250"},{"key":"4402_CR6","doi-asserted-by":"crossref","unstructured":"Fridovich-Keil, S., Yu, A., Tancik, M., Chen, Q., Recht, B., Kanazawa, A.: Plenoxels: radiance fields without neural networks. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5491\u20135500. IEEE, New Orleans, LA, USA (2022)","DOI":"10.1109\/CVPR52688.2022.00542"},{"key":"4402_CR7","doi-asserted-by":"publisher","unstructured":"Yu, A., Ye, V., Tancik, M., Kanazawa, A.: pixelNeRF: neural radiance fields from one or few images. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4576\u20134585. IEEE, Nashville, TN, USA (2021). https:\/\/doi.org\/10.1109\/CVPR46437.2021.00455","DOI":"10.1109\/CVPR46437.2021.00455"},{"key":"4402_CR8","doi-asserted-by":"publisher","unstructured":"Sun, C., Sun, M., Chen, H.: Direct voxel grid optimization: super-fast convergence for radiance fields reconstruction. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5449\u20135459. IEEE, New Orleans, LA, USA (2022). https:\/\/doi.org\/10.1109\/CVPR52688.2022.00538","DOI":"10.1109\/CVPR52688.2022.00538"},{"key":"4402_CR9","doi-asserted-by":"crossref","unstructured":"Chen, A., Xu, Z., Geiger, A., Yu, J., Su, H.: TensoRF: tensorial radiance fields. In: Computer Vision - ECCV 2022, pp. 333\u2013350. Springer, Tel Aviv, Israel (2022)","DOI":"10.1007\/978-3-031-19824-3_20"},{"issue":"4","key":"4402_CR10","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3528223.3530127","volume":"41","author":"T Muller","year":"2022","unstructured":"Muller, T., Evans, A., Schied, C., Keller, A.: Instant neural graphics primitives with a multiresolution hash encoding. ACM Trans. Graph. 41(4), 1\u201315 (2022). https:\/\/doi.org\/10.1145\/3528223.3530127","journal-title":"ACM Trans. Graph."},{"key":"4402_CR11","doi-asserted-by":"publisher","unstructured":"Barron, J.T., Mildenhall, B., Tancik, M., Hedman, P., Martin-Brualla, R., Srinivasan, P.P.: Mip-NeRF: a multiscale representation for anti-aliasing neural radiance fields. In: IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 5835\u20135844. IEEE, Montreal, QC, Canada (2021). https:\/\/doi.org\/10.1109\/ICCV48922.2021.00580","DOI":"10.1109\/ICCV48922.2021.00580"},{"key":"4402_CR12","doi-asserted-by":"crossref","unstructured":"Barron, J.T., Mildenhall, B., Verbin, D., Srinivasan, P.P., Hedman, P.: Mip-NeRF 360: unbounded anti-aliased neural radiance fields. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5460\u20135469. IEEE, New Orleans, LA, USA (2022)","DOI":"10.1109\/CVPR52688.2022.00539"},{"key":"4402_CR13","doi-asserted-by":"crossref","unstructured":"Barron, J.T., Mildenhall, B., Verbin, D., Srinivasan, P.P., Hedman, P.: Zip-NeRF: anti-aliased grid-based neural radiance fields. In: IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 19640\u201319648. IEEE, Paris, France (2023)","DOI":"10.1109\/ICCV51070.2023.01804"},{"issue":"4","key":"4402_CR14","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3592433","volume":"42","author":"B Kerbl","year":"2023","unstructured":"Kerbl, B., Kopanas, G., Leimkuehler, T., Drettakis, G.: 3D Gaussian splatting for real-time radiance field rendering. ACM Trans. Graph. 42(4), 1\u201314 (2023). https:\/\/doi.org\/10.1145\/3592433","journal-title":"ACM Trans. Graph."},{"issue":"3","key":"4402_CR15","doi-asserted-by":"publisher","first-page":"294","DOI":"10.1109\/TIP.2004.838698","volume":"14","author":"RJ Radke","year":"2005","unstructured":"Radke, R.J., Andra, S., Al-Kofahi, O., Roysam, B.: Image change detection algorithms: a systematic survey. IEEE Trans. Image Process. 14(3), 294\u2013307 (2005)","journal-title":"IEEE Trans. Image Process."},{"key":"4402_CR16","doi-asserted-by":"crossref","unstructured":"Caye Daudt, R., Le Saux, B., Boulch, A.: Fully convolutional Siamese networks for change detection. In: IEEE International Conference on Image Processing (ICIP), pp. 4063\u20134067. IEEE, Athens, Greece (2018)","DOI":"10.1109\/ICIP.2018.8451652"},{"key":"4402_CR17","first-page":"1","volume":"19","author":"S Fang","year":"2022","unstructured":"Fang, S., Li, K., Shao, J., Li, Z.: SNUNet-CD: a densely connected Siamese network for change detection of VHR images. IEEE Geosci. Remote Sens. Lett. 19, 1\u20135 (2022)","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"4402_CR18","first-page":"1323","volume-title":"IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS)","author":"L Yen-Chen","year":"2021","unstructured":"Yen-Chen, L., Florence, P., Barron, J.T., Rodriguez, A., Isola, P., Lin, T.: iNeRF: inverting neural radiance fields for pose estimation. In: IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 1323\u20131330. IEEE, Prague, Czech Republic (2021)"},{"key":"4402_CR19","doi-asserted-by":"crossref","unstructured":"Schonberger, J.L., Frahm, J.: Structure-from-motion revisited. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4104\u20134113. IEEE, Las Vegas, NV, USA (2016)","DOI":"10.1109\/CVPR.2016.445"},{"key":"4402_CR20","doi-asserted-by":"crossref","unstructured":"Kendall, A., Grimes, M., Cipolla, R.: PoseNet: A convolutional network for real-time 6-DOF camera relocalization. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 2938\u20132946 (2015)","DOI":"10.1109\/ICCV.2015.336"},{"key":"4402_CR21","doi-asserted-by":"publisher","first-page":"104032","DOI":"10.1016\/j.dsp.2023.104032","volume":"137","author":"X Zhang","year":"2023","unstructured":"Zhang, X., Liu, G., Huang, L., Ren, Q., Bavirisetti, D.P.: IVOMFuse: an image fusion method based on infrared-to-visible object mapping. Digital Signal Process. 137, 104032 (2023). https:\/\/doi.org\/10.1016\/j.dsp.2023.104032","journal-title":"Digital Signal Process."},{"key":"4402_CR22","doi-asserted-by":"publisher","first-page":"104801","DOI":"10.1016\/j.dsp.2024.104801","volume":"156","author":"X Zhang","year":"2025","unstructured":"Zhang, X., Liu, G., Li, M., Ren, Q., Tang, H., Bavirisetti, D.P.: FusionNGFPE: an image fusion approach driven by non-global fuzzy pre-enhancement framework. Digi. Signal Process. 156, 104801 (2025). https:\/\/doi.org\/10.1016\/j.dsp.2024.104801","journal-title":"Digi. Signal Process."},{"issue":"2","key":"4402_CR23","doi-asserted-by":"publisher","first-page":"291","DOI":"10.1109\/TG.2023.3263001","volume":"16","author":"X Gu","year":"2024","unstructured":"Gu, X., Liu, G., Zhang, X., Tang, L., Zhou, X., Qiu, W.: Infrared-visible synthetic data from game engine for image fusion improvement. IEEE Trans. Games 16(2), 291\u2013302 (2024). https:\/\/doi.org\/10.1109\/TG.2023.3263001","journal-title":"IEEE Trans. Games"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-026-04402-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-026-04402-4","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-026-04402-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,25]],"date-time":"2026-03-25T16:20:04Z","timestamp":1774455604000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-026-04402-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,3]]},"references-count":23,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2026,3]]}},"alternative-id":["4402"],"URL":"https:\/\/doi.org\/10.1007\/s00371-026-04402-4","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,3]]},"assertion":[{"value":"22 October 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 February 2026","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 March 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest. The authors declare they have no financial interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"192"}}