{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,2]],"date-time":"2025-10-02T13:43:48Z","timestamp":1759412628229,"version":"3.37.3"},"reference-count":27,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2023,9,26]],"date-time":"2023-09-26T00:00:00Z","timestamp":1695686400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,9,26]],"date-time":"2023-09-26T00:00:00Z","timestamp":1695686400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001868","name":"National Science Council","doi-asserted-by":"publisher","award":["110-2221-E-027-079-MY2"],"award-info":[{"award-number":["110-2221-E-027-079-MY2"]}],"id":[{"id":"10.13039\/501100001868","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2024,6]]},"DOI":"10.1007\/s00371-023-03077-5","type":"journal-article","created":{"date-parts":[[2023,9,26]],"date-time":"2023-09-26T09:02:27Z","timestamp":1695718947000},"page":"4201-4220","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Reinforcement learning-based approach for plastic texture surface defects inspection"],"prefix":"10.1007","volume":"40","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-7315-6116","authenticated-orcid":false,"given":"Chao-Ching","family":"Ho","sequence":"first","affiliation":[]},{"given":"Yuan-Cheng","family":"Chiao","sequence":"additional","affiliation":[]},{"given":"Eugene","family":"Su","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,9,26]]},"reference":[{"key":"3077_CR1","doi-asserted-by":"publisher","first-page":"348","DOI":"10.1016\/j.asoc.2016.10.030","volume":"52","author":"C Jian","year":"2017","unstructured":"Jian, C., Gao, J., Ao, Y.: Automatic surface defect detection for mobile phone screen glass based on machine vision. Appl. Soft Comput. 52, 348\u2013358 (2017)","journal-title":"Appl. Soft Comput."},{"issue":"3","key":"3077_CR2","doi-asserted-by":"publisher","first-page":"1","DOI":"10.5565\/rev\/elcvia.268","volume":"7","author":"X Xie","year":"2008","unstructured":"Xie, X.: A review of recent advances in surface defect detection using texture analysis techniques. ELCVIA: Electron. Lett. Comput. Vision Image Anal. 7(3), 1\u201322 (2008)","journal-title":"ELCVIA: Electron. Lett. Comput. Vision Image Anal."},{"issue":"9","key":"3077_CR3","doi-asserted-by":"publisher","first-page":"1575","DOI":"10.3390\/app8091575","volume":"8","author":"X Tao","year":"2018","unstructured":"Tao, X., Zhang, D., Ma, W., Liu, X., Xu, D.: Automatic metallic surface defect detection and recognition with convolutional neural networks. Appl. Sci. 8(9), 1575 (2018)","journal-title":"Appl. Sci."},{"doi-asserted-by":"crossref","unstructured":"Aittala, M. Weyrich, T. and Lehtinen, J.: Two-shot SVBRDF capture for stationary materials, ACM Trans. Graph: Association for Computing Machinery Transactions on Graphics Archive, 34(4), 110\u2013111, (2015).","key":"3077_CR4","DOI":"10.1145\/2766967"},{"doi-asserted-by":"crossref","unstructured":"Li, Z., Sunkavalli, K. and Chandraker, M.: Materials for masses: SVBRDF acquisition with a single mobile phone image, European Conference on Computer Vision (ECCV). pp. 72\u201387, (2018).","key":"3077_CR5","DOI":"10.1007\/978-3-030-01219-9_5"},{"doi-asserted-by":"crossref","unstructured":"Yu, Y., Debevec, P., Malik, J. and Hawkins, T.: Inverse global illumination: recovering reflectance models of real scenes from photographs, the 26th annual conference on Computer graphics and interactive techniques, pp. 215\u2013224, (1999).","key":"3077_CR6","DOI":"10.1145\/311535.311559"},{"issue":"3","key":"3077_CR7","first-page":"334","volume":"22","author":"L Wang","year":"2003","unstructured":"Wang, L., Wang, X., Tong, X., Lin, S., Hu, S., Guo, B., Shum, H.-Y.: View-dependent displacement mapping. Assoc. Comput. Mach. Transact. Gr. (TOG) 22(3), 334\u2013339 (2003)","journal-title":"Assoc. Comput. Mach. Transact. Gr. (TOG)"},{"doi-asserted-by":"crossref","unstructured":"Mujeeb, A., Dai, W., Erdt, M., and Sourin, A., Unsupervised surface defect detection using deep autoencoders and data augmentation, 2018 International Conference on Cyberworlds (CW). IEEE, pp. 391\u2013398, (2018).","key":"3077_CR8","DOI":"10.1109\/CW.2018.00076"},{"doi-asserted-by":"crossref","unstructured":"Chen, Y.-F., Yang, F.-H., Suc, E. and Ho, C.-C., Automatic defect detection system based on deep convolutional neural networks, International Conference on Engineering, Science, and Industrial Applications (ICESI), 22\u201324, Tokyo, Japan (2019).","key":"3077_CR9","DOI":"10.1109\/ICESI.2019.8863029"},{"issue":"15","key":"3077_CR10","doi-asserted-by":"publisher","first-page":"3159","DOI":"10.3390\/app9153159","volume":"9","author":"F Zhou","year":"2019","unstructured":"Zhou, F., Liu, G., Xu, F., Deng, H.: A generic automated surface defect detection based on a bilinear model. Appl. Sci. 9(15), 3159 (2019)","journal-title":"Appl. Sci."},{"doi-asserted-by":"crossref","unstructured":"Inoue, T., Chaudhury, S., Magistris, G. D. and Dasgupta, S.: Transfer learning from synthetic to real images using variational autoencoders for robotic applications, Computer Vision and Pattern Recognition arXiv:1709.06762, (2017).","key":"3077_CR11","DOI":"10.1109\/ICIP.2018.8451064"},{"doi-asserted-by":"crossref","unstructured":"Chen, J., Hu, K., Yang, Y., Liu, Y. and Xuan, Q.: Collective transfer learning for defect prediction, Neurocomputing, vol. 25, (2019).","key":"3077_CR12","DOI":"10.1109\/ACCESS.2018.2890733"},{"doi-asserted-by":"crossref","unstructured":"Shrivastava, A., Pfister, T., Tuzel, O., Susskind, J., Wang, W. and Webb, R. Learning from simulated and unsupervised images through adversarial training, IEEE conference on computer vision and pattern recognition, pp. 2107\u20132116, (2017).","key":"3077_CR13","DOI":"10.1109\/CVPR.2017.241"},{"doi-asserted-by":"crossref","unstructured":"Tremblay, J., Prakash, A., Acuna, D., Brophy, M., Jampani, V., Anil, C., To, T., Cameracci, E., Boochoon, S. and Birchfield, S.: Training deep networks with synthetic data: bridging the reality gap by domain randomization, IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 969\u2013977, (2018).","key":"3077_CR14","DOI":"10.1109\/CVPRW.2018.00143"},{"doi-asserted-by":"crossref","unstructured":"Ho, C.-C., Hernandez, M. A. B., Chen, Y.-F., Lin, C.-J. and Chen, C.-S.: Deep residual neural network-based defect detection on complex backgrounds. IEEE Transactions on Instrumentation & Measurement, vol. 71, pp. 5005210, (2022).","key":"3077_CR15","DOI":"10.1109\/TIM.2022.3144224"},{"unstructured":"Marceau, L., Qiu, L., Vandewiele, N. and Charton, E. J. A. P. A.: A comparison of Deep Learning performances with other machine learning algorithms on credit scoring unbalanced data, arXiv preprint arXiv:.12363, (2019).","key":"3077_CR16"},{"doi-asserted-by":"crossref","unstructured":"Lin, L. and Guo, S.-X.: Text classification feature extraction method based on deep learning for unbalanced data sets, in International Conference on Advanced Hybrid Information Processing, pp. 320\u2013331: Springer (2020).","key":"3077_CR17","DOI":"10.1007\/978-3-030-67871-5_29"},{"doi-asserted-by":"crossref","unstructured":"Zhu, J.-Y., Park, T., Isola, P. and Efros, A. A. Unpaired image-to-image translation using cycle-consistent adversarial networks, IEEE international conference on computer vision, pp. 2223\u20132232, (2017).","key":"3077_CR18","DOI":"10.1109\/ICCV.2017.244"},{"doi-asserted-by":"crossref","unstructured":"Huang, C.-C. and Lin X.-P., Study on machine learning based intelligent defect detection system, in MATEC Web of Conferences, vol. 201, p. 01010: EDP Sciences (2018).","key":"3077_CR19","DOI":"10.1051\/matecconf\/201820101010"},{"doi-asserted-by":"crossref","unstructured":"Bi, S., Sunkavalli, K., Perazzi, F., Shechtman, E., Kim, V. G., and Ramamoorthi, R.: Deep cg2real: synthetic-to-real translation via image disentanglement, in Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 2730\u20132739 (2019).","key":"3077_CR20","DOI":"10.1109\/ICCV.2019.00282"},{"issue":"12","key":"3077_CR21","doi-asserted-by":"publisher","first-page":"5645","DOI":"10.1109\/TNNLS.2020.2969327","volume":"31","author":"S Alonso-Monsalve","year":"2020","unstructured":"Alonso-Monsalve, S., Whitehead, L.H.: Image-based model parameter optimization using model-assisted generative adversarial networks. IEEE Transact. Neural Netw. Learn. Syst. 31(12), 5645\u20135650 (2020)","journal-title":"IEEE Transact. Neural Netw. Learn. Syst."},{"doi-asserted-by":"crossref","unstructured":"Wang, H., Liang, W., Shen, J., Van Gool, L. and Wang, W.: Counterfactual cycle-consistent learning for instruction following and generation in vision-language navigation,\" in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1\u201315 (2022).","key":"3077_CR22","DOI":"10.1109\/CVPR52688.2022.01503"},{"unstructured":"Jiang, L., Dai, B., Wu, W., and Loy, C. C.: Deceive D: adaptive pseudo augmentation for GAN training with limited data, in Proceedings of the Conference on Neural Information Processing Systems (NeurIPS), pp. 21655\u201321667 (2021).","key":"3077_CR23"},{"key":"3077_CR24","doi-asserted-by":"publisher","first-page":"108","DOI":"10.1016\/j.future.2021.04.017","volume":"123","author":"G Andresini","year":"2021","unstructured":"Andresini, G., Appice, A., De Rose, L., Malerba, D.: GAN augmentation to deal with imbalance in imaging-based intrusion detection. Futur. Gener. Comput. Syst. 123, 108\u2013127 (2021)","journal-title":"Futur. Gener. Comput. Syst."},{"doi-asserted-by":"crossref","unstructured":"Zhang, S. Yu, D., Zhou, Y., Wu, Y. and Ma, Y.: Enhanced visual perception for underwater images based on multistage generative adversarial network, The Visual Computer, pp. 1\u201313, 2022.","key":"3077_CR25","DOI":"10.1007\/s00371-022-02665-1"},{"doi-asserted-by":"crossref","unstructured":"Sharma, V., Tripathi, A. K., Mittal, H., Parmar, A., Soni, A., and Amarwal, R. WeedGAN: a novel generative adversarial network for cotton weed identification, The Visual Computer, pp. 1\u201317, (2022).","key":"3077_CR26","DOI":"10.1007\/s00371-022-02742-5"},{"doi-asserted-by":"crossref","unstructured":"Gutierrez, J., Rabin, J., Galerne, B. and Hurtut, T.: On demand solid texture synthesis using deep 3d networks, in Computer Graphics Forum, vol. 39, no. 1, pp. 511\u2013530: Wiley Online Library (2020).","key":"3077_CR27","DOI":"10.1111\/cgf.13889"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-03077-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-023-03077-5\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-03077-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,6,6]],"date-time":"2024-06-06T11:13:48Z","timestamp":1717672428000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-023-03077-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,9,26]]},"references-count":27,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2024,6]]}},"alternative-id":["3077"],"URL":"https:\/\/doi.org\/10.1007\/s00371-023-03077-5","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"type":"print","value":"0178-2789"},{"type":"electronic","value":"1432-2315"}],"subject":[],"published":{"date-parts":[[2023,9,26]]},"assertion":[{"value":"18 August 2023","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 September 2023","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}