{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,25]],"date-time":"2026-03-25T16:07:58Z","timestamp":1774454878647,"version":"3.50.1"},"publisher-location":"Cham","reference-count":29,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783030720728","type":"print"},{"value":"9783030720735","type":"electronic"}],"license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021]]},"DOI":"10.1007\/978-3-030-72073-5_2","type":"book-chapter","created":{"date-parts":[[2021,3,17]],"date-time":"2021-03-17T18:02:58Z","timestamp":1616004178000},"page":"13-25","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":25,"title":["Traffic-Sign Recognition Using Deep Learning"],"prefix":"10.1007","author":[{"given":"Zhongbing","family":"Qin","sequence":"first","affiliation":[]},{"given":"Wei Qi","family":"Yan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,3,18]]},"reference":[{"issue":"4","key":"2_CR1","doi-asserted-by":"publisher","first-page":"1484","DOI":"10.1109\/TITS.2012.2209421","volume":"13","author":"A Mogelmose","year":"2012","unstructured":"Mogelmose, A., Trivedi, M., Moeslund, T.B.: Vision-based traffic sign detection and analysis for intelligent driver assistance systems: perspectives and survey. IEEE Trans. Intell. Transp. Syst. 13(4), 1484\u20131497 (2012)","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"2_CR2","doi-asserted-by":"publisher","first-page":"758","DOI":"10.1016\/j.neucom.2016.07.009","volume":"214","author":"Y Zhu","year":"2016","unstructured":"Zhu, Y., Zhang, C., Zhou, D., Wang, X., Bai, X., Liu, W.: Traffic sign detection and recognition using fully convolutional network guided proposals. Neurocomputing 214, 758\u2013766 (2016)","journal-title":"Neurocomputing"},{"issue":"7","key":"2_CR3","doi-asserted-by":"publisher","first-page":"2022","DOI":"10.1109\/TITS.2015.2482461","volume":"17","author":"Y Yang","year":"2015","unstructured":"Yang, Y., Luo, H., Xu, H., Wu, F.: Towards real-time traffic sign detection and classification. IEEE Trans. Intell. Transp. Syst. 17(7), 2022\u20132031 (2015)","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"issue":"4","key":"2_CR4","doi-asserted-by":"publisher","first-page":"127","DOI":"10.3390\/a10040127","volume":"10","author":"J Zhang","year":"2017","unstructured":"Zhang, J., Huang, M., Jin, X., Li, X.: A real-time Chinese traffic sign detection algorithm based on modified YOLOv2. Algorithms 10(4), 127 (2017)","journal-title":"Algorithms"},{"key":"2_CR5","doi-asserted-by":"publisher","first-page":"323","DOI":"10.1016\/j.neunet.2012.02.016","volume":"32","author":"J Stallkamp","year":"2012","unstructured":"Stallkamp, J., Schlipsing, M., Salmen, J., Igel, C.: Man vs. computer: benchmarking machine learning algorithms for traffic sign recognition. Neural Netw. 32, 323\u2013332 (2012). https:\/\/doi.org\/10.1016\/j.neunet.2012.02.016","journal-title":"Neural Netw."},{"key":"2_CR6","doi-asserted-by":"crossref","unstructured":"Stallkamp, J., Schlipsing, M., Salmen, J., Igel, C.: The German traffic sign recognition benchmark: a multi-class classification competition. In: International Joint Conference on Neural Networks (2011)","DOI":"10.1109\/IJCNN.2011.6033395"},{"key":"2_CR7","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"238","DOI":"10.1007\/978-3-642-21227-7_23","volume-title":"Image Analysis","author":"F Larsson","year":"2011","unstructured":"Larsson, F., Felsberg, M.: Using Fourier descriptors and spatial models for traffic sign recognition. In: Heyden, A., Kahl, F. (eds.) SCIA 2011. LNCS, vol. 6688, pp. 238\u2013249. Springer, Heidelberg (2011). https:\/\/doi.org\/10.1007\/978-3-642-21227-7_23"},{"key":"2_CR8","doi-asserted-by":"crossref","unstructured":"Wang, G., Ren, G., Quan, T.: A traffic sign detection method with high accuracy and efficiency. In: International Conference on Computer Science and Electronics Engineering (2013)","DOI":"10.2991\/iccsee.2013.359"},{"key":"2_CR9","doi-asserted-by":"crossref","unstructured":"Sermanet, P., LeCun, Y.: Traffic sign recognition with multi-scale convolutional networks. In: International Joint Conference on Neural Networks (2011)","DOI":"10.1109\/IJCNN.2011.6033589"},{"key":"2_CR10","unstructured":"Mao, X., Hijazi, S., Casas, R., Kaul, P., Kumar, R., Rowen, C.: Hierarchical CNN for traffic sign recognition. In: IEEE Intelligent Vehicles Symposium (IV) (2016)"},{"key":"2_CR11","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: unified, real-time object detection. In: IEEE CVPR, pp. 779\u2013788 (2016)","DOI":"10.1109\/CVPR.2016.91"},{"key":"2_CR12","doi-asserted-by":"crossref","unstructured":"Redmon, J., Farhadi, A.: YOLO9000: better, faster, stronger. In: IEEE CVPR, pp. 7263\u20137271 (2017)","DOI":"10.1109\/CVPR.2017.690"},{"key":"2_CR13","doi-asserted-by":"crossref","unstructured":"Girshick, R.: Fast R-CNN. In: IEEE ICCV, pp. 1440\u20131448 (2015)","DOI":"10.1109\/ICCV.2015.169"},{"key":"2_CR14","unstructured":"Maas, A.L., Hannun, A.Y., Ng, A.Y.: Rectifier nonlinearities improve neural network acoustic models. In: ICML (2013)"},{"key":"2_CR15","doi-asserted-by":"crossref","unstructured":"Rezatofighi, H., Tsoi, N., Gwak, J., Sadeghian, A., Reid, I., Savarese, S.: Generalized intersection over union: a metric and a loss for bounding box regression. In: IEEE Conference on Computer Vision and Pattern Recognition (2019)","DOI":"10.1109\/CVPR.2019.00075"},{"key":"2_CR16","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-61081-4","volume-title":"Computational Methods for Deep Learning - Theoretic","author":"WQ Yan","year":"2021","unstructured":"Yan, W.Q.: Computational Methods for Deep Learning - Theoretic. Practice and Applications. Springer, Heidelberg (2021). https:\/\/doi.org\/10.1007\/978-3-030-61081-4"},{"key":"2_CR17","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-60228-8","volume-title":"Introduction to Intelligent Surveillance - Surveillance Data Capture, Transmission, and Analytics","author":"WQ Yan","year":"2019","unstructured":"Yan, W.Q.: Introduction to Intelligent Surveillance - Surveillance Data Capture, Transmission, and Analytics, 3rd edn. Springer, Heidelberg (2019). https:\/\/doi.org\/10.1007\/978-3-319-60228-8","edition":"3"},{"issue":"27-28","key":"2_CR18","doi-asserted-by":"publisher","first-page":"19925","DOI":"10.1007\/s11042-020-08866-x","volume":"79","author":"C Pan","year":"2020","unstructured":"Pan, C., Yan, W.Q.: Object detection based on saturation of visual perception. Multimed. Tools Appl. 79(27\u201328), 19925\u201319944 (2020). https:\/\/doi.org\/10.1007\/s11042-020-08866-x","journal-title":"Multimed. Tools Appl."},{"key":"2_CR19","doi-asserted-by":"crossref","unstructured":"Pan, C., Li, X., Yan, W.: A learning-based positive feedback approach in salient object detection. In: IEEE IVCNZ (2018)","DOI":"10.1109\/IVCNZ.2018.8634717"},{"key":"2_CR20","doi-asserted-by":"crossref","unstructured":"Liu, X., Yan, W., Kasabov, N.: Vehicle-related scene segmentation using CapsNets. In: IEEE IVCNZ (2020)","DOI":"10.1109\/IVCNZ51579.2020.9290664"},{"key":"2_CR21","series-title":"Communications in Computer and Information Science","doi-asserted-by":"publisher","first-page":"61","DOI":"10.1007\/978-981-15-3651-9_7","volume-title":"Pattern Recognition","author":"X Liu","year":"2020","unstructured":"Liu, X., Neuyen, M., Yan, W.: Vehicle-related scene understanding using deep learning. In: Cree, M., Huang, F., Yuan, J., Yan, W.Q. (eds.) ACPR 2019. CCIS, vol. 1180, pp. 61\u201373. Springer, Singapore (2020). https:\/\/doi.org\/10.1007\/978-981-15-3651-9_7"},{"issue":"2","key":"2_CR22","doi-asserted-by":"publisher","first-page":"1679","DOI":"10.1007\/s11042-017-4356-z","volume":"77","author":"J Wang","year":"2017","unstructured":"Wang, J., Bacic, B., Yan, W.Q.: An effective method for plate number recognition. Multimed. Tools Appl. 77(2), 1679\u20131692 (2017). https:\/\/doi.org\/10.1007\/s11042-017-4356-z","journal-title":"Multimed. Tools Appl."},{"issue":"3","key":"2_CR23","doi-asserted-by":"publisher","first-page":"224","DOI":"10.1109\/TETCI.2017.2778716","volume":"2","author":"K Zheng","year":"2018","unstructured":"Zheng, K., Yan, W., Nand, P.: Video dynamics detection using deep neural networks. IEEE Trans. Emerg. Top. Comput. Intell. 2(3), 224\u2013234 (2018)","journal-title":"IEEE Trans. Emerg. Top. Comput. Intell."},{"key":"2_CR24","doi-asserted-by":"crossref","unstructured":"Shen, Y., Yan, W.: Blind spot monitoring using deep learning. In: IEEE IVCNZ (2018)","DOI":"10.1109\/IVCNZ.2018.8634716"},{"key":"2_CR25","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"439","DOI":"10.1007\/978-3-319-92753-4_34","volume-title":"Image and Video Technology","author":"G Qin","year":"2018","unstructured":"Qin, G., Yang, J., Yan, W., Li, Y., Klette, R.: Local fast R-CNN flow for object-centric event recognition in complex traffic scenes. In: Satoh, S. (ed.) PSIVT 2017. LNCS, vol. 10799, pp. 439\u2013452. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-319-92753-4_34"},{"issue":"3","key":"2_CR26","doi-asserted-by":"publisher","first-page":"34","DOI":"10.4018\/IJDCF.2016070103","volume":"8","author":"J Wang","year":"2016","unstructured":"Wang, J., Yan, W.: BP-neural network for plate number recognition. Int. J. Digit. Crime Forensics 8(3), 34\u201345 (2016)","journal-title":"Int. J. Digit. Crime Forensics"},{"key":"2_CR27","doi-asserted-by":"crossref","unstructured":"An, N., Yan, W.: Multitarget tracking using Siamese neural networks. ACM TOMM (2021)","DOI":"10.1145\/3441656"},{"key":"2_CR28","doi-asserted-by":"crossref","unstructured":"Liu, X., Yan, W.: Traffic-light sign recognition using Capsule network. MTAP (2021)","DOI":"10.1007\/s11042-020-10455-x"},{"key":"2_CR29","doi-asserted-by":"crossref","unstructured":"Xing, J., Yan, W.: Traffic sign recognition using guided image filtering. In: ISGV (2021)","DOI":"10.1007\/978-3-030-72073-5_7"}],"container-title":["Communications in Computer and Information Science","Geometry and Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-72073-5_2","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,12,21]],"date-time":"2022-12-21T20:11:55Z","timestamp":1671653515000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-72073-5_2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"ISBN":["9783030720728","9783030720735"],"references-count":29,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-72073-5_2","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"value":"1865-0929","type":"print"},{"value":"1865-0937","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021]]},"assertion":[{"value":"18 March 2021","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ISGV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Symposium on Geometry and Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Auckland","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"New Zealand","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 January 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 January 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"isgv2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/cerv.aut.ac.nz\/isgv2021","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"50","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"29","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"58% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.72","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Due to the COVID-19 pandemic the conference was partly held in a virtual format.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}