{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T14:45:50Z","timestamp":1742913950086,"version":"3.40.3"},"publisher-location":"Cham","reference-count":31,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783031301100"},{"type":"electronic","value":"9783031301117"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-30111-7_8","type":"book-chapter","created":{"date-parts":[[2023,4,12]],"date-time":"2023-04-12T05:02:51Z","timestamp":1681275771000},"page":"85-96","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["CCN: Pavement Crack Detection with\u00a0Context Contrasted Net"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-5328-7151","authenticated-orcid":false,"given":"Yihuan","family":"Zhu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4600-2928","authenticated-orcid":false,"given":"Sheng","family":"Zhang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2487-7152","authenticated-orcid":false,"given":"Chengfeng","family":"Ruan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,4,13]]},"reference":[{"issue":"7533","key":"8_CR1","doi-asserted-by":"publisher","first-page":"436","DOI":"10.1038\/nature14539","volume":"521","author":"Y LeCun","year":"2015","unstructured":"LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521(7533), 436\u2013444 (2015)","journal-title":"Nature"},{"key":"8_CR2","unstructured":"Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: towards real-time object detection with region proposal networks. In: Neural Information Processing Systems, vol. 28, pp. 91\u201399 (2015)"},{"key":"8_CR3","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: Unified, real-time object detection. In: CVPR, pp. 779\u2013788 (2016)","DOI":"10.1109\/CVPR.2016.91"},{"key":"8_CR4","doi-asserted-by":"crossref","unstructured":"Lin, T. Y., Goyal, P., Girshick, R., He, K., Doll\u00e1r, P. : Focal loss for dense object detection. In: ICCV, pp. 2980\u20132988 (2017)","DOI":"10.1109\/ICCV.2017.324"},{"key":"8_CR5","doi-asserted-by":"crossref","unstructured":"Tian, Z., Shen, C., Chen, H., He, T. : FCOS: fully convolutional one-stage object detection. In: ICCV, pp. 9626\u20139635 (2019)","DOI":"10.1109\/ICCV.2019.00972"},{"key":"8_CR6","doi-asserted-by":"crossref","unstructured":"Tan, M., Pang, R., Le, Q.V.: EfficientDet: scalable and efficient object detection. In: CVPR, pp. 10781\u201310790 (2020)","DOI":"10.1109\/CVPR42600.2020.01079"},{"key":"8_CR7","doi-asserted-by":"crossref","unstructured":"Nguyen, N.T.H., Le, T.H., Perry, S., Nguyen, T.T.: Pavement crack detection using convolutional neural network. In: SoICT, pp. 251\u2013256 (2018)","DOI":"10.1145\/3287921.3287949"},{"key":"8_CR8","doi-asserted-by":"crossref","unstructured":"Gou, C., Peng, B., Li, T., Gao, Z.: Pavement crack detection based on the improved faster-RCNN. In: ISKE, pp. 962\u2013967 (2019)","DOI":"10.1109\/ISKE47853.2019.9170456"},{"key":"8_CR9","doi-asserted-by":"crossref","unstructured":"Yusof, N.A.M., et al.: Automated asphalt pavement crack detection and classification using deep convolution neural network. In: ICCSCE, pp. 215\u2013220 (2019)","DOI":"10.1109\/ICCSCE47578.2019.9068551"},{"issue":"4","key":"8_CR10","doi-asserted-by":"publisher","first-page":"1525","DOI":"10.1109\/TITS.2019.2910595","volume":"21","author":"F Yang","year":"2019","unstructured":"Yang, F., Zhang, L., Yu, S., Prokhorov, D., Mei, X., Ling, H.: Feature pyramid and hierarchical boosting network for pavement crack detection. IEEE Trans. Intell. Transp. Syst. 21(4), 1525\u20131535 (2019)","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"issue":"8","key":"8_CR11","doi-asserted-by":"publisher","first-page":"1580","DOI":"10.1049\/iet-ipr.2019.0973","volume":"14","author":"X Xiang","year":"2020","unstructured":"Xiang, X., Zhang, Y., El Saddik, A.: Pavement crack detection network based on pyramid structure and attention mechanism. IET Image Process. 14(8), 1580\u20131586 (2020)","journal-title":"IET Image Process."},{"key":"8_CR12","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"664","DOI":"10.1007\/978-3-030-60639-8_55","volume-title":"Pattern Recognition and Computer Vision","author":"J Wang","year":"2020","unstructured":"Wang, J., Liu, F., Yang, W., Xu, G., Tao, Z.: Pavement crack detection using attention U-Net with multiple sources. In: Peng, Y., et al. (eds.) PRCV 2020. LNCS, vol. 12306, pp. 664\u2013672. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-60639-8_55"},{"issue":"2","key":"8_CR13","doi-asserted-by":"publisher","first-page":"1306","DOI":"10.1109\/TITS.2020.2990703","volume":"22","author":"K Zhang","year":"2020","unstructured":"Zhang, K., Zhang, Y., Cheng, H.D.: Crack-GAN: pavement crack detection using partially accurate ground truths based on generative adversarial learning. IEEE Trans. Intell. Transp. Syst. 22(2), 1306\u20131319 (2020)","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"8_CR14","doi-asserted-by":"crossref","unstructured":"Cheng, W., Zhou, Y.: Automatic pavement crack detection based on hierarchical feature augmentation. In: ICAIIS, pp. 1\u20137 (2021)","DOI":"10.1145\/3469213.3470392"},{"key":"8_CR15","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/978-3-030-01234-2_1","volume-title":"Computer Vision \u2013 ECCV 2018","author":"S Woo","year":"2018","unstructured":"Woo, S., Park, J., Lee, J.-Y., Kweon, I.S.: CBAM: convolutional block attention module. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11211, pp. 3\u201319. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01234-2_1"},{"key":"8_CR16","doi-asserted-by":"crossref","unstructured":"He, K., Gkioxari, G., Doll\u00e1r, P., Girshick, R.: Mask R-CNN. In: ICCV, pp. 2961\u20132969 (2017)","DOI":"10.1109\/ICCV.2017.322"},{"key":"8_CR17","doi-asserted-by":"crossref","unstructured":"Pang, J., Chen, K., Shi, J., Feng, H., Ouyang, W., Lin, D.: Libra R-CNN: towards balanced learning for object detection. In: CVPR, pp. 821\u2013830 (2019)","DOI":"10.1109\/CVPR.2019.00091"},{"key":"8_CR18","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Doll\u00e1r, P., Girshick, R.B., He, K., Hariharan, B., Belongie, S.J.: Feature pyramid networks for object detection. In: CVPR, pp. 936\u2013944 (2017)","DOI":"10.1109\/CVPR.2017.106"},{"key":"8_CR19","unstructured":"Park, J., Woo, S., Lee, J.Y., Kweon, I.S.: Bam: bottleneck attention module. arXiv preprint arXiv:1807.06514 (2018)"},{"key":"8_CR20","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"8_CR21","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"8_CR22","unstructured":"Tan, M., Le, Q.: EfficientNet: rethinking model scaling for convolutional neural networks. In: ICML, pp. 6105\u20136114 (2019)"},{"key":"8_CR23","doi-asserted-by":"crossref","unstructured":"Li, Z., Liu, Y., Li, B., Hu, W., Miao, Y., Zhang, H.: DSIC: dynamic sample-individualized connector for multi-scale object detection. In: ICME (2021)","DOI":"10.1109\/ICME51207.2021.9428167"},{"key":"8_CR24","doi-asserted-by":"crossref","unstructured":"Liu, S., Qi, L., Qin, H., Shi, J., Jia, J.: Path aggregation network for instance segmentation. In: CVPR, pp. 8759\u20138768 (2018)","DOI":"10.1109\/CVPR.2018.00913"},{"key":"8_CR25","doi-asserted-by":"crossref","unstructured":"Ghiasi, G., Lin, T.Y., Le, Q.V.: NAS-FPN: learning scalable feature pyramid architecture for object detection. In: CVPR, pp. 7036\u20137045 (2019)","DOI":"10.1109\/CVPR.2019.00720"},{"key":"8_CR26","doi-asserted-by":"crossref","unstructured":"Arya, D., Maeda, H., Ghosh, S.K., et al.: Global road damage detection: state-of-the-art solutions. In: IEEE BigData, pp. 5533\u20135539 (2020)","DOI":"10.1109\/BigData50022.2020.9377790"},{"key":"8_CR27","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: ImageNet: a large-scale hierarchical image database. In: CVPR, pp. 248\u2013255 (2009)","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"8_CR28","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"421","DOI":"10.1007\/978-3-642-35289-8_25","volume-title":"Neural Networks: Tricks of the Trade","author":"L Bottou","year":"2012","unstructured":"Bottou, L.: Stochastic gradient descent tricks. In: Montavon, G., Orr, G.B., M\u00fcller, K.-R. (eds.) Neural Networks: Tricks of the Trade. LNCS, vol. 7700, pp. 421\u2013436. Springer, Heidelberg (2012). https:\/\/doi.org\/10.1007\/978-3-642-35289-8_25"},{"key":"8_CR29","unstructured":"Loshchilov, I., Hutter, F.: SGDR: stochastic gradient descent with warm restarts. arXiv preprint arXiv:1608.03983 (2016)"},{"key":"8_CR30","doi-asserted-by":"crossref","unstructured":"Peng, C., et al.: MegDet: a large mini-batch object detector. In: CVPR, pp. 6181\u20136189 (2018)","DOI":"10.1109\/CVPR.2018.00647"},{"key":"8_CR31","unstructured":"Chen, K., Wang, J., Pang, J., et al.: MMDetection: open MMLab detection toolbox and benchmark. CoRR (2019)"}],"container-title":["Lecture Notes in Computer Science","Neural Information Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-30111-7_8","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,4,12]],"date-time":"2023-04-12T05:04:23Z","timestamp":1681275863000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-30111-7_8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031301100","9783031301117"],"references-count":31,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-30111-7_8","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"13 April 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICONIP","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Neural Information Processing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"New Delhi","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"India","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 November 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26 November 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iconip2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/iconip2022.apnns.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Easy Chair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"810","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"359","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"44% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.65","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"ICONIP 2022 consists of a two-volume set, LNCS & CCIS, which includes 146 and 213 papers","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}