{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T22:49:00Z","timestamp":1743029340996,"version":"3.40.3"},"publisher-location":"Cham","reference-count":34,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783031189067"},{"type":"electronic","value":"9783031189074"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-18907-4_2","type":"book-chapter","created":{"date-parts":[[2022,10,26]],"date-time":"2022-10-26T23:03:53Z","timestamp":1666825433000},"page":"17-28","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Dual-Rank Attention Module for\u00a0Fine-Grained Vehicle Model Recognition"],"prefix":"10.1007","author":[{"given":"Wen","family":"Cai","sequence":"first","affiliation":[]},{"given":"Wenjia","family":"Zhu","sequence":"additional","affiliation":[]},{"given":"Bo","family":"Cheng","sequence":"additional","affiliation":[]},{"given":"Longdao","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Ye","family":"Yu","sequence":"additional","affiliation":[]},{"given":"Qiang","family":"Lu","sequence":"additional","affiliation":[]},{"given":"Wei","family":"Jia","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,27]]},"reference":[{"doi-asserted-by":"crossref","unstructured":"Yang, Z., et al.: Learning to navigate for fine-grained classification. In: ECCV, pp. 420\u2013435 (2018)","key":"2_CR1","DOI":"10.1007\/978-3-030-01264-9_26"},{"issue":"2","key":"2_CR2","doi-asserted-by":"publisher","first-page":"749","DOI":"10.1109\/TITS.2018.2833620","volume":"20","author":"FC Soon","year":"2019","unstructured":"Soon, F.C., et al.: PCANet-based convolutional neural network architecture for a vehicle model recognition system. IEEE Trans. Intell. Transp. Syst. 20(2), 749\u2013759 (2019)","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"doi-asserted-by":"publisher","unstructured":"Yang, L., et al.: A large-scale car dataset for fine-grained categorization and verification. In: CVPR (2015). https:\/\/doi.org\/10.1109\/CVPR.2015.7299023","key":"2_CR3","DOI":"10.1109\/CVPR.2015.7299023"},{"doi-asserted-by":"publisher","unstructured":"Fu, J., et al.: Look closer to see better: recurrent attention convolutional neural network for fine-grained image recognition. In: CVPR (2017). https:\/\/doi.org\/10.1109\/CVPR.2017.476","key":"2_CR4","DOI":"10.1109\/CVPR.2017.476"},{"unstructured":"Jonathan, K., et al.: 3D object representations for fine-grained categorization. In: 3DRR (2013)","key":"2_CR5"},{"doi-asserted-by":"publisher","unstructured":"Em, Y., et al.: Incorporating intra-class variance to fine-grained visual recognition. In: IEEE International Conference on Multimedia and Expo (ICME), vol. 1, pp. 1452\u20131457 (2017). https:\/\/doi.org\/10.1109\/ICME.2017.8019371","key":"2_CR6","DOI":"10.1109\/ICME.2017.8019371"},{"doi-asserted-by":"crossref","unstructured":"Hu, J., et al.: Squeeze-and-excitation networks. In: CVPR (2018)","key":"2_CR7","DOI":"10.1109\/CVPR.2018.00745"},{"unstructured":"Lin, Z., et al.: Space: unsupervised object-oriented scene representation via spatial attention and decomposition. arXiv preprint arXiv:2001.02407 (2020)","key":"2_CR8"},{"issue":"1","key":"2_CR9","first-page":"113","volume":"28","author":"Z Yousong","year":"2018","unstructured":"Yousong, Z., et al.: Attention CoupleNet: fully convolutional attention coupling network for object detection. IEEE Trans. Image Process. 28(1), 113\u2013126 (2018)","journal-title":"IEEE Trans. Image Process."},{"unstructured":"Heliang, Z., et al.: Learning multi-attention convolutional neural network for fine-grained image recognition. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp. 5209\u20135217 (2017)","key":"2_CR10"},{"issue":"6","key":"2_CR11","doi-asserted-by":"publisher","first-page":"1245","DOI":"10.1109\/TMM.2017.2648498","volume":"19","author":"Z Bo","year":"2017","unstructured":"Bo, Z., et al.: Diversified visual attention networks for fine-grained object classification. IEEE Trans. Multimedia 19(6), 1245\u20131256 (2017)","journal-title":"IEEE Trans. Multimedia"},{"issue":"8","key":"2_CR12","doi-asserted-by":"publisher","first-page":"1213","DOI":"10.1109\/LSP.2017.2715076","volume":"24","author":"Z Yuanyi","year":"2017","unstructured":"Yuanyi, Z., et al.: Toward end-to-end face recognition through alignment learning. IEEE Signal Process. Lett. 24(8), 1213\u20131217 (2017)","journal-title":"IEEE Signal Process. Lett."},{"issue":"11","key":"2_CR13","doi-asserted-by":"publisher","first-page":"2905","DOI":"10.1109\/TMM.2020.2965491","volume":"22","author":"L Hao","year":"2020","unstructured":"Hao, L., et al.: STNReID: deep convolutional networks with pairwise spatial transformer networks for partial person re-identification. IEEE Trans. Multimedia 22(11), 2905\u20132913 (2020)","journal-title":"IEEE Trans. Multimedia"},{"unstructured":"Max, J., et al.: Spatial transformer networks. In: Advances in Neural Information Processing Systems, vol. 28, pp. 2017\u20132025 (2015)","key":"2_CR14"},{"unstructured":"Jiaqi, W., et al.: CARAFE: Content-Aware ReAssembly of FEatures. In: ICCV, pp. 3007\u20133016 (2019)","key":"2_CR15"},{"unstructured":"Jun, F., et al.: Dual attention network for scene segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3146\u20133154 (2020)","key":"2_CR16"},{"unstructured":"Lin, M., et al.: Network in network. arXiv preprint arXiv:1312.4400 (2013)","key":"2_CR17"},{"doi-asserted-by":"crossref","unstructured":"Ghassemi, S., et al.: Fine-grained vehicle classification using deep residual networks with multiscale attention windows. In: IEEE International Workshop on Multimedia Signal Processing, pp. 1\u20136 (2017)","key":"2_CR18","DOI":"10.1109\/MMSP.2017.8122262"},{"unstructured":"Liu, X., et al.: Fully convolutional attention localization networks: efficient attention localization for fine-grained recognition. arXiv preprint arXiv:1603.06765, vol. 1, no. 2 (2016)","key":"2_CR19"},{"unstructured":"Ruyi, J., et al.: Attention convolutional binary neural tree for fine-grained visual categorization. In: CVPR, pp. 10468\u201310477 (2020)","key":"2_CR20"},{"issue":"5","key":"2_CR21","first-page":"1791","volume":"49","author":"W Lin","year":"2018","unstructured":"Lin, W., et al.: Deep attention-based spatially recursive networks for fine-grained visual recognition. IEEE Trans. Cybern. 49(5), 1791\u20131802 (2018)","journal-title":"IEEE Trans. Cybern."},{"key":"2_CR22","doi-asserted-by":"publisher","first-page":"104027","DOI":"10.1016\/j.imavis.2020.104027","volume":"104","author":"Y Ye","year":"2020","unstructured":"Ye, Y., et al.: CAM: a fine-grained vehicle model recognition method based on visual attention model. Image Vis. Comput. 104, 104027 (2020)","journal-title":"Image Vis. Comput."},{"unstructured":"Dubey, A., et al.: Maximum-entropy fine grained classification. In: Advances in Neural Information Processing Systems, vol. 31, pp. 635\u2013645 (2018)","key":"2_CR23"},{"key":"2_CR24","doi-asserted-by":"publisher","first-page":"117944","DOI":"10.1109\/ACCESS.2019.2936118","volume":"7","author":"T Min","year":"2019","unstructured":"Min, T., et al.: Fine-grained classification via hierarchical bilinear pooling with aggregated slack mask. IEEE Access 7, 117944\u2013117953 (2019)","journal-title":"IEEE Access"},{"doi-asserted-by":"crossref","unstructured":"Shijin, L.I., et al.: Research on fine-grain model recognition based on branch feedback convolution neural network. In: ICCSE, pp. 47\u201351 (2019)","key":"2_CR25","DOI":"10.1109\/ICCSE.2019.8845061"},{"doi-asserted-by":"crossref","unstructured":"Cao, J., et al.: End-to-end view-aware vehicle classification via progressive CNN learning. In: CCF Chinese Conference on Computer Vision, pp. 729\u2013737 (2017)","key":"2_CR26","DOI":"10.1007\/978-981-10-7299-4_61"},{"key":"2_CR27","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"664","DOI":"10.1007\/978-3-030-11018-5_54","volume-title":"Computer Vision \u2013 ECCV 2018 Workshops","author":"S Elkerdawy","year":"2019","unstructured":"Elkerdawy, S., Ray, N., Zhang, H.: Fine-grained vehicle classification with unsupervised parts co-occurrence learning. In: Leal-Taix\u00e9, L., Roth, S. (eds.) ECCV 2018. LNCS, vol. 11132, pp. 664\u2013670. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-11018-5_54"},{"doi-asserted-by":"crossref","unstructured":"Yanling, T., et al.: Selective multi-convolutional region feature extraction based iterative discrimination CNN for fine-grained vehicle model recognition. In: International Conference on Pattern Recognition (ICPR), pp. 3279\u20133284 (2018)","key":"2_CR28","DOI":"10.1109\/ICPR.2018.8545375"},{"unstructured":"Ye, Y., et al.: Embedding pose information for multiview vehicle model recognition. IEEE Trans. Circuits Syst. Video Technol. (TCSVT) (2022)","key":"2_CR29"},{"issue":"3","key":"2_CR30","first-page":"593","volume":"25","author":"Y Changdong","year":"2020","unstructured":"Changdong, Y., et al.: A method of enhancing data based on AT-PGGAN for fine-grained recognition of vehicle models. J. Image Graph. 25(3), 593\u2013604 (2020)","journal-title":"J. Image Graph."},{"issue":"5","key":"2_CR31","doi-asserted-by":"publisher","first-page":"3123","DOI":"10.1109\/TITS.2020.2981737","volume":"22","author":"Y Ye","year":"2021","unstructured":"Ye, Y., et al.: A multilayer pyramid network based on learning for vehicle logo recognition. IEEE Trans. Intell. Transp. Syst. 22(5), 3123\u20133134 (2021)","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"2_CR32","doi-asserted-by":"publisher","first-page":"273","DOI":"10.1016\/j.compeleceng.2018.07.045","volume":"71","author":"Y Ye","year":"2018","unstructured":"Ye, Y., et al.: Vehicle logo recognition based on overlapping enhanced patterns of oriented edge magnitudes. Comput. Electr. Eng. 71, 273\u2013283 (2018)","journal-title":"Comput. Electr. Eng."},{"doi-asserted-by":"crossref","unstructured":"Tanveer, M., et al.: Fine-tuning DARTS for image classification. In: 2020 25th International Conference on Pattern Recognition (ICPR), pp. 4789\u20134796 (2021)","key":"2_CR33","DOI":"10.1109\/ICPR48806.2021.9412221"},{"unstructured":"Zhang, Y., et al.: Bamboo: Building Mega-Scale Vision Dataset Continually with Human-Machine Synergy (2022)","key":"2_CR34"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-18907-4_2","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,6]],"date-time":"2024-10-06T16:33:00Z","timestamp":1728232380000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-18907-4_2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031189067","9783031189074"],"references-count":34,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-18907-4_2","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"27 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Shenzhen","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"14 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/en.prcv.cn\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"microsoft","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"564","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"233","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"41% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.03","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.35","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}