{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,6]],"date-time":"2025-05-06T05:10:27Z","timestamp":1746508227570,"version":"3.40.3"},"publisher-location":"Cham","reference-count":26,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031159336"},{"type":"electronic","value":"9783031159343"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-15934-3_7","type":"book-chapter","created":{"date-parts":[[2022,9,6]],"date-time":"2022-09-06T00:02:53Z","timestamp":1662422573000},"page":"74-86","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["Data Augmented Dual-Attention Interactive Image Classification Network"],"prefix":"10.1007","author":[{"given":"Qiangxi","family":"Zhu","sequence":"first","affiliation":[]},{"given":"Zhixin","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,9,15]]},"reference":[{"key":"7_CR1","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"7_CR2","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"7_CR3","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"834","DOI":"10.1007\/978-3-319-10590-1_54","volume-title":"Computer Vision \u2013 ECCV 2014","author":"N Zhang","year":"2014","unstructured":"Zhang, N., Donahue, J., Girshick, R., Darrell, T.: Part-based R-CNNs for fine-grained category detection. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8689, pp. 834\u2013849. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10590-1_54"},{"key":"7_CR4","doi-asserted-by":"crossref","unstructured":"Branson, S., Van Horn, G., Belongie, S., Perona, P.: Bird species categorization using pose normalized deep convolutional nets. arXiv preprint arXiv:1406.2952 (2014)","DOI":"10.5244\/C.28.87"},{"issue":"11","key":"7_CR5","doi-asserted-by":"publisher","first-page":"7903","DOI":"10.1007\/s10489-021-02280-y","volume":"51","author":"C Liu","year":"2021","unstructured":"Liu, C., Huang, L., Wei, Z., Zhang, W.: Subtler mixed attention network on fine-grained image classification. Appl. Intell. 51(11), 7903\u20137916 (2021). https:\/\/doi.org\/10.1007\/s10489-021-02280-y","journal-title":"Appl. Intell."},{"key":"7_CR6","doi-asserted-by":"crossref","unstructured":"Luo, W., et al.: Cross-X learning for fine-grained visual categorization. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 8242\u20138251 (2019)","DOI":"10.1109\/ICCV.2019.00833"},{"key":"7_CR7","doi-asserted-by":"publisher","first-page":"4683","DOI":"10.1109\/TIP.2020.2973812","volume":"29","author":"D Chang","year":"2020","unstructured":"Chang, D., et al.: The devil is in the channels: mutual-channel loss for fine-grained image classification. IEEE Trans. Image Process. 29, 4683\u20134695 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"7_CR8","doi-asserted-by":"crossref","unstructured":"Lin, T.-Y., Chowdhury, A.R., Maji, S.: Bilinear CNN models for fine-grained visual recognition. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1449\u20131457 (2015)","DOI":"10.1109\/ICCV.2015.170"},{"key":"7_CR9","doi-asserted-by":"publisher","unstructured":"Yu, C., Zhao, X., Zheng, Q., Zhang, P., You, X.: Hierarchical bilinear pooling for fine-grained visual recognition. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) Computer Vision \u2013 ECCV 2018. ECCV 2018. Lecture Notes in Computer Science, vol. 11220. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01270-0_35","DOI":"10.1007\/978-3-030-01270-0_35"},{"key":"7_CR10","doi-asserted-by":"crossref","unstructured":"Wang, H., Wang, Q., Gao, M., Li, P., Zuo, W.: Multi-scale location-aware kernel representation for object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1248\u20131257 (2018)","DOI":"10.1109\/CVPR.2018.00136"},{"key":"7_CR11","doi-asserted-by":"publisher","unstructured":"Guo, C., Lin, Y., Xu, M., et al.: Inverse transformation sampling-based attentive cutout for fine-grained visual recognition. Vis. Comput., 1\u201312 (2022). https:\/\/doi.org\/10.1007\/s00371-022-02481-7","DOI":"10.1007\/s00371-022-02481-7"},{"issue":"1","key":"7_CR12","first-page":"1","volume":"17","author":"Z Li","year":"2021","unstructured":"Li, Z., Lin, L., Zhang, C., Ma, H., Zhao, W., Shi, Z.: A semi-supervised learning approach based on adaptive weighted fusion for automatic image annotation. ACM Trans. Multimed. Comput. Commun. Appl. 17(1), 1\u201323 (2021)","journal-title":"ACM Trans. Multimed. Comput. Commun. Appl."},{"issue":"9","key":"7_CR13","doi-asserted-by":"publisher","first-page":"6871","DOI":"10.1007\/s11042-019-08568-z","volume":"79","author":"T Zhou","year":"2020","unstructured":"Zhou, T., Li, Z., Zhang, C., Ma, H.: Classify multi-label images via improved CNN model with adversarial network. Multimedia Tools Appl. 79(9), 6871\u20136890 (2020)","journal-title":"Multimedia Tools Appl."},{"key":"7_CR14","doi-asserted-by":"crossref","unstructured":"Zhang, F., Li, M., Zhai, G., Liu, Y.: Multi-branch and multi-scale attention learning for fine-grained visual categorization. arXiv preprint arXiv:2003.09150 (2020)","DOI":"10.1007\/978-3-030-67832-6_12"},{"key":"7_CR15","doi-asserted-by":"crossref","unstructured":"Fu, J., Zheng, H., Mei, T.: Look closer to see better: recurrent attention convolutional neural network for fine-grained image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4438\u20134446 (2017)","DOI":"10.1109\/CVPR.2017.476"},{"key":"7_CR16","unstructured":"Wah, C., Branson, S., Welinder, P., Perona, P., Belongie, S.: The caltech-UCSD birds-200-2011 dataset. Technical report 2010\u2013001, California Institute of Technology (2011)"},{"key":"7_CR17","unstructured":"Maji, S., Rahtu, E., Kannala, J., Blaschko, M., Vedaldi, A.: Fine-grained visual classification of aircraft. arXiv preprint arXiv:1306.5151 (2013)"},{"key":"7_CR18","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"337","DOI":"10.1007\/978-3-319-48674-1_30","volume-title":"Cloud Computing and Security","author":"M Liu","year":"2016","unstructured":"Liu, M., Yu, C., Ling, H., Lei, J.: Hierarchical joint CNN-based models for fine-grained cars recognition. In: Sun, X., Liu, A., Chao, H.-C., Bertino, E. (eds.) ICCCS 2016. LNCS, vol. 10040, pp. 337\u2013347. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-48674-1_30"},{"key":"7_CR19","doi-asserted-by":"crossref","unstructured":"Chen, Y., Bai, Y., Zhang, W., Mei, T.: Destruction and construction learning for fine-grained image recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5157\u20135166 (2019)","DOI":"10.1109\/CVPR.2019.00530"},{"issue":"10","key":"7_CR20","doi-asserted-by":"publisher","first-page":"2972","DOI":"10.1109\/TCSVT.2018.2875449","volume":"29","author":"K Song","year":"2018","unstructured":"Song, K., Yang, H., Yin, Z.: Multi-scale attention deep neural network for fast accurate object detection. IEEE Trans. Circ. Syst. Video Technol. 29(10), 2972\u20132985 (2018)","journal-title":"IEEE Trans. Circ. Syst. Video Technol."},{"key":"7_CR21","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"438","DOI":"10.1007\/978-3-030-01264-9_26","volume-title":"Computer Vision \u2013 ECCV 2018","author":"Z Yang","year":"2018","unstructured":"Yang, Z., Luo, T., Wang, D., Hu, Z., Gao, J., Wang, L.: Learning to navigate for fine-grained classification. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) Computer Vision \u2013 ECCV 2018. LNCS, vol. 11218, pp. 438\u2013454. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01264-9_26"},{"key":"7_CR22","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"834","DOI":"10.1007\/978-3-030-01270-0_49","volume-title":"Computer Vision \u2013 ECCV 2018","author":"M Sun","year":"2018","unstructured":"Sun, M., Yuan, Y., Zhou, F., Ding, E.: Multi-attention multi-class constraint for fine-grained image recognition. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11220, pp. 834\u2013850. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01270-0_49"},{"key":"7_CR23","doi-asserted-by":"crossref","unstructured":"Ding, Y., Zhou, Y., Zhu, Y., Ye, Q., Jiao, J.: Selective sparse sampling for fine-grained image recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 6599\u20136608 (2019)","DOI":"10.1109\/ICCV.2019.00670"},{"key":"7_CR24","doi-asserted-by":"crossref","unstructured":"Zheng, Y., Chang, D., Xie, J., Ma, Z.: IU-module: intersection and union module for fine-grained visual classification. In Proceedings of the IEEE International Conference on Multimedia and Expo, pp. 1\u20136. IEEE (2020)","DOI":"10.1109\/ICME46284.2020.9102839"},{"key":"7_CR25","doi-asserted-by":"crossref","unstructured":"Li, X., Yang, C., Chen, S.-L., Zhu, C., Yin, X.-C.: Semantic bilinear pooling for fine-grained recognition. In: Proceedings of the 2020 25th International Conference on Pattern Recognition, pp. 3660\u20133666. IEEE (2021)","DOI":"10.1109\/ICPR48806.2021.9412252"},{"key":"7_CR26","doi-asserted-by":"crossref","unstructured":"Zhou, B., Khosla, A., Lapedriza, A., Oliva, A., Torralba, A.: Learning deep features for discriminative localization. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2921\u20132929 (2016)","DOI":"10.1109\/CVPR.2016.319"}],"container-title":["Lecture Notes in Computer Science","Artificial Neural Networks and Machine Learning \u2013 ICANN 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-15934-3_7","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,9,14]],"date-time":"2022-09-14T06:06:07Z","timestamp":1663135567000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-15934-3_7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031159336","9783031159343"],"references-count":26,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-15934-3_7","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"15 September 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICANN","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Artificial Neural Networks","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Bristol","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"United Kingdom","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6 September 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"9 September 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"31","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icann2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/e-nns.org\/icann2022\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"561","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"255","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"45% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}