{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T16:00:48Z","timestamp":1742918448126,"version":"3.40.3"},"publisher-location":"Cham","reference-count":33,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030930486"},{"type":"electronic","value":"9783030930493"}],"license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021]]},"DOI":"10.1007\/978-3-030-93049-3_7","type":"book-chapter","created":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T06:00:24Z","timestamp":1641016824000},"page":"79-90","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Attribute and Identity Are Equally Important: Person Re-identification with More Powerful Pedestrian Attributes"],"prefix":"10.1007","author":[{"given":"Shuangye","family":"Chen","sequence":"first","affiliation":[]},{"given":"Kai","family":"Xu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,1,1]]},"reference":[{"issue":"4","key":"7_CR1","doi-asserted-by":"publisher","first-page":"1156","DOI":"10.1109\/TPAMI.2019.2947427","volume":"43","author":"A Siarohin","year":"2019","unstructured":"Siarohin, A., Lathuili\u00e8re, S., Sangineto, E., et al.: Appearance and pose-conditioned human image generation using deformable GANs. IEEE Trans. Pattern Anal. Mach. Intell. 43(4), 1156\u20131171 (2019)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"doi-asserted-by":"crossref","unstructured":"Wojke, N., Bewley, A., Paulus, D.: Simple online and realtime tracking with a deep association metric. In: 2017 IEEE International Conference on Image Processing (ICIP), pp. 3645\u20133649. IEEE (2017)","key":"7_CR2","DOI":"10.1109\/ICIP.2017.8296962"},{"doi-asserted-by":"crossref","unstructured":"Ristani, E., Tomasi, C.: Features for multi-target multi-camera tracking and re-identification. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6036\u20136046 (2018)","key":"7_CR3","DOI":"10.1109\/CVPR.2018.00632"},{"doi-asserted-by":"crossref","unstructured":"Li, W.H., Hong, F.T., Zheng, W.S.: Learning to learn relation for important people detection in still images. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5003\u20135011 (2019)","key":"7_CR4","DOI":"10.1109\/CVPR.2019.00514"},{"issue":"1","key":"7_CR5","first-page":"1","volume":"14","author":"Z Zheng","year":"2017","unstructured":"Zheng, Z., Zheng, L., Yang, Y.: A discriminatively learned CNN embedding for person reidentification. ACM Trans. Multimedia Comput. Commun. Appl. (TOMM) 14(1), 1\u201320 (2017)","journal-title":"ACM Trans. Multimedia Comput. Commun. Appl. (TOMM)"},{"doi-asserted-by":"crossref","unstructured":"Li, D., Chen, X., Huang, K.: Multi-attribute learning for pedestrian attribute recognition in surveillance scenarios. In: 2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR), pp. 111\u2013115. IEEE (2015)","key":"7_CR6","DOI":"10.1109\/ACPR.2015.7486476"},{"unstructured":"Liu, P., Liu, X., Yan, J., et al.: Localization guided learning for pedestrian attribute recognition. arXiv preprint arXiv:1808.09102 (2018)","key":"7_CR7"},{"doi-asserted-by":"crossref","unstructured":"Li, D., Chen, X., Zhang, Z., et al.: Pose guided deep model for pedestrian attribute recognition in surveillance scenarios. In: 2018 IEEE International Conference on Multimedia and Expo (ICME), pp. 1\u20136. IEEE (2018)","key":"7_CR8","DOI":"10.1109\/ICME.2018.8486604"},{"doi-asserted-by":"crossref","unstructured":"Liu, X., Zhao, H., Tian, M., et al.: HydraPlus-Net: attentive deep features for pedestrian analysis. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 350\u2013359 (2017)","key":"7_CR9","DOI":"10.1109\/ICCV.2017.46"},{"unstructured":"Sarfraz, M.S., Schumann, A., Wang, Y., et al.: Deep view-sensitive pedestrian attribute inference in an end-to-end model. arXiv preprint arXiv:1707.06089 (2017)","key":"7_CR10"},{"unstructured":"Wang, X., Zheng, S., Yang, R., et al.: Pedestrian attribute recognition: a survey. arXiv preprint arXiv:1901.07474 (2019)","key":"7_CR11"},{"issue":"10","key":"7_CR12","doi-asserted-by":"publisher","first-page":"3398","DOI":"10.1109\/TCSVT.2020.2982962","volume":"30","author":"J Wu","year":"2020","unstructured":"Wu, J., Liu, H., Jiang, J., et al.: Person attribute recognition by sequence contextual relation learning. IEEE Trans. Circuits Syst. Video Technol. 30(10), 3398\u20133412 (2020)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"doi-asserted-by":"crossref","unstructured":"Wu, S., Chen, Y.C., Li, X., et al.: An enhanced deep feature representation for person re-identification. In: 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 1\u20138. IEEE (2016)","key":"7_CR13","DOI":"10.1109\/WACV.2016.7477681"},{"doi-asserted-by":"crossref","unstructured":"Wei, L., Zhang, S., Yao, H., et al.: GLAD: global-local-alignment descriptor for pedestrian retrieval. In: Proceedings of the 25th ACM International Conference on Multimedia, pp. 420\u2013428 (2017)","key":"7_CR14","DOI":"10.1145\/3123266.3123279"},{"key":"7_CR15","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"34","DOI":"10.1007\/978-3-319-46466-4_3","volume-title":"Computer Vision \u2013 ECCV 2016","author":"E Insafutdinov","year":"2016","unstructured":"Insafutdinov, E., Pishchulin, L., Andres, B., Andriluka, M., Schiele, B.: DeeperCut: a deeper, stronger, and faster multi-person pose estimation model. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9910, pp. 34\u201350. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46466-4_3"},{"issue":"9","key":"7_CR16","doi-asserted-by":"publisher","first-page":"4500","DOI":"10.1109\/TIP.2019.2910414","volume":"28","author":"L Zheng","year":"2019","unstructured":"Zheng, L., Huang, Y., Lu, H., et al.: Pose-invariant embedding for deep person re-identification. IEEE Trans. Image Process. 28(9), 4500\u20134509 (2019)","journal-title":"IEEE Trans. Image Process."},{"key":"7_CR17","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"384","DOI":"10.1007\/978-3-030-01225-0_23","volume-title":"Computer Vision \u2013 ECCV 2018","author":"C Wang","year":"2018","unstructured":"Wang, C., Zhang, Q., Huang, C., Liu, W., Wang, X.: Mancs: a multi-task attentional network with curriculum sampling for person re-identification. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11208, pp. 384\u2013400. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01225-0_23"},{"key":"7_CR18","doi-asserted-by":"publisher","first-page":"151","DOI":"10.1016\/j.patcog.2019.06.006","volume":"95","author":"Y Lin","year":"2019","unstructured":"Lin, Y., Zheng, L., Zheng, Z., et al.: Improving person re-identification by attribute and identity learning. Pattern Recogn. 95, 151\u2013161 (2019)","journal-title":"Pattern Recogn."},{"doi-asserted-by":"crossref","unstructured":"Zhang, X., Pala, F., Bhanu, B.: Attributes co-occurrence pattern mining for video-based person re-identification. In: 2017 14th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), pp. 1\u20136. IEEE (2017)","key":"7_CR19","DOI":"10.1109\/AVSS.2017.8078469"},{"doi-asserted-by":"crossref","unstructured":"Tay, C.P., Roy, S., Yap, K.H.: AANet: attribute attention network for person re-identifications. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7134\u20137143 (2019)","key":"7_CR20","DOI":"10.1109\/CVPR.2019.00730"},{"issue":"11","key":"7_CR21","doi-asserted-by":"publisher","first-page":"1254","DOI":"10.1109\/34.730558","volume":"20","author":"L Itti","year":"1998","unstructured":"Itti, L., Koch, C., Niebur, E.: A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach. Intell. 20(11), 1254\u20131259 (1998)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"doi-asserted-by":"crossref","unstructured":"Chennupati, S., Sistu, G., Yogamani, S., et al.: MultiNet++: multi-stream feature aggregation and geometric loss strategy for multi-task learning. arXiv preprint arXiv:1904.08492 (2019)","key":"7_CR22","DOI":"10.1109\/CVPRW.2019.00159"},{"doi-asserted-by":"crossref","unstructured":"Chattopadhay, A., Sarkar, A., Howlader, P., et al.: Grad-CAM++: generalized gradient-based visual explanations for deep convolutional networks. In: 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 839\u2013847. IEEE (2018)","key":"7_CR23","DOI":"10.1109\/WACV.2018.00097"},{"unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)","key":"7_CR24"},{"key":"7_CR25","doi-asserted-by":"publisher","first-page":"126116","DOI":"10.1109\/ACCESS.2019.2939071","volume":"7","author":"S Zhang","year":"2019","unstructured":"Zhang, S., He, Y., Wei, J., et al.: Person re-identification with joint verification and identification of identity-attribute labels. IEEE Access 7, 126116\u2013126126 (2019)","journal-title":"IEEE Access"},{"issue":"11","key":"7_CR26","doi-asserted-by":"publisher","first-page":"3607","DOI":"10.1007\/s10489-020-01752-x","volume":"50","author":"J Yin","year":"2020","unstructured":"Yin, J., Fan, Z., Chen, S., Wang, Y.: In-depth exploration of attribute information for person re-identification. Appl. Intell. 50(11), 3607\u20133622 (2020). https:\/\/doi.org\/10.1007\/s10489-020-01752-x","journal-title":"Appl. Intell."},{"issue":"10","key":"7_CR27","doi-asserted-by":"publisher","first-page":"3016","DOI":"10.1109\/TCSVT.2018.2872503","volume":"29","author":"C Shen","year":"2018","unstructured":"Shen, C., Qi, G.J., Jiang, R., et al.: Sharp attention network via adaptive sampling for person re-identification. IEEE Trans. Circuits Syst. Video Technol. 29(10), 3016\u20133027 (2018)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"7_CR28","doi-asserted-by":"publisher","first-page":"57594","DOI":"10.1109\/ACCESS.2020.2971612","volume":"8","author":"Y Ha","year":"2020","unstructured":"Ha, Y., Tian, J., Miao, Q., et al.: Part-based enhanced super resolution network for low-resolution person re-identification. IEEE Access 8, 57594\u201357605 (2020)","journal-title":"IEEE Access"},{"doi-asserted-by":"crossref","unstructured":"Liu, Z., Wang, J., Gong, S., et al.: Deep reinforcement active learning for human-in-the-loop person re-identification. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 6122\u20136131 (2019)","key":"7_CR29","DOI":"10.1109\/ICCV.2019.00622"},{"key":"7_CR30","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"418","DOI":"10.1007\/978-3-030-01264-9_25","volume-title":"Computer Vision \u2013 ECCV 2018","author":"Y Suh","year":"2018","unstructured":"Suh, Y., Wang, J., Tang, S., Mei, T., Lee, K.M.: Part-aligned bilinear representations for person re-identification. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) Computer Vision \u2013 ECCV 2018. LNCS, vol. 11218, pp. 418\u2013437. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01264-9_25"},{"doi-asserted-by":"crossref","unstructured":"Yu, T., Li, D., Yang, Y., et al.: Robust person re-identification by modelling feature uncertainty. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 552\u2013561 (2019)","key":"7_CR31","DOI":"10.1109\/ICCV.2019.00064"},{"doi-asserted-by":"crossref","unstructured":"Li, Y.J., Chen, Y.C., Lin, Y.Y., et al.: Recover and identify: a generative dual model for cross-resolution person re-identification. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 8090\u20138099 (2019)","key":"7_CR32","DOI":"10.1109\/ICCV.2019.00818"},{"doi-asserted-by":"crossref","unstructured":"Sarfraz, M.S., Schumann, A., Eberle, A., et al.: A pose-sensitive embedding for person re-identification with expanded cross neighborhood re-ranking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 420\u2013429 (2018)","key":"7_CR33","DOI":"10.1109\/CVPR.2018.00051"}],"container-title":["Lecture Notes in Computer Science","Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-93049-3_7","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T06:25:09Z","timestamp":1641018309000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-93049-3_7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"ISBN":["9783030930486","9783030930493"],"references-count":33,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-93049-3_7","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2021]]},"assertion":[{"value":"1 January 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"CICAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"CAAI International Conference on Artificial Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Hangzhou","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5 June 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6 June 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"cicai2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/cicai.caai.cn\/#\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"307","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"105","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"34% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.2","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5.3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}