{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T16:21:39Z","timestamp":1775578899903,"version":"3.50.1"},"publisher-location":"Cham","reference-count":61,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031197741","type":"print"},{"value":"9783031197758","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19775-8_33","type":"book-chapter","created":{"date-parts":[[2022,10,22]],"date-time":"2022-10-22T12:12:59Z","timestamp":1666440779000},"page":"562-579","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":34,"title":["Label2Label: A Language Modeling Framework for\u00a0Multi-attribute Learning"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2730-0543","authenticated-orcid":false,"given":"Wanhua","family":"Li","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0261-4105","authenticated-orcid":false,"given":"Zhexuan","family":"Cao","sequence":"additional","affiliation":[]},{"given":"Jianjiang","family":"Feng","sequence":"additional","affiliation":[]},{"given":"Jie","family":"Zhou","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6121-5529","authenticated-orcid":false,"given":"Jiwen","family":"Lu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,23]]},"reference":[{"issue":"11","key":"33_CR1","first-page":"1949","volume":"17","author":"AH Abdulnabi","year":"2015","unstructured":"Abdulnabi, A.H., Wang, G., Lu, J., Jia, K.: Multi-task cnn model for attribute prediction. TMM 17(11), 1949\u20131959 (2015)","journal-title":"TMM"},{"key":"33_CR2","doi-asserted-by":"crossref","unstructured":"Ak, K.E., Kassim, A.A., Lim, J.H., Tham, J.Y.: Learning attribute representations with localization for flexible fashion search. In: CVPR, pp. 7708\u20137717 (2018)","DOI":"10.1109\/CVPR.2018.00804"},{"key":"33_CR3","unstructured":"Bao, H., Dong, L., Wei, F.: Beit: bert pre-training of image transformers. arXiv preprint arXiv:2106.08254 (2021)"},{"key":"33_CR4","unstructured":"Brown, T.B., et al.: Language models are few-shot learners. In: NeurIPS (2020)"},{"key":"33_CR5","doi-asserted-by":"crossref","unstructured":"Cao, J., Li, Y., Zhang, Z.: Partially shared multi-task convolutional neural network with local constraint for face attribute learning. In: CVPR, pp. 4290\u20134299 (2018)","DOI":"10.1109\/CVPR.2018.00451"},{"key":"33_CR6","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"213","DOI":"10.1007\/978-3-030-58452-8_13","volume-title":"Computer Vision \u2013 ECCV 2020","author":"N Carion","year":"2020","unstructured":"Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., Zagoruyko, S.: End-to-end object detection with transformers. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 213\u2013229. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_13"},{"key":"33_CR7","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"609","DOI":"10.1007\/978-3-642-33712-3_44","volume-title":"Computer Vision \u2013 ECCV 2012","author":"H Chen","year":"2012","unstructured":"Chen, H., Gallagher, A., Girod, B.: Describing clothing by semantic attributes. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012. LNCS, vol. 7574, pp. 609\u2013623. Springer, Heidelberg (2012). https:\/\/doi.org\/10.1007\/978-3-642-33712-3_44"},{"key":"33_CR8","unstructured":"Chen, T., Saxena, S., Li, L., Fleet, D.J., Hinton, G.: Pix2seq: a language modeling framework for object detection. arXiv preprint arXiv:2109.10852 (2021)"},{"key":"33_CR9","doi-asserted-by":"crossref","unstructured":"Cheng, X., et al.: Mltr: multi-label classification with transformer. arXiv preprint arXiv:2106.06195 (2021)","DOI":"10.1109\/ICME52920.2022.9860016"},{"key":"33_CR10","doi-asserted-by":"crossref","unstructured":"Cubuk, E.D., Zoph, B., Shlens, J., Le, Q.: Randaugment: practical automated data augmentation with a reduced search space. In: NeurIPS, pp. 18613\u201318624 (2020)","DOI":"10.1109\/CVPRW50498.2020.00359"},{"key":"33_CR11","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: Bert: pre-training of deep bidirectional transformers for language understanding. In: NAACL (2019)"},{"key":"33_CR12","unstructured":"Doersch, C., Gupta, A., Zisserman, A.: Crosstransformers: spatially-aware few-shot transfer. In: NeurIPS (2020)"},{"key":"33_CR13","unstructured":"Dosovitskiy, A., et al.: An image is worth 16$$\\times $$16 words: transformers for image recognition at scale. In: ICLR (2021)"},{"key":"33_CR14","unstructured":"Duan, K., Parikh, D., Crandall, D., Grauman, K.: Discovering localized attributes for fine-grained recognition. In: CVPR, pp. 3474\u20133481 (2012)"},{"key":"33_CR15","doi-asserted-by":"crossref","unstructured":"Farhadi, A., Endres, I., Hoiem, D., Forsyth, D.: Describing objects by their attributes. In: CVPR, pp. 1778\u20131785 (2009)","DOI":"10.1109\/CVPR.2009.5206772"},{"key":"33_CR16","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-50077-5","volume-title":"Visual Attributes","author":"RS Feris","year":"2017","unstructured":"Feris, R.S., Lampert, C., Parikh, D.: Visual Attributes. Springer, Cham (2017). https:\/\/doi.org\/10.1007\/978-3-319-50077-5"},{"key":"33_CR17","doi-asserted-by":"crossref","unstructured":"Guo, H., Zheng, K., Fan, X., Yu, H., Wang, S.: Visual attention consistency under image transforms for multi-label image classification. In: CVPR, pp. 729\u2013739 (2019)","DOI":"10.1109\/CVPR.2019.00082"},{"key":"33_CR18","doi-asserted-by":"crossref","unstructured":"Hand, E.M., Chellappa, R.: Attributes for improved attributes: a multi-task network utilizing implicit and explicit relationships for facial attribute classification. In: AAAI (2017)","DOI":"10.1609\/aaai.v31i1.11229"},{"key":"33_CR19","doi-asserted-by":"crossref","unstructured":"He, K., Xinlei, C., Xie, S., Li, Y., Doll\u00e1r, P., Girshick, R.: Masked autoencoders are scalable vision learners. arXiv preprint arXiv:2106.08254 (2021)","DOI":"10.1109\/CVPR52688.2022.01553"},{"key":"33_CR20","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"33_CR21","doi-asserted-by":"crossref","unstructured":"He, K., et al.: Harnessing synthesized abstraction images to improve facial attribute recognition. In: IJCAI, pp. 733\u2013740 (2018)","DOI":"10.24963\/ijcai.2018\/102"},{"key":"33_CR22","doi-asserted-by":"crossref","unstructured":"Huang, S., Li, X., Cheng, Z.Q., Zhang, Z., Hauptmann, A.: GNAS: a greedy neural architecture search method for multi-attribute learning. In: ACM MM, pp. 2049\u20132057 (2018)","DOI":"10.1145\/3240508.3240588"},{"key":"33_CR23","doi-asserted-by":"crossref","unstructured":"Jia, J., Chen, X., Huang, K.: Spatial and semantic consistency regularizations for pedestrian attribute recognition. In: ICCV, pp. 962\u2013971 (2021)","DOI":"10.1109\/ICCV48922.2021.00100"},{"key":"33_CR24","doi-asserted-by":"crossref","unstructured":"Kalayeh, M.M., Gong, B., Shah, M.: Improving facial attribute prediction using semantic segmentation. In: CVPR, pp. 6942\u20136950 (2017)","DOI":"10.1109\/CVPR.2017.450"},{"key":"33_CR25","doi-asserted-by":"crossref","unstructured":"Lanchantin, J., Wang, T., Ordonez, V., Qi, Y.: General multi-label image classification with transformers. In: CVPR, pp. 16478\u201316488 (2021)","DOI":"10.1109\/CVPR46437.2021.01621"},{"key":"33_CR26","doi-asserted-by":"crossref","unstructured":"Li, D., Chen, X., Huang, K.: Multi-attribute learning for pedestrian attribute recognition in surveillance scenarios. In: ACPR, pp. 111\u2013115 (2015)","DOI":"10.1109\/ACPR.2015.7486476"},{"key":"33_CR27","doi-asserted-by":"crossref","unstructured":"Li, D., Chen, X., Zhang, Z., Huang, K.: Pose guided deep model for pedestrian attribute recognition in surveillance scenarios. In: ICME, pp. 1\u20136 (2018)","DOI":"10.1109\/ICME.2018.8486604"},{"issue":"4","key":"33_CR28","first-page":"1575","volume":"28","author":"D Li","year":"2018","unstructured":"Li, D., Zhang, Z., Chen, X., Huang, K.: A richly annotated pedestrian dataset for person retrieval in real surveillance scenarios. TIP 28(4), 1575\u20131590 (2018)","journal-title":"TIP"},{"issue":"9","key":"33_CR29","first-page":"4651","volume":"27","author":"J Li","year":"2018","unstructured":"Li, J., Zhao, F., Feng, J., Roy, S., Yan, S., Sim, T.: Landmark free face attribute prediction. TIP 27(9), 4651\u20134662 (2018)","journal-title":"TIP"},{"key":"33_CR30","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"18","DOI":"10.1007\/978-3-030-58555-6_2","volume-title":"Computer Vision \u2013 ECCV 2020","author":"W Li","year":"2020","unstructured":"Li, W., Duan, Y., Lu, J., Feng, J., Zhou, J.: Graph-based social relation reasoning. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12360, pp. 18\u201334. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58555-6_2"},{"key":"33_CR31","doi-asserted-by":"crossref","unstructured":"Li, W., Huang, X., Lu, J., Feng, J., Zhou, J.: Learning probabilistic ordinal embeddings for uncertainty-aware regression. In: CVPR, pp. 13896\u201313905 (2021)","DOI":"10.1109\/CVPR46437.2021.01368"},{"key":"33_CR32","doi-asserted-by":"crossref","unstructured":"Li, W., Lu, J., Feng, J., Xu, C., Zhou, J., Tian, Q.: Bridgenet: a continuity-aware probabilistic network for age estimation. In: CVPR, pp. 1145\u20131154 (2019)","DOI":"10.1109\/CVPR.2019.00124"},{"key":"33_CR33","unstructured":"Liu, P., Liu, X., Yan, J., Shao, J.: Localization guided learning for pedestrian attribute recognition. In: BMVC (2018)"},{"key":"33_CR34","unstructured":"Liu, S., Zhang, L., Yang, X., Su, H., Zhu, J.: Query2label: a simple transformer way to multi-label classification. arXiv preprint arXiv:2107.10834 (2021)"},{"key":"33_CR35","doi-asserted-by":"crossref","unstructured":"Liu, X., et al.: Hydraplus-net: attentive deep features for pedestrian analysis. In: ICCV, pp. 350\u2013359 (2017)","DOI":"10.1109\/ICCV.2017.46"},{"key":"33_CR36","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"33_CR37","doi-asserted-by":"crossref","unstructured":"Liu, Z., Luo, P., Qiu, S., Wang, X., Tang, X.: Deepfashion: powering robust clothes recognition and retrieval with rich annotations. In: CVPR, pp. 1096\u20131104 (2016)","DOI":"10.1109\/CVPR.2016.124"},{"key":"33_CR38","doi-asserted-by":"crossref","unstructured":"Liu, Z., Luo, P., Wang, X., Tang, X.: Deep learning face attributes in the wild. In: ICCV, pp. 3730\u20133738 (2015)","DOI":"10.1109\/ICCV.2015.425"},{"key":"33_CR39","unstructured":"Mao, L., Yan, Y., Xue, J.H., Wang, H.: Deep multi-task multi-label CNN for effective facial attribute classification. TAC (2020)"},{"key":"33_CR40","doi-asserted-by":"crossref","unstructured":"Meng, Z., Adluru, N., Kim, H.J., Fung, G., Singh, V.: Efficient relative attribute learning using graph neural networks. In: ECCV, pp. 552\u2013567 (2018)","DOI":"10.1007\/978-3-030-01264-9_34"},{"key":"33_CR41","doi-asserted-by":"crossref","unstructured":"Nguyen, H.D., Vu, X.S., Le, D.T.: Modular graph transformer networks for multi-label image classification. In: AAAI, pp. 9092\u20139100 (2021)","DOI":"10.1609\/aaai.v35i10.17098"},{"key":"33_CR42","doi-asserted-by":"crossref","unstructured":"Perrett, T., Masullo, A., Burghardt, T., Mirmehdi, M., Damen, D.: Temporal-relational crosstransformers for few-shot action recognition. In: CVPR, pp. 475\u2013484 (2021)","DOI":"10.1109\/CVPR46437.2021.00054"},{"key":"33_CR43","doi-asserted-by":"crossref","unstructured":"Peters, M.E., et al.: Deep contextualized word representations. In: NAACL (2018)","DOI":"10.18653\/v1\/N18-1202"},{"key":"33_CR44","unstructured":"Radford, A., Narasimhan, K., Salimans, T., Sutskever, I.: Improving language understanding by generative pre-training (2018)"},{"key":"33_CR45","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"19","DOI":"10.1007\/978-3-319-46454-1_2","volume-title":"Computer Vision \u2013 ECCV 2016","author":"EM Rudd","year":"2016","unstructured":"Rudd, E.M., G\u00fcnther, M., Boult, T.E.: MOON: a mixed objective optimization network for the recognition of facial attributes. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9909, pp. 19\u201335. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46454-1_2"},{"key":"33_CR46","doi-asserted-by":"crossref","unstructured":"Sarafianos, N., Xu, X., Kakadiaris, I.A.: Deep imbalanced attribute classification using visual attention aggregation. In: ECCV, pp. 680\u2013697 (2018)","DOI":"10.1007\/978-3-030-01252-6_42"},{"key":"33_CR47","unstructured":"Sarfraz, M.S., Schumann, A., Wang, Y., Stiefelhagen, R.: Deep view-sensitive pedestrian attribute inference in an end-to-end model. In: BMVC (2017)"},{"key":"33_CR48","doi-asserted-by":"crossref","unstructured":"Shao, J., Kang, K., Loy, C.C., Wang, X.: Deeply learned attributes for crowded scene understanding. In: CVPR, pp. 4657\u20134666 (2015)","DOI":"10.1109\/CVPR.2015.7299097"},{"key":"33_CR49","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"509","DOI":"10.1007\/978-3-030-58621-8_30","volume-title":"Computer Vision \u2013 ECCV 2020","author":"M Shin","year":"2020","unstructured":"Shin, M.: Semi-supervised learning with a teacher-student network for generalized attribute prediction. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12356, pp. 509\u2013525. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58621-8_30"},{"key":"33_CR50","doi-asserted-by":"crossref","unstructured":"Shu, Y., Yan, Y., Chen, S., Xue, J.H., Shen, C., Wang, H.: Learning spatial-semantic relationship for facial attribute recognition with limited labeled data. In: CVPR, pp. 11916\u201311925 (2021)","DOI":"10.1109\/CVPR46437.2021.01174"},{"key":"33_CR51","doi-asserted-by":"crossref","unstructured":"Tang, C., Sheng, L., Zhang, Z., Hu, X.: Improving pedestrian attribute recognition with weakly-supervised multi-scale attribute-specific localization. In: ICCV, pp. 4997\u20135006 (2019)","DOI":"10.1109\/ICCV.2019.00510"},{"issue":"4","key":"33_CR52","doi-asserted-by":"publisher","first-page":"415","DOI":"10.1177\/107769905303000401","volume":"30","author":"WL Taylor","year":"1953","unstructured":"Taylor, W.L.: \u201ccloze procedure\u2019\u2019: a new tool for measuring readability. Journalism Q. 30(4), 415\u2013433 (1953)","journal-title":"Journalism Q."},{"key":"33_CR53","unstructured":"Vaswani, A., et al.: Attention is all you need. In: NeurIPS, pp. 5998\u20136008 (2017)"},{"key":"33_CR54","doi-asserted-by":"crossref","unstructured":"Wang, Y., et al.: End-to-end video instance segmentation with transformers. In: CVPR, pp. 8741\u20138750 (2021)","DOI":"10.1109\/CVPR46437.2021.00863"},{"key":"33_CR55","doi-asserted-by":"crossref","unstructured":"Wang, Z.J., Turko, R., Chau, D.H.: Dodrio: exploring transformer models with interactive visualization. In: ACL (2021)","DOI":"10.18653\/v1\/2021.acl-demo.16"},{"key":"33_CR56","doi-asserted-by":"crossref","unstructured":"Yu, B., Li, W., Li, X., Lu, J., Zhou, J.: Frequency-aware spatiotemporal transformers for video inpainting detection. In: ICCV, pp. 8188\u20138197 (2021)","DOI":"10.1109\/ICCV48922.2021.00808"},{"key":"33_CR57","doi-asserted-by":"crossref","unstructured":"Zhang, C., Gupta, A., Zisserman, A.: Temporal query networks for fine-grained video understanding. In: CVPR, pp. 4486\u20134496 (2021)","DOI":"10.1109\/CVPR46437.2021.00446"},{"key":"33_CR58","unstructured":"Zhang, H., Cisse, M., Dauphin, Y.N., Lopez-Paz, D.: mixup: beyond empirical risk minimization. In: ICLR (2018)"},{"key":"33_CR59","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Zhang, P., Yuan, C., Wang, Z.: Texture and shape biased two-stream networks for clothing classification and attribute recognition. In: CVPR, pp. 13538\u201313547 (2020)","DOI":"10.1109\/CVPR42600.2020.01355"},{"key":"33_CR60","doi-asserted-by":"crossref","unstructured":"Zhao, X., et al.: Recognizing part attributes with insufficient data. In: ICCV, pp. 350\u2013360 (2019)","DOI":"10.1109\/ICCV.2019.00044"},{"key":"33_CR61","doi-asserted-by":"crossref","unstructured":"Zheng, S., et al.: Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers. In: CVPR, pp. 6881\u20136890 (2021)","DOI":"10.1109\/CVPR46437.2021.00681"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19775-8_33","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,13]],"date-time":"2024-03-13T13:48:38Z","timestamp":1710337718000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19775-8_33"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031197741","9783031197758"],"references-count":61,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19775-8_33","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"23 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}