{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,10]],"date-time":"2026-04-10T10:05:14Z","timestamp":1775815514188,"version":"3.50.1"},"publisher-location":"Cham","reference-count":47,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031560590","type":"print"},{"value":"9783031560606","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-56060-6_2","type":"book-chapter","created":{"date-parts":[[2024,3,15]],"date-time":"2024-03-15T15:02:17Z","timestamp":1710514937000},"page":"18-33","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":9,"title":["CrisisKAN: Knowledge-Infused and\u00a0Explainable Multimodal Attention Network for\u00a0Crisis Event Classification"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-4908-843X","authenticated-orcid":false,"given":"Shubham","family":"Gupta","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1736-943X","authenticated-orcid":false,"given":"Nandini","family":"Saini","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7856-4768","authenticated-orcid":false,"given":"Suman","family":"Kundu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6205-4096","authenticated-orcid":false,"given":"Debasis","family":"Das","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,3,16]]},"reference":[{"key":"2_CR1","doi-asserted-by":"crossref","unstructured":"Abavisani, M., Wu, L., Hu, S., Tetreault, J., Jaimes, A.: Multimodal categorization of crisis events in social media. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 14679\u201314689 (2020)","DOI":"10.1109\/CVPR42600.2020.01469"},{"key":"2_CR2","doi-asserted-by":"publisher","first-page":"52138","DOI":"10.1109\/ACCESS.2018.2870052","volume":"6","author":"A Adadi","year":"2018","unstructured":"Adadi, A., Berrada, M.: Peeking inside the black-box: a survey on explainable artificial intelligence (XAI). IEEE Access 6, 52138\u201352160 (2018)","journal-title":"IEEE Access"},{"key":"2_CR3","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"479","DOI":"10.1007\/978-3-030-45439-5_32","volume-title":"Advances in Information Retrieval","author":"M Agarwal","year":"2020","unstructured":"Agarwal, M., Leekha, M., Sawhney, R., Ratn Shah, R., Kumar Yadav, R., Kumar Vishwakarma, D.: MEMIS: multimodal emergency management information system. In: Jose, J.M., et al. (eds.) ECIR 2020. LNCS, vol. 12035, pp. 479\u2013494. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-45439-5_32"},{"key":"2_CR4","doi-asserted-by":"crossref","unstructured":"Alam, F., Ofli, F., Imran, M.: CrisisMMD: multimodal twitter datasets from natural disasters. In: Proceedings of the International AAAI Conference on Web and Social Media. vol. 12 (2018)","DOI":"10.1609\/icwsm.v12i1.14983"},{"key":"2_CR5","unstructured":"Anonymous: EA2n: Evidence-based AMR attention network for fake news detection. In: Submitted to The Twelfth International Conference on Learning Representations (2023). https:\/\/openreview.net\/forum?id=5rrYpa2vts, under review"},{"key":"2_CR6","unstructured":"Arevalo, J., Solorio, T., Montes-y G\u00f3mez, M., Gonz\u00e1lez, F.A.: Gated multimodal units for information fusion (2017). arXiv preprint arXiv:1702.01992"},{"key":"2_CR7","doi-asserted-by":"publisher","unstructured":"Bandyopadhyay, D., Kumari, G., Ekbal, A., Pal, S., Chatterjee, A., BN, V.: A knowledge infusion based multitasking system for sarcasm detection in meme. In: Kamps, J., et al. Advances in Information Retrieval. ECIR 2023. LNCS, vol. 13980. Springer, Cham (2023). https:\/\/doi.org\/10.1007\/978-3-031-28244-7_7","DOI":"10.1007\/978-3-031-28244-7_7"},{"key":"2_CR8","doi-asserted-by":"publisher","unstructured":"Chu, S.Y., Lee, M.S.: MT-DETR: robust end-to-end multimodal detection with confidence fusion. In: 2023 IEEE\/CVF Winter Conference on Applications of Computer Vision (WACV), pp. 5241\u20135250 (2023). https:\/\/doi.org\/10.1109\/WACV56688.2023.00522","DOI":"10.1109\/WACV56688.2023.00522"},{"key":"2_CR9","unstructured":"Clark, K., Luong, M., Le, Q.V., Manning, C.D.: ELECTRA: pre-training text encoders as discriminators rather than generators. In: ICLR. OpenReview.net (2020)"},{"key":"2_CR10","doi-asserted-by":"crossref","unstructured":"Dai, Z., Yang, Z., Yang, Y., Carbonell, J., Le, Q.V., Salakhutdinov, R.: Transformer-XL: Attentive language models beyond a fixed-length context (2019). arXiv preprint arXiv:1901.02860","DOI":"10.18653\/v1\/P19-1285"},{"key":"2_CR11","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: Pre-training of deep bidirectional transformers for language understanding (2018). arXiv preprint arXiv:1810.04805"},{"key":"2_CR12","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: Pre-training of deep bidirectional transformers for language understanding. In: NAACL Volume 1 (Long and Short Papers), pp. 4171\u20134186. ACL, Minneapolis, Minnesota (2019)"},{"key":"2_CR13","doi-asserted-by":"crossref","unstructured":"Ferragina, P., Scaiella, U.: TAGME: on-the-fly annotation of short text fragments (by wikipedia entities). In: ICIKM, pp. 1625\u20131628. CIKM 2010, ACM, New York, NY, USA (2010)","DOI":"10.1145\/1871437.1871689"},{"key":"2_CR14","doi-asserted-by":"crossref","unstructured":"Fukui, A., Park, D.H., Yang, D., Rohrbach, A., Darrell, T., Rohrbach, M.: Multimodal compact bilinear pooling for visual question answering and visual grounding (2016). arXiv preprint arXiv:1606.01847","DOI":"10.18653\/v1\/D16-1044"},{"key":"2_CR15","doi-asserted-by":"crossref","unstructured":"Gallo, I., Ria, G., Landro, N., La Grassa, R.: Image and text fusion for UPMC food-101 using BERT and CNNs. In: 2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ), pp. 1\u20136. IEEE (2020)","DOI":"10.1109\/IVCNZ51579.2020.9290622"},{"key":"2_CR16","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2023.120890","volume":"232","author":"S Gupta","year":"2023","unstructured":"Gupta, S., Kundu, S.: Interaction graph, topical communities, and efficient local event detection from social streams. Expert Syst. Appl. 232, 120890 (2023)","journal-title":"Expert Syst. Appl."},{"key":"2_CR17","doi-asserted-by":"crossref","unstructured":"Gupta, S., Yadav, N., Sainath Reddy, S., Kundu, S.: FakEDAMR: Fake news detection using abstract meaning representation (2023)","DOI":"10.36227\/techrxiv.24115896"},{"key":"2_CR18","doi-asserted-by":"crossref","unstructured":"Han, K., Wang, Y., Tian, Q., Guo, J., Xu, C., Xu, C.: GhostNet: more features from cheap operations. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1577\u20131586 (2020)","DOI":"10.1109\/CVPR42600.2020.00165"},{"key":"2_CR19","doi-asserted-by":"crossref","unstructured":"He, K., Gkioxari, G., Doll\u00e1r, P., Girshick, R.: Mask R-CNN. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2961\u20132969 (2017)","DOI":"10.1109\/ICCV.2017.322"},{"key":"2_CR20","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1016\/j.inffus.2021.01.008","volume":"71","author":"A Holzinger","year":"2021","unstructured":"Holzinger, A., Malle, B., Saranti, A., Pfeifer, B.: Towards multi-modal causability with graph neural networks enabling information fusion for explainable AI. Inf. Fusion 71, 28\u201337 (2021)","journal-title":"Inf. Fusion"},{"key":"2_CR21","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., Van Der Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2261\u20132269 (2017)","DOI":"10.1109\/CVPR.2017.243"},{"key":"2_CR22","unstructured":"Huang, Z., Zeng, Z., Liu, B., Fu, D., Fu, J.: Pixel-BERT: Aligning image pixels with text by deep multi-modal transformers (2020). arXiv preprint arXiv:2004.00849"},{"key":"2_CR23","doi-asserted-by":"publisher","unstructured":"Hubenthal, M., Kumar, S.: Image-text pre-training for logo recognition. In: 2023 IEEE\/CVF Winter Conference on Applications of Computer Vision (WACV), pp. 1145\u20131154 (2023). https:\/\/doi.org\/10.1109\/WACV56688.2023.00120","DOI":"10.1109\/WACV56688.2023.00120"},{"issue":"1","key":"2_CR24","doi-asserted-by":"publisher","first-page":"861","DOI":"10.1007\/s11069-020-04016-6","volume":"103","author":"K Hunt","year":"2020","unstructured":"Hunt, K., Wang, B., Zhuang, J.: Misinformation debunking and cross-platform information sharing through Twitter during Hurricanes Harvey and Irma: a case study on shelters and ID checks. Nat. Hazards 103(1), 861\u2013883 (2020). https:\/\/doi.org\/10.1007\/s11069-020-04016-6","journal-title":"Nat. Hazards"},{"key":"2_CR25","doi-asserted-by":"publisher","first-page":"59800","DOI":"10.1109\/ACCESS.2021.3070212","volume":"9","author":"G Joshi","year":"2021","unstructured":"Joshi, G., Walambe, R., Kotecha, K.: A review on explainability in multimodal deep neural nets. IEEE Access 9, 59800\u201359821 (2021). https:\/\/doi.org\/10.1109\/ACCESS.2021.3070212","journal-title":"IEEE Access"},{"key":"2_CR26","unstructured":"Kiela, D., Bhooshan, S., Firooz, H., Perez, E., Testuggine, D.: Supervised multimodal bitransformers for classifying images and text (2019). arXiv preprint arXiv:1909.02950"},{"key":"2_CR27","doi-asserted-by":"crossref","unstructured":"Kiela, D., Grave, E., Joulin, A., Mikolov, T.: Efficient large-scale multi-modal classification. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol. 32 (2018)","DOI":"10.1609\/aaai.v32i1.11945"},{"key":"2_CR28","unstructured":"Kim, W., Son, B., Kim, I.: ViLT: vision-and-language transformer without convolution or region supervision. In: International Conference on Machine Learning, pp. 5583\u20135594. PMLR (2021)"},{"key":"2_CR29","unstructured":"Krizhevsky, A.: One weird trick for parallelizing convolutional neural networks (2014). CoRR abs\/1404.5997"},{"key":"2_CR30","doi-asserted-by":"crossref","unstructured":"Kwan, J.S.L., Lim, K.H.: Understanding public sentiments, opinions and topics about COVID-19 using twitter. In: 2020 IEEE\/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM), pp. 623\u2013626. IEEE (2020)","DOI":"10.1109\/ASONAM49781.2020.9381384"},{"key":"2_CR31","doi-asserted-by":"crossref","unstructured":"Li, G., Duan, N., Fang, Y., Gong, M., Jiang, D.: Unicoder-VL: a universal encoder for vision and language by cross-modal pre-training. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol. 34, pp. 11336\u201311344 (2020)","DOI":"10.1609\/aaai.v34i07.6795"},{"key":"2_CR32","unstructured":"Li, L.H., Yatskar, M., Yin, D., Hsieh, C.J., Chang, K.W.: VisualBERT: A simple and performant baseline for vision and language (2019). arXiv preprint arXiv:1908.03557"},{"key":"2_CR33","doi-asserted-by":"publisher","unstructured":"Liang, T., Lin, G., Wan, M., Li, T., Ma, G., Lv, F.: Expanding large pre-trained unimodal models with multimodal information injection for image-text multimodal classification. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 15471\u201315480 (2022). https:\/\/doi.org\/10.1109\/CVPR52688.2022.01505","DOI":"10.1109\/CVPR52688.2022.01505"},{"key":"2_CR34","doi-asserted-by":"publisher","unstructured":"Long, S., Han, S.C., Wan, X., Poon, J.: GraDual: graph-based dual-modal representation for image-text matching. In: 2022 IEEE\/CVF Winter Conference on Applications of Computer Vision (WACV), pp. 2463\u20132472 (2022). https:\/\/doi.org\/10.1109\/WACV51458.2022.00252","DOI":"10.1109\/WACV51458.2022.00252"},{"key":"2_CR35","doi-asserted-by":"crossref","unstructured":"Mao, X., et al.: Towards robust vision transformer. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12042\u201312051 (2022)","DOI":"10.1109\/CVPR52688.2022.01173"},{"key":"2_CR36","doi-asserted-by":"publisher","first-page":"56044","DOI":"10.1109\/ACCESS.2022.3176965","volume":"10","author":"H Moraliyage","year":"2022","unstructured":"Moraliyage, H., Sumanasena, V., De Silva, D., Nawaratne, R., Sun, L., Alahakoon, D.: Multimodal classification of onion services for proactive cyber threat intelligence using explainable deep learning. IEEE Access 10, 56044\u201356056 (2022)","journal-title":"IEEE Access"},{"key":"2_CR37","doi-asserted-by":"publisher","unstructured":"Nazura, J., Muralidhara, B.L.: Semantic classification of tweets: a contextual knowledge based approach for tweet classification. In: 2017 8th International Conference on Information, Intelligence, Systems & Applications (IISA), pp. 1\u20136 (2017). https:\/\/doi.org\/10.1109\/IISA.2017.8316358","DOI":"10.1109\/IISA.2017.8316358"},{"key":"2_CR38","doi-asserted-by":"crossref","unstructured":"Petsiuk, V., et al.: Black-box explanation of object detectors via saliency maps. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11443\u201311452 (2021)","DOI":"10.1109\/CVPR46437.2021.01128"},{"key":"2_CR39","doi-asserted-by":"crossref","unstructured":"Sakaki, T., Okazaki, M., Matsuo, Y.: Earthquake shakes twitter users: real-time event detection by social sensors. In: Proceedings of the 19th International Conference on World Wide Web, pp. 851\u2013860 (2010)","DOI":"10.1145\/1772690.1772777"},{"key":"2_CR40","doi-asserted-by":"publisher","unstructured":"Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., Batra, D.: Grad-CAM: visual explanations from deep networks via gradient-based localization. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp. 618\u2013626 (2017). https:\/\/doi.org\/10.1109\/ICCV.2017.74","DOI":"10.1109\/ICCV.2017.74"},{"key":"2_CR41","doi-asserted-by":"crossref","unstructured":"Shu, K., Zhou, X., Wang, S., Zafarani, R., Liu, H.: The role of user profiles for fake news detection. In: IEEE\/ACM ASONAM, pp. 436\u2013439. ASONAM 2019, Association for Computing Machinery, New York, NY, USA (2020)","DOI":"10.1145\/3341161.3342927"},{"key":"2_CR42","doi-asserted-by":"crossref","unstructured":"Singh, A., et al.: FLAVA: a foundational language and vision alignment model. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 15638\u201315650 (2022)","DOI":"10.1109\/CVPR52688.2022.01519"},{"issue":"2","key":"2_CR43","doi-asserted-by":"publisher","first-page":"138","DOI":"10.5614\/itbj.ict.res.appl.2022.16.2.3","volume":"16","author":"B Tahayna","year":"2022","unstructured":"Tahayna, B., Ayyasamy, R., Akbar, R.: Context-aware sentiment analysis using tweet expansion method. J. ICT Res. Appl. 16(2), 138\u2013151 (2022)","journal-title":"J. ICT Res. Appl."},{"key":"2_CR44","doi-asserted-by":"crossref","unstructured":"Vielzeuf, V., Lechervy, A., Pateux, S., Jurie, F.: CentralNet: a multilayer approach for multimodal fusion. In: Proceedings of the European Conference on Computer Vision (ECCV) Workshops (2018)","DOI":"10.1007\/978-3-030-11024-6_44"},{"key":"2_CR45","doi-asserted-by":"crossref","unstructured":"Wang, Y., Xu, X., Yu, W., Xu, R., Cao, Z., Shen, H.T.: Combine early and late fusion together: a hybrid fusion framework for image-text matching. In: 2021 IEEE International Conference on Multimedia and Expo (ICME), pp. 1\u20136. IEEE (2021)","DOI":"10.1109\/ICME51207.2021.9428201"},{"key":"2_CR46","volume-title":"XLNet: Generalized Autoregressive Pretraining for Language Understanding","author":"Z Yang","year":"2019","unstructured":"Yang, Z., Dai, Z., Yang, Y., Carbonell, J., Salakhutdinov, R., Le, Q.V.: XLNet: Generalized Autoregressive Pretraining for Language Understanding. Curran Associates Inc., Red Hook, NY, USA (2019)"},{"key":"2_CR47","doi-asserted-by":"crossref","unstructured":"Zadeh, A., Chen, M., Poria, S., Cambria, E., Morency, L.P.: Tensor fusion network for multimodal sentiment analysis (2017). arXiv preprint arXiv:1707.07250","DOI":"10.18653\/v1\/D17-1115"}],"container-title":["Lecture Notes in Computer Science","Advances in Information Retrieval"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-56060-6_2","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,14]],"date-time":"2024-11-14T11:10:03Z","timestamp":1731582603000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-56060-6_2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9783031560590","9783031560606"],"references-count":47,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-56060-6_2","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"16 March 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECIR","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Information Retrieval","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Glasgow","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"United Kingdom","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"24 March 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 March 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ecir2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.ecir2024.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"578","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"110","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"69","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"19% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"31 (Tracks: Workshop, Tutorial, Industry, Doctoral Consortium)","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}