{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T16:42:43Z","timestamp":1743093763799,"version":"3.40.3"},"publisher-location":"Cham","reference-count":43,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783031065545"},{"type":"electronic","value":"9783031065552"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-06555-2_7","type":"book-chapter","created":{"date-parts":[[2022,5,17]],"date-time":"2022-05-17T09:10:13Z","timestamp":1652778613000},"page":"96-110","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Contrastive Graph Learning with\u00a0Graph Convolutional Networks"],"prefix":"10.1007","author":[{"given":"G.","family":"Nagendar","sequence":"first","affiliation":[]},{"given":"Ramachandrula","family":"Sitaram","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,5,18]]},"reference":[{"key":"7_CR1","doi-asserted-by":"crossref","unstructured":"Huang, Z., et al.: ICDAR2019 competition on scanned receipt OCR and information extraction. In: ICDAR (2019)","DOI":"10.1109\/ICDAR.2019.00244"},{"key":"7_CR2","doi-asserted-by":"crossref","unstructured":"Lample, G., Ballesteros, M., Subramanian, S., Kawakami, K., Dyer, C.: Neural architectures for named entity recognition. In: NAACL (2016)","DOI":"10.18653\/v1\/N16-1030"},{"key":"7_CR3","doi-asserted-by":"crossref","unstructured":"Esser, D., Schuster, D., Muthmann, K., Berger, M., Schill, A.: Automatic indexing of scanned documents: a layout-based approach. In: DRR (2012)","DOI":"10.1117\/12.908542"},{"key":"7_CR4","first-page":"102","volume":"6","author":"F Cesarini","year":"2003","unstructured":"Cesarini, F., Francesconi, E., Gori, M., Soda, G.: Analysis and understanding of multi-class invoices. DAS 6, 102\u2013114 (2003)","journal-title":"DAS"},{"key":"7_CR5","doi-asserted-by":"crossref","unstructured":"Simon, A., Pret, J.-C., Johnson, A.P.: A fast algorithm for bottom-up document layout analysis. In: PAMI (1997)","DOI":"10.1109\/34.584106"},{"key":"7_CR6","unstructured":"Kipf, T.N., Welling, M.: Semi-supervised classification with graph convolutional networks. In: ICLR (2017)"},{"key":"7_CR7","unstructured":"Veli\u010dkovi\u0107, P., Cucurull, G., Casanova, A., Romero, A., Li\u00f2, P., Bengio, Y.: Graph attention networks. In: ICLR (2017)"},{"key":"7_CR8","doi-asserted-by":"crossref","unstructured":"Xu, Y., et al.: LayoutLMv2: multi-modal pre-training for visually-rich document understanding. arXiv (2020)","DOI":"10.18653\/v1\/2021.acl-long.201"},{"key":"7_CR9","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"532","DOI":"10.1007\/978-3-030-86549-8_34","volume-title":"Document Analysis and Recognition \u2013 ICDAR 2021","author":"\u0141 Garncarek","year":"2021","unstructured":"Garncarek, \u0141, et al.: LAMBERT: layout-aware language modeling for information extraction. In: Llad\u00f3s, J., Lopresti, D., Uchida, S. (eds.) ICDAR 2021. LNCS, vol. 12821, pp. 532\u2013547. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-86549-8_34"},{"key":"7_CR10","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"548","DOI":"10.1007\/978-3-030-86549-8_35","volume-title":"Document Analysis and Recognition \u2013 ICDAR 2021","author":"W Lin","year":"2021","unstructured":"Lin, W., et al.: ViBERTgrid: a jointly trained multi-modal 2D document representation for key information extraction from documents. In: Llad\u00f3s, J., Lopresti, D., Uchida, S. (eds.) ICDAR 2021. LNCS, vol. 12821, pp. 548\u2013563. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-86549-8_35"},{"key":"7_CR11","doi-asserted-by":"crossref","unstructured":"Liu, X., Gao, F., Zhang, Q., Zhao, H.: Graph convolution for multimodal information extraction from visually rich documents. In: NAACL (2019)","DOI":"10.18653\/v1\/N19-2005"},{"key":"7_CR12","unstructured":"Yu, W., Lu, N., Qi, X., Gong, P., Xiao, R.: PICK: processing key information extraction from documents using improved graph learning-convolutional networks. In: ICPR (2020)"},{"key":"7_CR13","unstructured":"Henaff, M., Bruna, J., LeCun, Y.: Deep convolutional networks on graph-structured data. arXiv (2015)"},{"key":"7_CR14","unstructured":"Zhu, F., Huang, J., Li, R., Wang, S.: Adaptive graph convolutional neural networks. In: AAAI (2018)"},{"key":"7_CR15","doi-asserted-by":"crossref","unstructured":"Jiang, B., Zhang, Z., Lin, D., Tang, J., Luo, B.: Semi-supervised learning with graph learning-convolutional networks. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.01157"},{"key":"7_CR16","series-title":"Springer Series in Statistics","doi-asserted-by":"publisher","DOI":"10.1007\/978-0-387-21606-5","volume-title":"The Elements of Statistical Learning","author":"T Hastie","year":"2001","unstructured":"Hastie, T., Tibshirani, R., Friedman, J.: The Elements of Statistical Learning. Springer Series in Statistics, Springer, New York (2001)"},{"key":"7_CR17","unstructured":"Hadsell, R., Chopra, S., LeCun, Y.: Dimensionality reduction by learning an invariant mapping. In: CVPR (2006)"},{"key":"7_CR18","unstructured":"Khosla, P., et al.: Supervised contrastive learning. In: NIPS (2020)"},{"key":"7_CR19","unstructured":"Ge, Y., Zhu, F., Chen, D., Zhao, R., Li, H.: Self-paced contrastive learning with hybrid memory for domain adaptive object Re-ID. In: NIPS (2020)"},{"key":"7_CR20","unstructured":"Zhang, Y., Zhang, X., Qiu, R.C., Li, J., Xu, H., Tian, Q.: Semi-supervised contrastive learning with similarity co-calibration. CoRR abs\/2105.07387 (2021)"},{"key":"7_CR21","unstructured":"You, Y., Chen, T., Sui, Y., Chen, T., Wang, Z., Shen, Y.: Graph contrastive learning with augmentations. In: NIPS (2020)"},{"key":"7_CR22","unstructured":"You, Y., Chen, T., Shen, Y., Wang, Z.: Graph contrastive learning automated. In: ICML (2021)"},{"key":"7_CR23","unstructured":"Chen, T., Kornblith, S., Norouzi, M., Hinton, G.: A simple framework for contrastive learning of visual representations. arXiv (2020)"},{"key":"7_CR24","doi-asserted-by":"crossref","unstructured":"He, K., Fan, H., Wu, Y., Xie, S., Girshick, R.: Momentum contrast for unsupervised visual representation learning. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"7_CR25","unstructured":"Chen, T., Kornblith, S., Norouzi, M., Hinton, G.: A simple framework for contrastive learning of visual representations. In: ICML (2020)"},{"key":"7_CR26","unstructured":"Vaswani, A., et al.: Attention is all you need. In: NIPS (2017)"},{"key":"7_CR27","doi-asserted-by":"crossref","unstructured":"Graves, A., Schmidhuber, J.: Framewise phoneme classification with bidirectional LSTM networks. In: IJCNN (2005)","DOI":"10.1016\/j.neunet.2005.06.042"},{"key":"7_CR28","unstructured":"Lafferty, J., McCallum, A., Pereira, F.C.: Conditional random fields: probabilistic models for segmenting and labeling sequence data. In: ICML (2001)"},{"key":"7_CR29","doi-asserted-by":"crossref","unstructured":"Xu, Y., Li, M., Cui, L., Huang, S., Wei, F., Zhou, M.: LayoutLM: pre-training of text and layout for document image understanding. In: SIGKDD (2020)","DOI":"10.1145\/3394486.3403172"},{"key":"7_CR30","unstructured":"Devlin, J., Chang, M.-W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: NAACL (2019)"},{"key":"7_CR31","doi-asserted-by":"crossref","unstructured":"Ben-younes, H., Cadene, R., Thome, N., Cord, M.: BLOCK: bilinear superdiagonal fusion for visual question answering and visual relationship detection. In: AAAI (2019)","DOI":"10.1609\/aaai.v33i01.33018102"},{"key":"7_CR32","doi-asserted-by":"crossref","unstructured":"De Lathauwer, L.: Decompositions of a higher-order tensor in block terms part II: definitions and uniqueness. In: SIMAX (2008)","DOI":"10.1137\/070690729"},{"key":"7_CR33","unstructured":"Zhang, Z., Liu, Q.: Road extraction by deep residual U-Net. In: GRSL (2017)"},{"key":"7_CR34","doi-asserted-by":"crossref","unstructured":"Diakogiannis, F.I., Waldner, F., Caccetta, P., Wu, C.: ResUNet-a: a deep learning framework for semantic segmentation of remotely sensed data. In: ISPRS (2020)","DOI":"10.1016\/j.isprsjprs.2020.01.013"},{"key":"7_CR35","unstructured":"Hendrycks, D., Gimpel, K.: Gaussian error linear units (GELUs). arXiv (2016)"},{"key":"7_CR36","doi-asserted-by":"crossref","unstructured":"Liao, M., Wan, Z., Yao, C., Chen, K., Bai, X.: Real-time scene text detection with differentiable binarization. In: AAAI (2020)","DOI":"10.1609\/aaai.v34i07.6812"},{"key":"7_CR37","doi-asserted-by":"crossref","unstructured":"Shi, B., Bai, X., Yao, C.: An end-to-end trainable neural network for image-based sequence recognition and its application to scene text recognition. In: PAMI (2017)","DOI":"10.1109\/TPAMI.2016.2646371"},{"key":"7_CR38","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv (2014)"},{"key":"7_CR39","unstructured":"Park, S., et al.: CORD: a consolidated receipt dataset for post-OCR parsing. In: Document Intelligence Workshop at NeurIPS (2019)"},{"key":"7_CR40","doi-asserted-by":"crossref","unstructured":"Powalski, R., Borchmann, \u0141., Jurkiewicz, D., Dwojak, T., Pietruszka, M., Pa\u0142ka, G.: Going full-tilt boogie on document understanding with textimage-layout transformer. arXiv (2021)","DOI":"10.1007\/978-3-030-86331-9_47"},{"key":"7_CR41","doi-asserted-by":"crossref","unstructured":"Hwang, W., Yim, J., Park, S., Yang, S., Seo, M.: Spatial dependency parsing for semi-structured document information extraction. In: ACL-IJCNLP (2021)","DOI":"10.18653\/v1\/2021.findings-acl.28"},{"key":"7_CR42","doi-asserted-by":"crossref","unstructured":"Ma, X., Hovy, E.: End-to-end sequence labeling via bi-directional LSTM-CNNsCRF. In: ACL (2016)","DOI":"10.18653\/v1\/P16-1101"},{"key":"7_CR43","doi-asserted-by":"crossref","unstructured":"Appalaraju, S., Jasani, B., Kota, B.U., Xie, Y., Manmatha, R.: DocFormer: end-to-end transformer for document understanding. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00103"}],"container-title":["Lecture Notes in Computer Science","Document Analysis Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-06555-2_7","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,12]],"date-time":"2024-03-12T16:03:37Z","timestamp":1710259417000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-06555-2_7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031065545","9783031065552"],"references-count":43,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-06555-2_7","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"18 May 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"DAS","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Workshop on Document Analysis Systems","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"La Rochelle","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"France","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 May 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"25 May 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"das2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/das2022.univ-lr.fr\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"easychair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"94","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"52","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"16","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"55% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.85","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}