{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,24]],"date-time":"2025-11-24T21:40:11Z","timestamp":1764020411792,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":40,"publisher":"ACM","license":[{"start":{"date-parts":[[2020,10,19]],"date-time":"2020-10-19T00:00:00Z","timestamp":1603065600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2020,10,19]]},"DOI":"10.1145\/3340531.3411908","type":"proceedings-article","created":{"date-parts":[[2020,10,19]],"date-time":"2020-10-19T05:31:05Z","timestamp":1603085465000},"page":"1725-1734","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":53,"title":["Beyond 512 Tokens: Siamese Multi-depth Transformer-based Hierarchical Encoder for Long-Form Document Matching"],"prefix":"10.1145","author":[{"given":"Liu","family":"Yang","sequence":"first","affiliation":[{"name":"Google Research, Mountain View, CA, USA"}]},{"given":"Mingyang","family":"Zhang","sequence":"additional","affiliation":[{"name":"Google Research, Mountain View, CA, USA"}]},{"given":"Cheng","family":"Li","sequence":"additional","affiliation":[{"name":"Google Research, Mountain View, CA, USA"}]},{"given":"Michael","family":"Bendersky","sequence":"additional","affiliation":[{"name":"Google Research, Mountain View, CA, USA"}]},{"given":"Marc","family":"Najork","sequence":"additional","affiliation":[{"name":"Google Research, Mountain View, CA, USA"}]}],"member":"320","published-online":{"date-parts":[[2020,10,19]]},"reference":[{"volume-title":"EMNLP '15","author":"Bowman S. R.","key":"e_1_3_2_2_1_1","unstructured":"S. R. Bowman , G. Angeli , C. Potts , and C. D. Manning . 2015. A large annotated corpus for learning natural language inference . In EMNLP '15 . 632--642. S. R. Bowman, G. Angeli, C. Potts, and C. D. Manning. 2015. A large annotated corpus for learning natural language inference. In EMNLP '15. 632--642."},{"key":"e_1_3_2_2_2_1","unstructured":"R. Child S. Gray A. Radford and I. Sutskever. 2019. Generating Long Sequences with Sparse Transformers. (2019). arXiv:1904.10509  R. Child S. Gray A. Radford and I. Sutskever. 2019. Generating Long Sequences with Sparse Transformers. (2019). arXiv:1904.10509"},{"key":"e_1_3_2_2_3_1","volume-title":"Overview of the TREC 2019 deep learning track. (2020","author":"Craswell N.","year":"2020","unstructured":"N. Craswell , B. Mitra , E. Yilmaz , D. Campos , and E. M Voorhees . 2020 . Overview of the TREC 2019 deep learning track. (2020 ). arXiv:2003.07820 N. Craswell, B. Mitra, E. Yilmaz, D. Campos, and E. M Voorhees. 2020. Overview of the TREC 2019 deep learning track. (2020). arXiv:2003.07820"},{"key":"e_1_3_2_2_4_1","volume":"201","author":"Dai Z.","unstructured":"Z. Dai and J. Callan. 201 9. Deeper Text Understanding for IR with Contextual Neural Language Modeling. In SIGIR '19. Z. Dai and J. Callan. 2019. Deeper Text Understanding for IR with Contextual Neural Language Modeling. In SIGIR '19.","journal-title":"J. Callan."},{"key":"e_1_3_2_2_5_1","doi-asserted-by":"crossref","unstructured":"Z. Dai Z. Yang Y. Yang J. G. Carbonell Q. V. Le and R. Salakhutdinov. 2019. Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context.(2019). arXiv:1901.02860  Z. Dai Z. Yang Y. Yang J. G. Carbonell Q. V. Le and R. Salakhutdinov. 2019. Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context.(2019). arXiv:1901.02860","DOI":"10.18653\/v1\/P19-1285"},{"key":"e_1_3_2_2_6_1","volume-title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding.","author":"Devlin J.","year":"2018","unstructured":"J. Devlin , M. Chang , K. Lee , and K. Toutanova . 2018 . BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. (2018). arXiv:1810.04805 J. Devlin, M. Chang, K. Lee, and K. Toutanova. 2018. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. (2018). arXiv:1810.04805"},{"key":"e_1_3_2_2_7_1","volume-title":"IWP","author":"Dolan W. B.","year":"2005","unstructured":"W. B. Dolan and C. Brockett . 2005. Automatically Constructing a Corpus of Sentential Paraphrases . In IWP 2005 . 9--16. W. B. Dolan and C. Brockett. 2005. Automatically Constructing a Corpus of Sentential Paraphrases. In IWP 2005. 9--16."},{"volume-title":"CIKM '16","author":"Guo J.","key":"e_1_3_2_2_8_1","unstructured":"J. Guo , Y. Fan , Q. Ai , and W. B. Croft . 2016. A Deep Relevance Matching Model for Ad-hoc Retrieval . In CIKM '16 . 55--64. J. Guo, Y. Fan, Q. Ai, and W. B. Croft. 2016. A Deep Relevance Matching Model for Ad-hoc Retrieval. In CIKM '16. 55--64."},{"key":"e_1_3_2_2_9_1","unstructured":"J. Guo Y. Fan L. Pang L. Yang Q. Ai H. Zamani C. Wu W. B. Croft and X. Cheng. 2019. A Deep Look into Neural Ranking Models for Information Retrieval.(2019). arXiv:1903.06902  J. Guo Y. Fan L. Pang L. Yang Q. Ai H. Zamani C. Wu W. B. Croft and X. Cheng. 2019. A Deep Look into Neural Ranking Models for Information Retrieval.(2019). arXiv:1903.06902"},{"key":"e_1_3_2_2_10_1","volume":"201","author":"He K.","unstructured":"K. He , X. Zhang , S. Ren , and J. Sun. 201 5. Deep Residual Learning for Image Recognition. (2015). arXiv:1512.03385 K. He, X. Zhang, S. Ren, and J. Sun. 2015. Deep Residual Learning for Image Recognition. (2015). arXiv:1512.03385","journal-title":"J. Sun."},{"key":"e_1_3_2_2_11_1","unstructured":"J. Ho N. Kalchbrenner D. Weissenborn and T. Salimans. 2019. Axial Attentionin Multidimensional Transformers. (2019). arXiv:1912.12180  J. Ho N. Kalchbrenner D. Weissenborn and T. Salimans. 2019. Axial Attentionin Multidimensional Transformers. (2019). arXiv:1912.12180"},{"key":"e_1_3_2_2_12_1","volume-title":"Convolutional Neural Network Architectures for Matching Natural Language Sentences. In NIPS '14. 2042","author":"Hu B.","year":"2050","unstructured":"B. Hu , Z. Lu , H. Li , and Q. Chen . 2014 . Convolutional Neural Network Architectures for Matching Natural Language Sentences. In NIPS '14. 2042 -- 2050 . B. Hu, Z. Lu, H. Li, and Q. Chen. 2014. Convolutional Neural Network Architectures for Matching Natural Language Sentences. In NIPS '14. 2042--2050."},{"volume-title":"CIKM'13","author":"Huang P.","key":"e_1_3_2_2_13_1","unstructured":"P. Huang , X. He , J. Gao , L. Deng , A. Acero , and L. P. Heck . 2013. Learning Deep Structured Semantic Models for Web Search using Clickthrough Data . In CIKM'13 . 2333--2338. P. Huang, X. He, J. Gao, L. Deng, A. Acero, and L. P. Heck. 2013. Learning Deep Structured Semantic Models for Web Search using Clickthrough Data. In CIKM'13. 2333--2338."},{"volume-title":"Semantic Text Matching for Long-Form Documents. In WWW '19","author":"Jiang J.","key":"e_1_3_2_2_14_1","unstructured":"J. Jiang , M. Zhang , C. Li , M. Bendersky , N. Golbandi , and M. Najork . 2019 . Semantic Text Matching for Long-Form Documents. In WWW '19 . 795--806. J. Jiang, M. Zhang, C. Li, M. Bendersky, N. Golbandi, and M. Najork. 2019. Semantic Text Matching for Long-Form Documents. In WWW '19. 795--806."},{"key":"e_1_3_2_2_15_1","unstructured":"J. Johnson M. Douze and H. J\u00e9gou. 2017. Billion-scale similarity search with GPUs. (2017). arXiv:1702.08734  J. Johnson M. Douze and H. J\u00e9gou. 2017. Billion-scale similarity search with GPUs. (2017). arXiv:1702.08734"},{"volume-title":"Reformer: The Efficient Transformer. In ICLR '20","author":"Kitaev N.","key":"e_1_3_2_2_16_1","unstructured":"N. Kitaev , L. Kaiser , and A. Levskaya . 2020 . Reformer: The Efficient Transformer. In ICLR '20 . N. Kitaev, L. Kaiser, and A. Levskaya. 2020. Reformer: The Efficient Transformer. In ICLR '20."},{"key":"e_1_3_2_2_17_1","volume":"201","author":"Li H.","unstructured":"H. Li and J. Xu. 201 4.Semantic Matching in Search. Now Publishers Inc., Hanover,MA, USA. H. Li and J. Xu. 2014.Semantic Matching in Search. Now Publishers Inc., Hanover,MA, USA.","journal-title":"J. Xu."},{"key":"e_1_3_2_2_18_1","volume":"201","author":"Lowe R.","unstructured":"R. Lowe , N. Pow , I. Serban , and J. Pineau. 201 5. The Ubuntu Dialogue Corpus: A Large Dataset for Research in Unstructured Multi-Turn Dialogue Systems. (2015). arXiv:1506.08909 R. Lowe, N. Pow, I. Serban, and J. Pineau. 2015. The Ubuntu Dialogue Corpus: A Large Dataset for Research in Unstructured Multi-Turn Dialogue Systems. (2015). arXiv:1506.08909","journal-title":"J. Pineau."},{"key":"e_1_3_2_2_19_1","volume":"201","author":"Mikolov T.","unstructured":"T. Mikolov , I. Sutskever , K. Chen , G. S. Corrado , and J. Dean. 201 3. Distributed Representations of Words and Phrases and their Compositionality. In NIPS '13. 3111--3119. T. Mikolov, I. Sutskever, K. Chen, G. S. Corrado, and J. Dean. 2013. Distributed Representations of Words and Phrases and their Compositionality. In NIPS '13. 3111--3119.","journal-title":"J. Dean."},{"volume-title":"WWW '17","author":"Mitra B.","key":"e_1_3_2_2_20_1","unstructured":"B. Mitra , F. Diaz , and N. Craswell . 2017. Learning to Match Using Local and Distributed Representations of Text for Web Search . In WWW '17 . 1291--1299. B. Mitra, F. Diaz, and N. Craswell. 2017. Learning to Match Using Local and Distributed Representations of Text for Web Search. In WWW '17. 1291--1299."},{"volume-title":"Overview of the TREC 2008 BlogTrack. In TREC '08.","author":"Ounis I.","key":"e_1_3_2_2_21_1","unstructured":"I. Ounis , C. MacDonald , and I. Soboroff . 2008 . Overview of the TREC 2008 BlogTrack. In TREC '08. I. Ounis, C. MacDonald, and I. Soboroff. 2008. Overview of the TREC 2008 BlogTrack. In TREC '08."},{"volume-title":"AAAI '16","author":"Pang L.","key":"e_1_3_2_2_22_1","unstructured":"L. Pang , Y. Lan , J. Guo , J. Xu , S. Wan , and X. Cheng . 2016. Text Matching as Image Recognition . In AAAI '16 . 2793--2799. L. Pang, Y. Lan, J. Guo, J. Xu, S. Wan, and X. Cheng. 2016. Text Matching as Image Recognition. In AAAI '16. 2793--2799."},{"key":"e_1_3_2_2_23_1","doi-asserted-by":"crossref","unstructured":"M. E. Peters M. Neumann M. Iyyer M. Gardner C. Clark K. Lee and L. Zettle-moyer. 2018. Deep contextualized word representations. (2018). arXiv:1802.05365  M. E. Peters M. Neumann M. Iyyer M. Gardner C. Clark K. Lee and L. Zettle-moyer. 2018. Deep contextualized word representations. (2018). arXiv:1802.05365","DOI":"10.18653\/v1\/N18-1202"},{"key":"e_1_3_2_2_24_1","volume":"201","author":"Qiu J.","unstructured":"J. Qiu , H. Ma , O. Levy , S. W. Yih , S. Wang , and J. Tang. 201 9. Blockwise Self-Attention for Long Document Understanding. (2019). arXiv:1911.02972 J. Qiu, H. Ma, O. Levy, S. W. Yih, S. Wang, and J. Tang. 2019. Blockwise Self-Attention for Long Document Understanding. (2019). arXiv:1911.02972","journal-title":"J. Tang."},{"key":"e_1_3_2_2_25_1","doi-asserted-by":"crossref","unstructured":"D. R. Radev P. Muthukrishnan and V. Qazvinian. 2009. The ACL Anthology Network Corpus. In NLPIR4DL '09. 54--61.  D. R. Radev P. Muthukrishnan and V. Qazvinian. 2009. The ACL Anthology Network Corpus. In NLPIR4DL '09. 54--61.","DOI":"10.3115\/1699750.1699759"},{"key":"e_1_3_2_2_26_1","volume-title":"OpenAI.","author":"Radford A.","year":"2018","unstructured":"A. Radford . 2018. Improving Language Understanding by Generative Pre-Training. Preprint , OpenAI. ( 2018 ). A. Radford. 2018. Improving Language Understanding by Generative Pre-Training. Preprint, OpenAI. (2018)."},{"key":"e_1_3_2_2_27_1","unstructured":"J. W. Rae A. Potapenko S. M. Jayakumar and T. P. Lillicrap. 2019. Compressive Transformers for Long-Range Sequence Modelling. (2019). arXiv:1911.05507  J. W. Rae A. Potapenko S. M. Jayakumar and T. P. Lillicrap. 2019. Compressive Transformers for Long-Range Sequence Modelling. (2019). arXiv:1911.05507"},{"key":"e_1_3_2_2_28_1","unstructured":"A. Roy M. T. Saffar D. Grangier and A. Vaswani. 2020. Efficient Content-Based Sparse Attention with Routing Transformers. (2020). arXiv:2003.05997  A. Roy M. T. Saffar D. Grangier and A. Vaswani. 2020. Efficient Content-Based Sparse Attention with Routing Transformers. (2020). arXiv:2003.05997"},{"key":"e_1_3_2_2_29_1","doi-asserted-by":"crossref","unstructured":"S. Sukhbaatar E. Grave P. Bojanowski and A. Joulin. 2019. Adaptive Attention Span in Transformers. (2019). arXiv:1905.07799  S. Sukhbaatar E. Grave P. Bojanowski and A. Joulin. 2019. Adaptive Attention Span in Transformers. (2019). arXiv:1905.07799","DOI":"10.18653\/v1\/P19-1032"},{"volume-title":"NIPS '17","author":"Vaswani A.","key":"e_1_3_2_2_30_1","unstructured":"A. Vaswani , N. Shazeer , N. Parmar , J. Uszkoreit , L. Jones , A. N. Gomez , ?. Kaiser, and I. Polosukhin . 2017. Attention is All You Need . In NIPS '17 . A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, ?. Kaiser, and I. Polosukhin. 2017. Attention is All You Need. In NIPS '17."},{"volume-title":"Multiscale Quantization for Fast Similarity Search. In NIPS '17","author":"Wu X.","key":"e_1_3_2_2_31_1","unstructured":"X. Wu , R. Guo , A. Suresh , S. Kumar , D. Holtmann-Rice , D. Simcha , and F. Yu . 2017 . Multiscale Quantization for Fast Similarity Search. In NIPS '17 . 5745--5755. X. Wu, R. Guo, A. Suresh, S. Kumar, D. Holtmann-Rice, D. Simcha, and F. Yu. 2017. Multiscale Quantization for Fast Similarity Search. In NIPS '17. 5745--5755."},{"volume-title":"Sequential Matching Network: A New Architecture for Multi-turn Response Selection in Retrieval-Based Chatbots. In ACL '17","author":"Wu Y.","key":"e_1_3_2_2_32_1","unstructured":"Y. Wu , W. Wu , C. Xing , M. Zhou , and Z. Li . 2017 . Sequential Matching Network: A New Architecture for Multi-turn Response Selection in Retrieval-Based Chatbots. In ACL '17 . 163--197. Y. Wu, W. Wu, C. Xing, M. Zhou, and Z. Li. 2017. Sequential Matching Network: A New Architecture for Multi-turn Response Selection in Retrieval-Based Chatbots. In ACL '17. 163--197."},{"volume-title":"End-to-End Neural Ad-hoc Ranking with Kernel Pooling. In SIGIR '17","author":"Xiong C.","key":"e_1_3_2_2_33_1","unstructured":"C. Xiong , Z. Dai , J. Callan , Z. Liu , and R. Power . 2017 . End-to-End Neural Ad-hoc Ranking with Kernel Pooling. In SIGIR '17 . 55--64. C. Xiong, Z. Dai, J. Callan, Z. Liu, and R. Power. 2017. End-to-End Neural Ad-hoc Ranking with Kernel Pooling. In SIGIR '17. 55--64."},{"volume-title":"InCIKM '16","author":"Yang L.","key":"e_1_3_2_2_34_1","unstructured":"L. Yang , Q. Ai , J. Guo , and W. B. Croft . 2016. aNMM: Ranking Short Answer Texts with Attention-Based Neural Matching Model . InCIKM '16 . 287--296. L. Yang, Q. Ai, J. Guo, and W. B. Croft. 2016. aNMM: Ranking Short Answer Texts with Attention-Based Neural Matching Model. InCIKM '16. 287--296."},{"key":"e_1_3_2_2_35_1","volume-title":"WikiQA: A Challenge Dataset for Open-Domain Question Answering. In EMNLP '15. 2013","author":"Yang Y.","year":"2018","unstructured":"Y. Yang , W. Yih , and C. Meek . 2015 . WikiQA: A Challenge Dataset for Open-Domain Question Answering. In EMNLP '15. 2013 -- 2018 . Y. Yang, W. Yih, and C. Meek. 2015. WikiQA: A Challenge Dataset for Open-Domain Question Answering. In EMNLP '15. 2013--2018."},{"key":"e_1_3_2_2_36_1","unstructured":"Z. Yang Z. Dai Y. Yang J. G. Carbonell R. Salakhutdinov and Q. V. Le. 2019.XLNet: Generalized Autoregressive Pretraining for Language Understanding. (2019). arXiv:1906.08237  Z. Yang Z. Dai Y. Yang J. G. Carbonell R. Salakhutdinov and Q. V. Le. 2019.XLNet: Generalized Autoregressive Pretraining for Language Understanding. (2019). arXiv:1906.08237"},{"volume-title":"Hierarchical Attention Networks for Document Classification. In NAACL '16","author":"Yang Z.","key":"e_1_3_2_2_37_1","unstructured":"Z. Yang , D. Yang , C. Dyer , X. He , A. Smola , and E. Hovy . 2016 . Hierarchical Attention Networks for Document Classification. In NAACL '16 . 1480--1489. Z. Yang, D. Yang, C. Dyer, X. He, A. Smola, and E. Hovy. 2016. Hierarchical Attention Networks for Document Classification. In NAACL '16. 1480--1489."},{"volume-title":"Convolutional Neural Network for Paraphrase Identification. In NAACL '15","author":"Yin W.","key":"e_1_3_2_2_38_1","unstructured":"W. Yin and H. Sch\u00fctze . 2015 . Convolutional Neural Network for Paraphrase Identification. In NAACL '15 . 901--911. W. Yin and H. Sch\u00fctze. 2015. Convolutional Neural Network for Paraphrase Identification. In NAACL '15. 901--911."},{"volume-title":"Modelling Domain Relationships for Transfer Learning on Retrieval-based Question Answering Systems in E-commerce. In WSDM '18","author":"Yu J.","key":"e_1_3_2_2_39_1","unstructured":"J. Yu , M. Qiu , J. Jiang , J. Huang , S. Song , W. Chu , and H. Chen . 2018 . Modelling Domain Relationships for Transfer Learning on Retrieval-based Question Answering Systems in E-commerce. In WSDM '18 . 682--690. J. Yu, M. Qiu, J. Jiang, J. Huang, S. Song, W. Chu, and H. Chen. 2018. Modelling Domain Relationships for Transfer Learning on Retrieval-based Question Answering Systems in E-commerce. In WSDM '18. 682--690."},{"key":"e_1_3_2_2_40_1","volume-title":"HIBERT: Document Level Pre-training of Hierarchical Bidirectional Transformers for Document Summarization.","author":"Zhang X.","year":"2019","unstructured":"X. Zhang , F. Wei , and M. Zhou . 2019 . HIBERT: Document Level Pre-training of Hierarchical Bidirectional Transformers for Document Summarization. (2019).arXiv:1905.06566 X. Zhang, F. Wei, and M. Zhou. 2019. HIBERT: Document Level Pre-training of Hierarchical Bidirectional Transformers for Document Summarization. (2019).arXiv:1905.06566"}],"event":{"name":"CIKM '20: The 29th ACM International Conference on Information and Knowledge Management","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web","SIGIR ACM Special Interest Group on Information Retrieval"],"location":"Virtual Event Ireland","acronym":"CIKM '20"},"container-title":["Proceedings of the 29th ACM International Conference on Information &amp; Knowledge Management"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3340531.3411908","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3340531.3411908","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T22:01:22Z","timestamp":1750197682000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3340531.3411908"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,10,19]]},"references-count":40,"alternative-id":["10.1145\/3340531.3411908","10.1145\/3340531"],"URL":"https:\/\/doi.org\/10.1145\/3340531.3411908","relation":{},"subject":[],"published":{"date-parts":[[2020,10,19]]},"assertion":[{"value":"2020-10-19","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}