{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,26]],"date-time":"2025-06-26T09:32:23Z","timestamp":1750930343713,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":52,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,4,25]],"date-time":"2022-04-25T00:00:00Z","timestamp":1650844800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"Beijing Outstanding Young Scientist Program","award":["NO. BJJWZYJH012019100020098"],"award-info":[{"award-number":["NO. BJJWZYJH012019100020098"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["No. 61872370"],"award-info":[{"award-number":["No. 61872370"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,4,25]]},"DOI":"10.1145\/3485447.3511962","type":"proceedings-article","created":{"date-parts":[[2022,4,25]],"date-time":"2022-04-25T05:11:23Z","timestamp":1650863483000},"page":"339-347","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Socialformer: Social Network Inspired Long Document Modeling for Document Ranking"],"prefix":"10.1145","author":[{"given":"Yujia","family":"Zhou","sequence":"first","affiliation":[{"name":"School of Information, Renmin University of China, China"}]},{"given":"Zhicheng","family":"Dou","sequence":"additional","affiliation":[{"name":"Gaoling School of Artificial Intelligence, Renmin University of China, China and Beijing Key Laboratory of Big Data Management and Analysis Methods, China"}]},{"given":"Huaying","family":"Yuan","sequence":"additional","affiliation":[{"name":"College of Computer Science, Nankai University, China"}]},{"given":"Zhengyi","family":"Ma","sequence":"additional","affiliation":[{"name":"School of Information, Renmin University of China, China"}]}],"member":"320","published-online":{"date-parts":[[2022,4,25]]},"reference":[{"volume-title":"ECIR(Lecture Notes in Computer Science), Vol.\u00a010772","author":"Ai Qingyao","key":"e_1_3_2_1_1_1","unstructured":"Qingyao Ai, Brendan O\u2019Connor, and W.\u00a0Bruce Croft. 2018. A Neural Passage Model for Ad-hoc Document Retrieval. In ECIR(Lecture Notes in Computer Science), Vol.\u00a010772. Springer, 537\u2013543."},{"key":"e_1_3_2_1_2_1","volume-title":"ETC: Encoding Long and Structured Inputs in Transformers. In EMNLP (1)","author":"Ainslie Joshua","year":"2020","unstructured":"Joshua Ainslie, Santiago Onta\u00f1\u00f3n, Chris Alberti, Vaclav Cvicek, Zachary Fisher, Philip Pham, Anirudh Ravula, Sumit Sanghai, Qifan Wang, and Li Yang. 2020. ETC: Encoding Long and Structured Inputs in Transformers. In EMNLP (1). Association for Computational Linguistics, 268\u2013284."},{"key":"e_1_3_2_1_3_1","volume-title":"Longformer: The Long-Document Transformer. CoRR abs\/2004.05150(2020).","author":"Beltagy Iz","year":"2020","unstructured":"Iz Beltagy, Matthew\u00a0E. Peters, and Arman Cohan. 2020. Longformer: The Long-Document Transformer. CoRR abs\/2004.05150(2020)."},{"volume-title":"Passage-Level Evidence in Document Retrieval","author":"Callan P.","key":"e_1_3_2_1_4_1","unstructured":"James\u00a0P. Callan. 1994. Passage-Level Evidence in Document Retrieval. In SIGIR. ACM\/Springer, 302\u2013310."},{"key":"e_1_3_2_1_5_1","volume-title":"Overview of the TREC 2019 deep learning track. CoRR abs\/2003","author":"Craswell Nick","year":"2020","unstructured":"Nick Craswell, Bhaskar Mitra, Emine Yilmaz, Daniel Campos, and Ellen\u00a0M. Voorhees. 2020. Overview of the TREC 2019 deep learning track. CoRR abs\/2003.07820(2020). arXiv:2003.07820https:\/\/arxiv.org\/abs\/2003.07820"},{"key":"e_1_3_2_1_6_1","volume-title":"Overview of the TREC 2019 deep learning track.","author":"Craswell Nick","year":"2020","unstructured":"Nick Craswell, Bhaskar Mitra, Emine Yilmaz, Daniel Campos, and Ellen\u00a0M. Voorhees. 2020. Overview of the TREC 2019 deep learning track. (2020)."},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"crossref","unstructured":"Zhuyun Dai and Jamie Callan. 2019. Deeper Text Understanding for IR with Contextual Neural Language Modeling. In SIGIR. ACM 985\u2013988.","DOI":"10.1145\/3331184.3331303"},{"volume-title":"ACL (1)","author":"Dai Zihang","key":"e_1_3_2_1_8_1","unstructured":"Zihang Dai, Zhilin Yang, Yiming Yang, Jaime\u00a0G. Carbonell, Quoc\u00a0Viet Le, and Ruslan Salakhutdinov. 2019. Transformer-XL: Attentive Language Models beyond a Fixed-Length Context. In ACL (1). Association for Computational Linguistics, 2978\u20132988."},{"key":"e_1_3_2_1_9_1","volume-title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL-HLT (1)","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL-HLT (1). Association for Computational Linguistics, 4171\u20134186."},{"key":"e_1_3_2_1_10_1","first-page":"43","article-title":"Networks, crowds, and markets: Reasoning about a highly connected world","volume":"9","author":"Easley David","year":"2012","unstructured":"David Easley, Jon Kleinberg, 2012. Networks, crowds, and markets: Reasoning about a highly connected world. Significance 9, 1 (2012), 43\u201344.","journal-title":"Significance"},{"key":"e_1_3_2_1_11_1","first-page":"159","article-title":"The development of social network analysis","volume":"1","author":"Freeman Linton","year":"2004","unstructured":"Linton Freeman. 2004. The development of social network analysis. A Study in the Sociology of Science 1, 687 (2004), 159\u2013167.","journal-title":"A Study in the Sociology of Science"},{"key":"e_1_3_2_1_12_1","volume-title":"GMAT: Global Memory Augmentation for Transformers. CoRR abs\/2006.03274(2020).","author":"Gupta Ankit","year":"2020","unstructured":"Ankit Gupta and Jonathan Berant. 2020. GMAT: Global Memory Augmentation for Transformers. CoRR abs\/2006.03274(2020)."},{"key":"e_1_3_2_1_13_1","volume-title":"Strong ties or weak ties? Neighbourhood networks in a new perspective. Scandinavian Housing and planning research 13, 1","author":"Henning Cecilia","year":"1996","unstructured":"Cecilia Henning and Mats Lieberg. 1996. Strong ties or weak ties? Neighbourhood networks in a new perspective. Scandinavian Housing and planning research 13, 1 (1996), 3\u201326."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"crossref","unstructured":"Sebastian Hofst\u00e4tter Bhaskar Mitra Hamed Zamani Nick Craswell and Allan Hanbury. 2021. Intra-Document Cascading: Learning to Select Passages for Neural Document Ranking. In SIGIR. ACM 1349\u20131358.","DOI":"10.1145\/3404835.3462889"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/3397271.3401224"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1093\/geronb\/59.6.P278"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.brq.2018.10.004"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.412"},{"key":"e_1_3_2_1_20_1","volume-title":"Reformer: The Efficient Transformer. In ICLR. OpenReview.net.","author":"Kitaev Nikita","year":"2020","unstructured":"Nikita Kitaev, Lukasz Kaiser, and Anselm Levskaya. 2020. Reformer: The Efficient Transformer. In ICLR. OpenReview.net."},{"key":"e_1_3_2_1_21_1","volume-title":"Navigation in a small world. Nature 406, 6798","author":"Kleinberg M","year":"2000","unstructured":"Jon\u00a0M Kleinberg. 2000. Navigation in a small world. Nature 406, 6798 (2000), 845\u2013845."},{"key":"e_1_3_2_1_22_1","volume-title":"PARADE: Passage Representation Aggregation for Document Reranking. CoRR abs\/2008.09093(2020).","author":"Li Canjia","year":"2020","unstructured":"Canjia Li, Andrew Yates, Sean MacAvaney, Ben He, and Yingfei Sun. 2020. PARADE: Passage Representation Aggregation for Document Reranking. CoRR abs\/2008.09093(2020)."},{"key":"e_1_3_2_1_23_1","unstructured":"Ilya Loshchilov and Frank Hutter. 2019. Decoupled Weight Decay Regularization. In ICLR (Poster). OpenReview.net."},{"key":"e_1_3_2_1_24_1","unstructured":"Zhengyi Ma Zhicheng Dou Wei Xu Xinyu Zhang Hao Jiang Zhao Cao and Ji-Rong Wen. 2021. Pre-training for Ad-hoc Retrieval: Hyperlink is Also You Need. In CIKM. ACM 1212\u20131221."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/3331184.3331317"},{"key":"e_1_3_2_1_26_1","volume-title":"MS MARCO: A Human Generated MAchine Reading COmprehension Dataset. In NIPS","author":"Nguyen Tri","year":"2016","unstructured":"Tri Nguyen, Mir Rosenberg, Xia Song, 2016. MS MARCO: A Human Generated MAchine Reading COmprehension Dataset. In NIPS 2016."},{"key":"e_1_3_2_1_27_1","unstructured":"Rodrigo Nogueira and Kyunghyun Cho. 2019. Passage Re-ranking with BERT. CoRR abs\/1901.04085(2019)."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.232"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"crossref","unstructured":"Stephen Robertson Hugo Zaragoza 2009. The probabilistic relevance framework: BM25 and beyond. Foundations and Trends\u00ae in Information Retrieval 3 4(2009) 333\u2013389.","DOI":"10.1561\/1500000019"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00353"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"crossref","unstructured":"Koustav Rudra and Avishek Anand. 2020. Distant Supervision in BERT-based Adhoc Document Retrieval. In CIKM. ACM 2197\u20132200.","DOI":"10.1145\/3340531.3412124"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1177\/0038038588022001007"},{"key":"e_1_3_2_1_33_1","volume-title":"NIPS","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan\u00a0N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is All you Need. In NIPS 2017. 5998\u20136008."},{"key":"e_1_3_2_1_34_1","volume-title":"Collective dynamics of \u2018small-world\u2019networks. nature 393, 6684","author":"Watts J","year":"1998","unstructured":"Duncan\u00a0J Watts and Steven\u00a0H Strogatz. 1998. Collective dynamics of \u2018small-world\u2019networks. nature 393, 6684 (1998), 440\u2013442."},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.17730\/humo.57.4.pw61g8j535740428"},{"volume-title":"ACL\/IJCNLP (2)","author":"Wu Chuhan","key":"e_1_3_2_1_36_1","unstructured":"Chuhan Wu, Fangzhao Wu, Tao Qi, and Yongfeng Huang. 2021. Hi-Transformer: Hierarchical Interactive Transformer for Efficient and Effective Long Document Modeling. In ACL\/IJCNLP (2). Association for Computational Linguistics, 848\u2013853."},{"key":"e_1_3_2_1_37_1","unstructured":"Zhijing Wu Jiaxin Mao Yiqun Liu Jingtao Zhan Yukun Zheng Min Zhang and Shaoping Ma. 2020. Leveraging Passage-level Cumulative Gain for Document Ranking. In WWW. ACM \/ IW3C2 2421\u20132431."},{"key":"e_1_3_2_1_38_1","unstructured":"Lee Xiong Chenyan Xiong Ye Li Kwok-Fung Tang Jialin Liu 2020. Approximate Nearest Neighbor Negative Contrastive Learning for Dense Text Retrieval. CoRR abs\/2007.00808(2020)."},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"crossref","unstructured":"Liu Yang Mingyang Zhang Cheng Li Michael Bendersky and Marc Najork. 2020. Beyond 512 Tokens: Siamese Multi-depth Transformer-based Hierarchical Encoder for Long-Form Document Matching. In CIKM. ACM 1725\u20131734.","DOI":"10.1145\/3340531.3411908"},{"volume-title":"Hierarchical Attention Networks for Document Classification","author":"Yang Zichao","key":"e_1_3_2_1_40_1","unstructured":"Zichao Yang, Diyi Yang, Chris Dyer, Xiaodong He, Alexander\u00a0J. Smola, and Eduard\u00a0H. Hovy. 2016. Hierarchical Attention Networks for Document Classification. In HLT-NAACL. The Association for Computational Linguistics, 1480\u20131489."},{"volume-title":"EMNLP\/IJCNLP (3)","author":"Yilmaz Zeynep\u00a0Akkalyoncu","key":"e_1_3_2_1_41_1","unstructured":"Zeynep\u00a0Akkalyoncu Yilmaz, Shengjin Wang, Wei Yang, Haotian Zhang, and Jimmy Lin. 2019. Applying BERT to Document Retrieval with Birch. In EMNLP\/IJCNLP (3). Association for Computational Linguistics, 19\u201324."},{"key":"e_1_3_2_1_42_1","unstructured":"Chengxuan Ying Tianle Cai Shengjie Luo Shuxin Zheng Guolin Ke Di He Yanming Shen and Tie-Yan Liu. 2021. Do Transformers Really Perform Bad for Graph Representation?CoRR abs\/2106.05234(2021)."},{"key":"e_1_3_2_1_43_1","volume-title":"PGT: Pseudo Relevance Feedback Using a Graph-Based Transformer. In ECIR (2)(Lecture Notes in Computer Science), Vol.\u00a012657","author":"Yu HongChien","year":"2021","unstructured":"HongChien Yu, Zhuyun Dai, and Jamie Callan. 2021. PGT: Pseudo Relevance Feedback Using a Graph-Based Transformer. In ECIR (2)(Lecture Notes in Computer Science), Vol.\u00a012657. Springer, 440\u2013447."},{"key":"e_1_3_2_1_44_1","volume-title":"Big Bird: Transformers for Longer Sequences. In NeurIPS.","author":"Zaheer Manzil","year":"2020","unstructured":"Manzil Zaheer, Guru Guruganesh, Kumar\u00a0Avinava Dubey, Joshua Ainslie, Chris Alberti, Santiago Onta\u00f1\u00f3n, Philip Pham, Anirudh Ravula, Qifan Wang, Li Yang, and Amr Ahmed. 2020. Big Bird: Transformers for Longer Sequences. In NeurIPS."},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1145\/3130348.3130377"},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1145\/3397271.3401325"},{"key":"e_1_3_2_1_47_1","volume-title":"Poolingformer: Long Document Modeling with Pooling Attention. In ICML(Proceedings of Machine Learning Research), Vol.\u00a0139. PMLR, 12437\u201312446.","author":"Zhang Hang","year":"2021","unstructured":"Hang Zhang, Yeyun Gong, Yelong Shen, Weisheng Li, Jiancheng Lv, Nan Duan, and Weizhu Chen. 2021. Poolingformer: Long Document Modeling with Pooling Attention. In ICML(Proceedings of Machine Learning Research), Vol.\u00a0139. PMLR, 12437\u201312446."},{"key":"e_1_3_2_1_48_1","volume-title":"HIBERT: Document Level Pre-training of Hierarchical Bidirectional Transformers for Document Summarization. In ACL (1)","author":"Zhang Xingxing","year":"2019","unstructured":"Xingxing Zhang, Furu Wei, and Ming Zhou. 2019. HIBERT: Document Level Pre-training of Hierarchical Bidirectional Transformers for Document Summarization. In ACL (1). Association for Computational Linguistics, 5059\u20135069."},{"key":"e_1_3_2_1_49_1","unstructured":"Chen Zhao Chenyan Xiong Corby Rosset Xia Song Paul\u00a0N. Bennett and Saurabh Tiwary. 2020. Transformer-XH: Multi-Evidence Reasoning with eXtra Hop Attention. In ICLR. OpenReview.net."},{"key":"e_1_3_2_1_50_1","unstructured":"Guangxiang Zhao Junyang Lin Zhiyuan Zhang Xuancheng Ren Qi Su and Xu Sun. 2019. Explicit Sparse Transformer: Concentrated Attention Through Explicit Selection. CoRR abs\/1912.11637(2019)."},{"key":"e_1_3_2_1_51_1","volume-title":"Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting","author":"Zhou Haoyi","year":"2021","unstructured":"Haoyi Zhou, Shanghang Zhang, Jieqi Peng, Shuai Zhang, Jianxin Li, Hui Xiong, and Wancai Zhang. 2021. Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting. In AAAI. AAAI Press, 11106\u201311115."},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1145\/3404835.3462918"}],"event":{"name":"WWW '22: The ACM Web Conference 2022","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"],"location":"Virtual Event, Lyon France","acronym":"WWW '22"},"container-title":["Proceedings of the ACM Web Conference 2022"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3485447.3511962","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3485447.3511962","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T20:12:09Z","timestamp":1750191129000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3485447.3511962"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4,25]]},"references-count":52,"alternative-id":["10.1145\/3485447.3511962","10.1145\/3485447"],"URL":"https:\/\/doi.org\/10.1145\/3485447.3511962","relation":{},"subject":[],"published":{"date-parts":[[2022,4,25]]},"assertion":[{"value":"2022-04-25","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}