{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,20]],"date-time":"2026-03-20T17:07:02Z","timestamp":1774026422521,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":24,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,12,6]],"date-time":"2024-12-06T00:00:00Z","timestamp":1733443200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"JSPS Research on Academic Transforma tion Areas","award":["JP22H05194"],"award-info":[{"award-number":["JP22H05194"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,12,6]]},"DOI":"10.1145\/3709026.3709050","type":"proceedings-article","created":{"date-parts":[[2025,2,15]],"date-time":"2025-02-15T10:05:41Z","timestamp":1739613941000},"page":"367-373","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Extending Token Computation for LLM Reasoning"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0005-0474-7464","authenticated-orcid":false,"given":"Bingli","family":"Li","sequence":"first","affiliation":[{"name":"Laboratory of Intelligent Systems, The Graduate School of Information Science and Electrical Engineering, Kyushu University, Fukuoka, Fukuoka, Japan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7442-1279","authenticated-orcid":false,"given":"Danilo Vasconcellos","family":"Vargas","sequence":"additional","affiliation":[{"name":"Laboratory of Intelligent Systems, The Graduate School of Information Science and Electrical Engineering, Kyushu University, Fukuoka, Fukuoka, Japan"}]}],"member":"320","published-online":{"date-parts":[[2025,2,15]]},"reference":[{"key":"e_1_3_3_1_2_2","unstructured":"Yossi Adi Einat Kermany Yonatan Belinkov Ofer Lavi and Yoav Goldberg. 2016. Fine-grained analysis of sentence embeddings using auxiliary prediction tasks. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/1608.04207 (2016)."},{"key":"e_1_3_3_1_3_2","unstructured":"Yonatan Bisk Rowan Zellers Ronan Le\u00a0Bras Jianfeng Gao and Yejin Choi. 2019. PIQA: Reasoning about physical commonsense in natural language. (2019). arxiv:https:\/\/arXiv.org\/abs\/1911.11641\u00a0[cs.CL]"},{"key":"e_1_3_3_1_4_2","unstructured":"Tom Brown Benjamin Mann Nick Ryder Melanie Subbiah Jared\u00a0D Kaplan Prafulla Dhariwal Arvind Neelakantan Pranav Shyam Girish Sastry Amanda Askell et\u00a0al. 2020. Language models are few-shot learners. Advances in Neural Information Processing Systems 33 (2020) 1877\u20131901."},{"key":"e_1_3_3_1_5_2","unstructured":"Hyung\u00a0Won Chung Le Hou Shayne Longpre Barret Zoph Yi Tay William Fedus Yunxuan Li Xuezhi Wang Mostafa Dehghani Siddhartha Brahma et\u00a0al. 2022. Scaling instruction-finetuned language models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2210.11416 (2022)."},{"key":"e_1_3_3_1_6_2","doi-asserted-by":"crossref","unstructured":"Kevin Clark Urvashi Khandelwal Omer Levy and Christopher\u00a0D Manning. 2019. What does BERT look at? An analysis of BERT\u2019s attention. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/1906.04341 (2019).","DOI":"10.18653\/v1\/W19-4828"},{"key":"e_1_3_3_1_7_2","doi-asserted-by":"crossref","unstructured":"Mark Davies. 2010. The Corpus of Contemporary American English as the first reliable monitor corpus of English. Literary and Linguistic Computing 25 4 (2010) 447\u2013464.","DOI":"10.1093\/llc\/fqq018"},{"key":"e_1_3_3_1_8_2","unstructured":"Dan Hendrycks Collin Burns Steven Basart Andy Zou Mantas Mazeika Dawn Song and Jacob Steinhardt. 2020. Measuring massive multitask language understanding. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2009.03300 (2020)."},{"key":"e_1_3_3_1_9_2","unstructured":"Edward\u00a0J Hu Yelong Shen Phillip Wallis Zeyuan Allen-Zhu Yuanzhi Li Shean Wang Lu Wang and Weizhu Chen. 2021. Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2106.09685 (2021)."},{"key":"e_1_3_3_1_10_2","doi-asserted-by":"crossref","unstructured":"Daniel\u00a0Martin Katz Michael\u00a0James Bommarito Shang Gao and Pablo Arredondo. 2024. Gpt-4 passes the bar exam. Philosophical Transactions of the Royal Society A 382 2270 (2024) 20230254.","DOI":"10.1098\/rsta.2023.0254"},{"key":"e_1_3_3_1_11_2","unstructured":"Patrick Lewis Ethan Perez Aleksandra Piktus Fabio Petroni Vladimir Karpukhin Naman Goyal Heinrich K\u00fcttler Mike Lewis Wen-tau Yih Tim Rockt\u00e4schel et\u00a0al. 2020. Retrieval-augmented generation for knowledge-intensive NLP tasks. Advances in Neural Information Processing Systems 33 (2020) 9459\u20139474."},{"key":"e_1_3_3_1_12_2","unstructured":"Aitor Lewkowycz Anders Andreassen David Dohan Ethan Dyer Henryk Michalewski Vinay Ramasesh Ambrose Slone Cem Anil Imanol Schlag Theo Gutman-Solo et\u00a0al. 2022. Solving quantitative reasoning problems with language models. Advances in Neural Information Processing Systems 35 (2022) 3843\u20133857."},{"key":"e_1_3_3_1_13_2","unstructured":"Yiming Li and Zhao Zhang. 2024. The first place solution of WSDM Cup 2024: Leveraging large language models for conversational multi-doc QA. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2402.18385 (2024)."},{"key":"e_1_3_3_1_14_2","doi-asserted-by":"crossref","unstructured":"Bingli Liao and Danilo\u00a0Vasconcellos Vargas. 2023. Towards immersive computational storytelling: Card-framework for enhanced persona-driven dialogues. Authorea Preprints (2023).","DOI":"10.36227\/techrxiv.24605667.v1"},{"key":"e_1_3_3_1_15_2","unstructured":"R OpenAI. 2023. Gpt-4 technical report. View in Article 2 5 (2023)."},{"key":"e_1_3_3_1_16_2","unstructured":"Stanislas Polu and Ilya Sutskever. 2020. Generative language modeling for automated theorem proving. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2009.03393 (2020)."},{"key":"e_1_3_3_1_17_2","unstructured":"Alec Radford Karthik Narasimhan Tim Salimans Ilya Sutskever et\u00a0al. 2018. Improving language understanding by generative pre-training. (2018)."},{"key":"e_1_3_3_1_18_2","unstructured":"Baptiste Roziere Jonas Gehring Fabian Gloeckle Sten Sootla Itai Gat Xiaoqing\u00a0Ellen Tan Yossi Adi Jingyu Liu Tal Remez J\u00e9r\u00e9my Rapin et\u00a0al. 2023. Code llama: Open foundation models for code. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2308.12950 (2023)."},{"key":"e_1_3_3_1_19_2","unstructured":"Maarten Sap Hannah Rashkin Derek Chen Ronan LeBras and Yejin Choi. 2019. SocialIQA: Commonsense reasoning about social interactions. (2019). arxiv:https:\/\/arXiv.org\/abs\/1904.09728\u00a0[cs.CL]"},{"key":"e_1_3_3_1_20_2","unstructured":"Mingjie Sun Xinlei Chen J\u00a0Zico Kolter and Zhuang Liu. 2024. Massive activations in large language models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2402.17762 (2024)."},{"key":"e_1_3_3_1_21_2","unstructured":"Hugo Touvron Louis Martin Kevin Stone Peter Albert Amjad Almahairi Yasmine Babaei Nikolay Bashlykov Soumya Batra Prajjwal Bhargava Shruti Bhosale et\u00a0al. 2023. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2307.09288 (2023)."},{"key":"e_1_3_3_1_22_2","unstructured":"Jason Wei Xuezhi Wang Dale Schuurmans Maarten Bosma Fei Xia Ed Chi Quoc\u00a0V Le Denny Zhou et\u00a0al. 2022. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems 35 (2022) 24824\u201324837."},{"key":"e_1_3_3_1_23_2","unstructured":"Guangxuan Xiao Yuandong Tian Beidi Chen Song Han and Mike Lewis. 2023. Efficient streaming language models with attention sinks. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2309.17453 (2023)."},{"key":"e_1_3_3_1_24_2","unstructured":"Catherine Yeh Yida Chen Aoyu Wu Cynthia Chen Fernanda Vi\u00e9gas and Martin Wattenberg. 2023. Attentionviz: A global view of transformer attention. IEEE Transactions on Visualization and Computer Graphics (2023)."},{"key":"e_1_3_3_1_25_2","unstructured":"Eric Zelikman Yuhuai Wu Jesse Mu and Noah Goodman. 2022. Star: Bootstrapping reasoning with reasoning. Advances in Neural Information Processing Systems 35 (2022) 15476\u201315488."}],"event":{"name":"CSAI 2024: 2024 8th International Conference on Computer Science and Artificial Intelligence (CSAI)","location":"Beijing China","acronym":"CSAI 2024"},"container-title":["Proceedings of the 2024 8th International Conference on Computer Science and Artificial Intelligence"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3709026.3709050","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3709026.3709050","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:17:31Z","timestamp":1750295851000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3709026.3709050"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,6]]},"references-count":24,"alternative-id":["10.1145\/3709026.3709050","10.1145\/3709026"],"URL":"https:\/\/doi.org\/10.1145\/3709026.3709050","relation":{},"subject":[],"published":{"date-parts":[[2024,12,6]]},"assertion":[{"value":"2025-02-15","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}