{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,15]],"date-time":"2025-11-15T07:32:34Z","timestamp":1763191954175,"version":"3.45.0"},"reference-count":55,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,6,30]],"date-time":"2025-06-30T00:00:00Z","timestamp":1751241600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,6,30]],"date-time":"2025-06-30T00:00:00Z","timestamp":1751241600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,6,30]]},"DOI":"10.1109\/ijcnn64981.2025.11228570","type":"proceedings-article","created":{"date-parts":[[2025,11,14]],"date-time":"2025-11-14T18:46:15Z","timestamp":1763145975000},"page":"1-8","source":"Crossref","is-referenced-by-count":0,"title":["CoT-Planner: Chain-of-Thoughts as the Content Planner for Few-shot Table-to-Text Generation Reduces the Hallucinations from LLMs"],"prefix":"10.1109","author":[{"given":"Yupian","family":"Lin","sequence":"first","affiliation":[{"name":"East China University of Science and Technology,School of Information Science and Engineering,Shanghai,China"}]},{"given":"Yuang","family":"Bian","sequence":"additional","affiliation":[{"name":"East China University of Science and Technology,School of Information Science and Engineering,Shanghai,China"}]},{"given":"Guangya","family":"Yu","sequence":"additional","affiliation":[{"name":"East China University of Science and Technology,School of Information Science and Engineering,Shanghai,China"}]},{"given":"Dongge","family":"Xue","sequence":"additional","affiliation":[{"name":"East China University of Science and Technology,School of Information Science and Engineering,Shanghai,China"}]},{"given":"Wanpeng","family":"Lu","sequence":"additional","affiliation":[{"name":"East China University of Science and Technology,School of Information Science and Engineering,Shanghai,China"}]},{"given":"Jingping","family":"Liu","sequence":"additional","affiliation":[{"name":"East China University of Science and Technology,School of Information Science and Engineering,Shanghai,China"}]},{"given":"Tong","family":"Ruan","sequence":"additional","affiliation":[{"name":"East China University of Science and Technology,School of Information Science and Engineering,Shanghai,China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.3115\/1687878.1687893"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D17-1239"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.202"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.708"},{"article-title":"Open question answering over tables and text","year":"2020","author":"Chen","key":"ref5"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.naacl-main.68"},{"article-title":"Language models are unsupervised multitask learners","year":"2019","author":"Radford","key":"ref7"},{"issue":"140","key":"ref8","first-page":"1","article-title":"Exploring the limits of transfer learning with a unified text-to-text transformer","volume":"21","author":"Raffel","year":"2020","journal-title":"Journal of machine learning research"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.703"},{"article-title":"Few-shot table-to-text generation with prefix-controlled generator","year":"2022","author":"Luo","key":"ref10"},{"article-title":"A survey of large language models","year":"2023","author":"Zhao","key":"ref11"},{"article-title":"Language models are few-shot learners","year":"2020","author":"Mann","key":"ref12"},{"issue":"240","key":"ref13","first-page":"1","article-title":"Palm: Scaling language modeling with pathways","volume":"24","author":"Chowdhery","year":"2023","journal-title":"Journal of Machine Learning Research"},{"article-title":"Galactica: A large language model for science","year":"2022","author":"Taylor","key":"ref14"},{"article-title":"Llama: Open and efficient foundation language models","year":"2023","author":"Touvron","key":"ref15"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.emnlp-main.64"},{"article-title":"Gpt-4 technical report","year":"2023","author":"Achiam","key":"ref17"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.acl-long.115"},{"key":"ref19","first-page":"24 824","article-title":"Chain-of-thought prompting elicits reasoning in large language models","volume":"35","author":"Wei","year":"2022","journal-title":"Advances in neural information processing systems"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33016908"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/W19-8645"},{"article-title":"Step-by-step: Separating planning from realization in neural data-to-text generation","year":"2019","author":"Moryossef","key":"ref22"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.findings-emnlp.76"},{"article-title":"Llama 2: Open foundation and fine-tuned chat models","year":"2023","author":"Touvron","key":"ref24"},{"article-title":"Stanford alpaca: An instruction-following llama model","year":"2023","author":"Taori","key":"ref25"},{"key":"ref26","article-title":"Judging llm-as-a-judge with mt-bench and chatbot arena","volume":"36","author":"Zheng","year":"2024","journal-title":"Advances in Neural Information Processing Systems"},{"article-title":"Chatglm: A family of large language models from glm-130b to glm-4 all tools","year":"2024","author":"Glm","key":"ref27"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1197"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.18"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.findings-acl.136"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.coling-main.179"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.findings-emnlp.77"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2022\/580"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.findings-emnlp.347"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-long.894"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.emnlp-main.574"},{"article-title":"Self-consistency improves chain of thought reasoning in language models","year":"2022","author":"Wang","key":"ref37"},{"article-title":"Chain-of-symbol prompting elicits planning in large langauge models","year":"2023","author":"Hu","key":"ref38"},{"key":"ref39","article-title":"Tree of thoughts: Deliberate problem solving with large language models","volume":"36","author":"Yao","year":"2024","journal-title":"Advances in Neural Information Processing Systems"},{"article-title":"Skeleton-of-thought: Large language models can do parallel decoding","year":"2023","author":"Ning","key":"ref40"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i16.29720"},{"article-title":"Boosting logical reasoning in large language models through a new framework: The graph of thought","year":"2023","author":"Lei","key":"ref42"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-eacl.83"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.naacl-long.320"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1145\/3539618.3591708"},{"key":"ref46","article-title":"A survey of chain of thought reasoning: Advances, frontiers and future","volume":"abs\/2309.15402","author":"Chu","year":"2023","journal-title":"CoRR"},{"key":"ref47","first-page":"22 199","article-title":"Large language models are zero-shot reasoners","volume":"35","author":"Kojima","year":"2022","journal-title":"Advances in neural information processing systems"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.3115\/1073083.1073135"},{"first-page":"65","article-title":"Meteor: an automatic metric for mt evaluation with high levels of correlation with human judgments","author":"Banerjee","key":"ref49"},{"key":"ref50","first-page":"74","article-title":"Rouge: A package for automatic evaluation of summaries","author":"Lin","year":"2004"},{"article-title":"Bertscore: Evaluating text generation with bert","year":"2019","author":"Zhang","key":"ref51"},{"article-title":"Bert: Pre-training of deep bidirectional transformers for language understanding","year":"2018","author":"Devlin","key":"ref52"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P17-1099"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.emnlp-main.562"},{"article-title":"Efficient and effective text encoding for chinese llama and alpaca","year":"2023","author":"Cui","key":"ref55"}],"event":{"name":"2025 International Joint Conference on Neural Networks (IJCNN)","start":{"date-parts":[[2025,6,30]]},"location":"Rome, Italy","end":{"date-parts":[[2025,7,5]]}},"container-title":["2025 International Joint Conference on Neural Networks (IJCNN)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11227166\/11227148\/11228570.pdf?arnumber=11228570","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,15]],"date-time":"2025-11-15T07:27:38Z","timestamp":1763191658000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11228570\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,30]]},"references-count":55,"URL":"https:\/\/doi.org\/10.1109\/ijcnn64981.2025.11228570","relation":{},"subject":[],"published":{"date-parts":[[2025,6,30]]}}}