{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,13]],"date-time":"2026-03-13T07:08:54Z","timestamp":1773385734374,"version":"3.50.1"},"reference-count":68,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2026,3,13]],"date-time":"2026-03-13T00:00:00Z","timestamp":1773360000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2026,3,13]],"date-time":"2026-03-13T00:00:00Z","timestamp":1773360000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100001230","name":"Macquarie University","doi-asserted-by":"crossref","id":[{"id":"10.13039\/501100001230","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Cogn Comput"],"published-print":{"date-parts":[[2026,12]]},"DOI":"10.1007\/s12559-026-10568-9","type":"journal-article","created":{"date-parts":[[2026,3,13]],"date-time":"2026-03-13T02:10:12Z","timestamp":1773367812000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["LLM Alignment should go beyond Harmlessness\u2013Helpfulness and incorporate Human Agency"],"prefix":"10.1007","volume":"18","author":[{"given":"Usman","family":"Naseem","sequence":"first","affiliation":[]},{"given":"Tanmoy","family":"Chakraborty","sequence":"additional","affiliation":[]},{"given":"Kai-Wei","family":"Chang","sequence":"additional","affiliation":[]},{"given":"Mark","family":"Dras","sequence":"additional","affiliation":[]},{"given":"Preslav","family":"Nakov","sequence":"additional","affiliation":[]},{"given":"Nanyun","family":"Peng","sequence":"additional","affiliation":[]},{"given":"Soujanya","family":"Poria","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,3,13]]},"reference":[{"key":"10568_CR1","unstructured":"Brown TB, Mann B, Ryder N, Subbiah M, Kaplan J, Dhariwal P, Neelakantan A, Shyam P, Sastry G, Askell A, Agarwal S, Herbert-Voss A, Krueger G, Henighan T, Child R, Ramesh A, Ziegler DM, Wu J, Winter C, Hesse C, Chen M, Sigler E, Litwin M, Gray S, Chess B, Clark J, Berner C, McCandlish S, Radford A, Sutskever I, Amodei D. Language models are few-shot learners. Adv Neural Inf Process Syst. 2020;33:1877\u2013901. arXiv:2005.14165 [cs.CL]."},{"key":"10568_CR2","unstructured":"Achiam J, Adler S, Agarwal S, Ahmad L, Akkaya I, Aleman FL, Almeida D, Altenschmidt J, Altman S, Anadkat S, et al. Gpt-4 technical report. (2023). arXiv preprint arXiv:2303.08774."},{"key":"10568_CR3","unstructured":"Team G, Anil R, Borgeaud S, Wu Y, Alayrac J-B, Yu J, Soricut R, Schalkwyk J, Dai AM, Hauth A, Millican K, Silver D, Petrov S, Johnson M, Antonoglou I, Schrittwieser J, Glaese A, Chen J, Pitler E, Lillicrap T, Silver D, Kartsaklis D, Zisserman A, Sifre L, Kavukcuoglu K, Hassabis D, Cornebise J, Luong T. Gemini: A family of highly capable multimodal models. (2023). arXiv preprint arXiv:2312.11805 [cs.CL]."},{"key":"10568_CR4","unstructured":"Askell A, Bai Y, Chen A, Drain D, Ganguli D, Henighan T, Jones A, Joseph N, Mann B, DasSarma N, Elhage N, Hatfield-Dodds Z, Hernandez D, Kernion J, Ndousse K, Olsson C, Amodei D, Brown T, Clark J, McCandlish S, Olah C, Kaplan J. A general language assistant as a laboratory for alignment. (2021). arxiv:2112.00861."},{"key":"10568_CR5","unstructured":"Xie Z, Wu J, Shen Y, Xia Y, Li X, Chang A, Rossi R, Kumar S, Majumder BP, Shang J, Ammanabrolu P, McAuley J. A survey on personalized and pluralistic preference alignment in large language models. (2025). arXiv preprint arXiv:2404.07070 [cs.CL]."},{"key":"10568_CR6","unstructured":"Zeng W, Zhu H, Qin C, Wu H, Cheng Y, Zhang S, Jin X, Shen Y, Wang Z, Zhong F, Xiong H. Multi-level value alignment in agentic ai systems: Survey and perspectives. (2025). arXiv preprint arXiv:2406.09656 [cs.AI]."},{"key":"10568_CR7","doi-asserted-by":"crossref","unstructured":"Nussbaum MC. Creating Capabilities: The Human Development Approach and Its Implementation. Harvard University Press, ??? (2011).","DOI":"10.4159\/harvard.9780674061200"},{"key":"10568_CR8","unstructured":"Russell S. Human compatible. (2019). https:\/\/books.google.com\/books\/about\/Human_Compatible.html?id=8vm0DwAAQBAJ."},{"key":"10568_CR9","unstructured":"Kirk R, Mediratta I, Nalmpantis C, Luketina J, Hambro E, Grefenstette E, Raileanu R. Understanding the effects of rlhf on llm generalisation and diversity. (2024). arXiv preprint arXiv:2310.06452 [cs.LG]."},{"key":"10568_CR10","unstructured":"Zhou D, Zhang J, Feng T, Sun Y. A survey on alignment for large language model agents. arXiv preprint (2024). Reviewed on OpenReview."},{"key":"10568_CR11","unstructured":"Wei J, Tay Y, Bommasani R, Raffel C, Zoph B, Borgeaud S, Yogatama D, Bosma M, Zhou D, Metzler D, Chi E.H., Hashimoto T, Vinyals O, Liang P, Dean J, Fedus W. Emergent abilities of large language models. Trans Mach Learn Res. (2022). arXiv:2206.07682 [cs.CL]."},{"key":"10568_CR12","doi-asserted-by":"crossref","unstructured":"Rafailov R, Sharma A, Mitchell E, Ermon S, Manning CD, Finn C. Direct preference optimization: Your language model is secretly a reward model. Adv Neural Inf Process Syst. 2023;36:53728\u201341. arXiv:2305.18290 [cs.LG].","DOI":"10.52202\/075280-2338"},{"key":"10568_CR13","doi-asserted-by":"crossref","unstructured":"AlKhamissi B, ElNokrashy M, Alkhamissi M, Tan FA, Wattenberg M, Xie SM, Roux NL, Singh S. Investigating cultural alignment of large language models. In: Proceedings of the 62nd (vol 1: Long Papers). 2024. p. 12448\u201312471. arXiv:2402.13231 [cs.CL].","DOI":"10.18653\/v1\/2024.acl-long.671"},{"key":"10568_CR14","unstructured":"Wang Y, Zhong W, Li L, Mi F, Zeng X, Huang W, Shang L, Jiang X, Liu Q. Aligning large language models with human: A survey. 2023. arXiv preprint arXiv:2307.12966 [cs.CL]."},{"key":"10568_CR15","unstructured":"Perez E, Huang S, Song F, Cai T, Ring R, Aslanides J, Glaese A, McAleer N, Irving G. Discovering language model behaviors with model-written evaluations. 2022. arXiv preprint arXiv:2212.09251 [cs.CL]."},{"key":"10568_CR16","unstructured":"Greenblatt R, Denison C, Wright B, Roger F, MacDiarmid M, Marks S, Treutlein J, Belonax T, Chen J, Duvenaud D, et al. Alignment faking in large language models. 2024. arXiv preprint arXiv:2412.14093."},{"key":"10568_CR17","unstructured":"Yao J, Yi X, Wang X, Wang J, Xie X. From instructions to intrinsic human values - a survey of alignment goals for big models. 2023. arXiv preprint arXiv:2308.12014 [cs.AI]."},{"key":"10568_CR18","doi-asserted-by":"crossref","unstructured":"Shen C, Cheng L, Nguyen X-P, You Y, Bing L. Large language models are not yet human-level evaluators for abstractive summarization. In: Findings of the association for computational linguistics: EMNLP 2023. 2023. p. 4215\u20134233","DOI":"10.18653\/v1\/2023.findings-emnlp.278"},{"key":"10568_CR19","unstructured":"Yu T, Zhang Y.-F., Fu C, Wu J, Lu J, Wang K, Lu X, Shen Y, Zhang G, Song D, Yan Y, Xu T, Wen Q, Zhang Z, Huang Y, Wang L, Tan T. Aligning multimodal llm with human preference: A survey. 2025. arXiv preprint arXiv:2403.14504 [cs.CV]."},{"key":"10568_CR20","unstructured":"Zhang Z, Rossi RA, Kveton B, Shao Y, Yang D, Zamani H, Dernoncourt F, Barrow J, Yu T, Kim S, et al. Personalization of large language models: A survey. 2024. arXiv preprint arXiv:2411.00027."},{"key":"10568_CR21","unstructured":"Bai Y, Jones A, Ndousse K, Askell A, Chen A, DasSarma N, Drain D, Fort S, Ganguli D, Henighan T, et al. Training a helpful and harmless assistant with reinforcement learning from human feedback. 2022. arXiv Preprint arXiv:2204.05862."},{"key":"10568_CR22","doi-asserted-by":"crossref","unstructured":"Ji J, Qiu T, Chen B, Zhang B, Lou H, Wang K, Duan Y, He Z, Zhou J, Zhang Z, Zeng F, Ng KY, Dai J, Pan X, O\u2019Gara A, Xu H, Tse B, Fu J, McAleer S, Yang Y, Wang Y, Zhu S-C, Guo Y, Gao W. Ai alignment: A comprehensive survey. 2024. arXiv preprint arXiv:2310.19852 [cs.AI].","DOI":"10.1145\/3770749"},{"key":"10568_CR23","unstructured":"Sorensen T, Moore J, Fisher J, Gordon ML, Mireshghallah N, Rytting CM, Ye A, Jiang L, Lu X, Dziri N, Althoff T, Choi Y. Position: A roadmap to pluralistic alignment. 2024. arXiv preprint arXiv:2402.05070."},{"key":"10568_CR24","unstructured":"Chakraborty S, Qiu J, Yuan H, Koppel A, Huang F, Manocha D, Bedi AS, Wang M. Maxmin-rlhf: Alignment with diverse human preferences. 2024. arXiv preprint arXiv:2402.08925."},{"key":"10568_CR25","doi-asserted-by":"crossref","unstructured":"Perez E, Huang S, Song F, Cai T, Ring R, Aslanides J, Glaese A, McAleese N, Irving G. Red teaming language models with language models. In: Proceedings of the 2022 conference on empirical methods in natural language processing. 2022. p. 3419\u20133448.","DOI":"10.18653\/v1\/2022.emnlp-main.225"},{"key":"10568_CR26","first-page":"9274","volume":"35","author":"D Ziegler","year":"2022","unstructured":"Ziegler D, Nix S, Chan L, Bauman T, Schmidt-Nielsen P, Lin T, et al. Adversarial training for high-stakes reliability. Adv Neural Inf Process Syst. 2022;35:9274\u201386.","journal-title":"Adv Neural Inf Process Syst."},{"key":"10568_CR27","first-page":"27730","volume":"35","author":"L Ouyang","year":"2022","unstructured":"Ouyang L, Wu J, Jiang X, Almeida D, Wainwright C, Mishkin P, et al. Training language models to follow instructions with human feedback. Adv Neural Inf Process Syst. 2022;35:27730\u201344.","journal-title":"Adv Neural Inf Process Syst."},{"key":"10568_CR28","unstructured":"Christiano PF, Leike J, Brown T, Martic M, Legg S, Amodei D. Deep reinforcement learning from human preferences. Adv Neural Inf Process Syst. 2017;30. arXiv:1706.03741 [cs.LG]."},{"key":"10568_CR29","unstructured":"Leike J, Krueger D, Everitt T, Martic M, Maini V, Legg S. Scalable agent alignment via reward modeling: a research direction. 2018. arXiv preprint arXiv:1811.07871."},{"key":"10568_CR30","first-page":"9460","volume":"35","author":"J Skalse","year":"2022","unstructured":"Skalse J, Howe N, Krasheninnikov D, Krueger D. Defining and characterizing reward gaming. Adv Neural Inf Process Syst. 2022;35:9460\u201371.","journal-title":"Adv Neural Inf Process Syst"},{"key":"10568_CR31","unstructured":"Gao L, Schulman J, Hilton J. Scaling laws for reward model overoptimization. In: International conference on machine learning. 2023. p. 10835\u201310866. PMLR."},{"key":"10568_CR32","unstructured":"Bai Y, Kadavath S, Kundu S, Askell A, Kernion J, Jones A, Chen A, Goldie A, Mirhoseini A, McKinnon C, Chen C, Olsson C, Olah C, Hernandez D, Drain D, Ganguli D, Li D, Tran-Johnson E, Perez E, Kerr J, Mueller J, Ladish J, Landau J, Ndousse K, Lukosuite K, Lovitt L, Sellitto M, Elhage N, Schiefer N, Dreksler N, DasSarma N, Rausch O, Larson R, Lasenby R, Yampolskiy R, Emmons S, Ringer S, Kundu S, Kadavath S, Yang S, Changpinyo S, Hawkins W, Katz Y, Bai Y, Witten Z, Rahtz M, McCandlish S, Amodei D, Clark J. Constitutional ai: Harmlessness from ai feedback. 2022. arXiv preprint arXiv:2212.08073 [cs.CL]."},{"key":"10568_CR33","unstructured":"Zhao Y, Joshi R, Liu T, Khalman M, Saleh M, Liu PJ. Slic-hf: Sequence likelihood calibration with human feedback. 2023. arXiv preprint arXiv:2305.10425 [cs.CL]."},{"key":"10568_CR34","unstructured":"Tang Y, Guo DZ, Zheng Z, Calandriello D, Cao Y, Tarassov E, Munos R, Pires B\u00c1, Valko M, Cheng Y, et al. Understanding the performance gap between online and offline alignment algorithms. 2024. arXiv preprint arXiv:2405.08448."},{"key":"10568_CR35","doi-asserted-by":"crossref","unstructured":"Zhang J, Marone M, Li T, Van Durme B, Khashabi D. Verifiable by design: Aligning language models to quote from pre-training data. 2024. arXiv preprint arXiv:2404.03862.","DOI":"10.18653\/v1\/2025.naacl-long.191"},{"key":"10568_CR36","doi-asserted-by":"crossref","unstructured":"Abdulhai M, Crepy C, Valter D, Canny J, Jaques N. Moral foundations of large language models. In: AAAI 2023 Workshop on representation learning for responsible human-centric AI. 2022.","DOI":"10.18653\/v1\/2024.emnlp-main.982"},{"key":"10568_CR37","unstructured":"Hendrycks D, Mazeika M, Zou A, Patel S, Zhu C, Navarro J, Song D, Li B, Steinhardt J. What would jiminy cricket do? towards agents that behave morally. 2021. arXiv preprint arXiv:2110.13136."},{"key":"10568_CR38","doi-asserted-by":"crossref","unstructured":"Gabriel I. Artificial intelligence, values and alignment. Minds Mach. 2020;30:411\u201337. arXiv:2001.09768 [cs.CY].","DOI":"10.1007\/s11023-020-09539-2"},{"issue":"4","key":"10568_CR39","doi-asserted-by":"publisher","first-page":"328","DOI":"10.1086\/256963","volume":"58","author":"KJ Arrow","year":"1950","unstructured":"Arrow KJ. A difficulty in the concept of social welfare. J Polit Econ. 1950;58(4):328\u201346.","journal-title":"J Polit Econ"},{"key":"10568_CR40","doi-asserted-by":"crossref","unstructured":"Birhane A. Algorithmic injustice: a relational ethics approach. Patterns. 2021;2(2).","DOI":"10.1016\/j.patter.2021.100205"},{"key":"10568_CR41","unstructured":"Sen A. Development as Freedom. Anchor Books, ???. 1999."},{"key":"10568_CR42","unstructured":"Hilliard E, Jagadeesh A, Cook A, Billings S, Skytland N, Llewellyn A, Paull J, Paull N, Kurylo N, Nesbitt K, et al. Measuring ai alignment with human flourishing. 2025. arXiv preprint arXiv:2507.07787."},{"key":"10568_CR43","doi-asserted-by":"crossref","unstructured":"Rawls J. A Theory of Justice. Harvard University Press, ???. 1971.","DOI":"10.4159\/9780674042605"},{"key":"10568_CR44","unstructured":"Berlin I. Two concepts of liberty. Four Essays on Liberty. 1969."},{"key":"10568_CR45","doi-asserted-by":"publisher","unstructured":"Yuan Y, Xiao T, Yunfan L, Xu B, Tao S, Qiu Y, Shen H, Cheng X. Inference-time alignment in continuous space. 2025. CoRR arxiv:2505.20081. https:\/\/doi.org\/10.48550\/ARXIV.2505.20081.","DOI":"10.48550\/ARXIV.2505.20081"},{"key":"10568_CR46","unstructured":"Liu T, Guo S, Bianco L, Calandriello D, Berthet Q, Llinares-L\u00f3pez F, Hoffmann J, Dixon L, Valko M, Blondel M. Decoding-time realignment of language models. In: Salakhutdinov R, Kolter Z, Heller K, Weller A, Oliver N, Scarlett J, Berkenkamp F, editors. Proceedings of the 41st international conference on machine learning. Proceedings of machine learning research. 2024. vol 235, p. 31015\u201331031. PMLR, ???. https:\/\/proceedings.mlr.press\/v235\/liu24r.html."},{"key":"10568_CR47","unstructured":"Khanov M, Burapacheep J, Li Y. ARGS: alignment as reward-guided search. In: The Twelfth International Conference on Learning Representations, ICLR 2024, Vienna, Austria, May 7\u201311, 2024. OpenReview.net, ???. 2024. https:\/\/openreview.net\/forum?id=shgx0eqdw6."},{"key":"10568_CR48","doi-asserted-by":"publisher","unstructured":"Dutta S, Kaufmann T, Glava\u0161 G, Habernal I, Kersting K, Kreuter F, Mezini M, Gurevych I, H\u00fcllermeier E, Sch\u00fctze H. Problem solving through human\u2013ai preference-based cooperation. Computat Linguist. 2025;1\u201336. https:\/\/doi.org\/10.1162\/COLI.a.19.","DOI":"10.1162\/COLI.a.19"},{"key":"10568_CR49","doi-asserted-by":"crossref","unstructured":"Guan MY, Joglekar M, Wallace E, Jain S, Barak B, Helyar A, Dias R, Vallone A, Ren H, Wei J, et al. Deliberative alignment: Reasoning enables safer language models. 2024. arXiv preprint arXiv:2412.16339.","DOI":"10.70777\/si.v2i3.15159"},{"key":"10568_CR50","doi-asserted-by":"crossref","unstructured":"Raina V, Liusie A, Gales M. Is LLM-as-a-judge robust? investigating universal adversarial attacks on zero-shot LLM assessment. In: Al-Onaizan Y, Bansal M, Chen Y-N, editors. Proceedings of the 2024 conference on empirical methods in natural language processing. 2024. p. 7499\u20137517. Association for Computational Linguistics, Miami, Florida, USA. https:\/\/doi.org\/10.18653\/v1\/2024.emnlp-main.427. https:\/\/aclanthology.org\/2024.emnlp-main.427\/.","DOI":"10.18653\/v1\/2024.emnlp-main.427"},{"key":"10568_CR51","doi-asserted-by":"publisher","unstructured":"Zhao Y, Liu H, Yu D, Kung SY, Mi H, Yu D. One token to fool llm-as-a-judge. 2025. CoRR arxiv:2507.08794. https:\/\/doi.org\/10.48550\/ARXIV.2507.08794.","DOI":"10.48550\/ARXIV.2507.08794"},{"key":"10568_CR52","doi-asserted-by":"publisher","unstructured":"Liu Z, Wang P, Xu R, Ma S, Ruan C, Li P, Liu Y, Wu Y. Inference-time scaling for generalist reward modeling. 2025. CoRR arxiv:2504.02495. https:\/\/doi.org\/10.48550\/ARXIV.2504.02495.","DOI":"10.48550\/ARXIV.2504.02495"},{"key":"10568_CR53","doi-asserted-by":"publisher","unstructured":"Chen X, Li G, Wang Z, Jin B, Qian C, Wang Y, Wang H, Zhang Y, Zhang D, Zhang T, Tong H, Ji H. RM-R1: reward modeling as reasoning. 2025. CoRR arxiv:2505.02387. https:\/\/doi.org\/10.48550\/ARXIV.2505.02387.","DOI":"10.48550\/ARXIV.2505.02387"},{"key":"10568_CR54","doi-asserted-by":"publisher","unstructured":"Guo J, Chi Z, Dong L, Dong Q, Wu X, Huang S, Wei F. Reward reasoning model. 2025. CoRR arxiv:2505.14674. https:\/\/doi.org\/10.48550\/ARXIV.2505.14674.","DOI":"10.48550\/ARXIV.2505.14674"},{"key":"10568_CR55","doi-asserted-by":"publisher","unstructured":"Zhao J, Liu R, Zhang K, Zhou Z, Gao J, Li D, Lyu J, Qian Z, Qi B, Li X, Zhou B. Genprm: Scaling test-time compute of process reward models via generative reasoning. 2025. CoRR arxiv:2504.00891. https:\/\/doi.org\/10.48550\/ARXIV.2504.00891.","DOI":"10.48550\/ARXIV.2504.00891"},{"key":"10568_CR56","unstructured":"Anderljung M, Barnhart J, Leung J, Korinek A, O\u2019Keefe C, Whittlestone J, Avin S, Brundage M, Bullock J, Cass-Beggs D, et al. Frontier ai regulation: Managing emerging risks to public safety. 2023. arXiv preprint arXiv:2307.03718."},{"key":"10568_CR57","doi-asserted-by":"crossref","unstructured":"Feng S, Sorensen T, Liu Y, Fisher J, Park CY, Choi Y, Tsvetkov Y. Modular pluralism: Pluralistic alignment via multi-llm collaboration. 2024. arXiv preprint arXiv:2406.15951.","DOI":"10.18653\/v1\/2024.emnlp-main.240"},{"key":"10568_CR58","doi-asserted-by":"publisher","first-page":"71","DOI":"10.1007\/s11023-012-9281-3","volume":"22","author":"N Bostrom","year":"2012","unstructured":"Bostrom N. The superintelligent will: Motivation and instrumental rationality in advanced artificial agents. Minds Mach. 2012;22:71\u201385.","journal-title":"Minds Mach"},{"key":"10568_CR59","doi-asserted-by":"crossref","unstructured":"Hendrycks D, Mazeika M, Woodside T. An overview of catastrophic ai risks. 2023. arXiv preprint arXiv:2306.12001.","DOI":"10.1201\/9781003530336-1"},{"key":"10568_CR60","unstructured":"Carlsmith J. Is power-seeking ai an existential risk? 2022. arXiv preprint arXiv:2206.13353."},{"key":"10568_CR61","unstructured":"Michael J, Mahdi S, Rein D, Petty J, Dirani J, Padmakumar V, Bowman SR. Debate helps supervise unreliable experts. 2023. arXiv preprint arXiv:2311.08702."},{"key":"10568_CR62","unstructured":"Lanham T, Chen A, Radhakrishnan A, Steiner B, Denison C, Hernandez D, Li D, Durmus E, Hubinger E, Kernion J, et al. Measuring faithfulness in chain-of-thought reasoning. 2023. arXiv preprint arXiv:2307.13702."},{"key":"10568_CR63","doi-asserted-by":"crossref","unstructured":"Olah C, Cammarata N, Schubert L, Goh G, Petrov M, Carter S. Zoom in An introduction to circuits. Distill. 2020;5(3):00024\u2013001.","DOI":"10.23915\/distill.00024.001"},{"key":"10568_CR64","unstructured":"Elhage N, Hume T, Olsson C, Schiefer N, Henighan T, Kravec S, Hatfield-Dodds Z, Lasenby R, Drain D, Chen C, et al. Toy models of superposition. 2022. arXiv preprint arXiv:2209.10652."},{"key":"10568_CR65","doi-asserted-by":"crossref","unstructured":"Hubinger E. Anthropic: Responsible scaling policy. SuperIntell-Robot-Safety Alignm. 2025;2(1).","DOI":"10.70777\/si.v2i1.13657"},{"key":"10568_CR66","unstructured":"Commission E. EU AI Act. 2024. https:\/\/digital-strategy.ec.europa.eu\/en\/policies\/european-approach-artificial-intelligence."},{"key":"10568_CR67","unstructured":"OpenAI Safety Evaluations Hub. 2025. https:\/\/openai.com\/safety\/evaluations-hub\/."},{"key":"10568_CR68","unstructured":"Kirk HR, Whitefield A, R\u00f6ttger P, Bean A, Ciro J, Mosquera R, Bartolo M, Williams A, He H, et al. The prism alignment project: What participatory, representative and individualised human feedback reveals about the subjective and multicultural alignment of large language models. 2024. arXiv preprint arXiv:2404.16019."}],"container-title":["Cognitive Computation"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s12559-026-10568-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s12559-026-10568-9","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s12559-026-10568-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,13]],"date-time":"2026-03-13T02:10:32Z","timestamp":1773367832000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s12559-026-10568-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,3,13]]},"references-count":68,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,12]]}},"alternative-id":["10568"],"URL":"https:\/\/doi.org\/10.1007\/s12559-026-10568-9","relation":{},"ISSN":["1866-9956","1866-9964"],"issn-type":[{"value":"1866-9956","type":"print"},{"value":"1866-9964","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,3,13]]},"assertion":[{"value":"11 January 2026","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"15 February 2026","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"13 March 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing Interests"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics approval and consent to participate"}}],"article-number":"26"}}