{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,30]],"date-time":"2026-03-30T16:00:22Z","timestamp":1774886422406,"version":"3.50.1"},"reference-count":40,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2024,12,20]],"date-time":"2024-12-20T00:00:00Z","timestamp":1734652800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2024,12,20]],"date-time":"2024-12-20T00:00:00Z","timestamp":1734652800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"name":"National Key Research and Development Program of China","award":["2022ZD0160103"],"award-info":[{"award-number":["2022ZD0160103"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["L2224015"],"award-info":[{"award-number":["L2224015"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Academic Divisions of the Chinese Academy of Sciences \u201cDigital Technology Ethics Research\u201d"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Ethics Inf Technol"],"published-print":{"date-parts":[[2025,3]]},"DOI":"10.1007\/s10676-024-09818-x","type":"journal-article","created":{"date-parts":[[2024,12,20]],"date-time":"2024-12-20T13:56:50Z","timestamp":1734703010000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":10,"title":["Possibilities and challenges in the moral growth of large language models: a philosophical perspective"],"prefix":"10.1007","volume":"27","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-4330-2629","authenticated-orcid":false,"given":"Guoyu","family":"Wang","sequence":"first","affiliation":[]},{"given":"Wei","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Yiqin","family":"Cao","sequence":"additional","affiliation":[]},{"given":"Yan","family":"Teng","sequence":"additional","affiliation":[]},{"given":"Qianyu","family":"Guo","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3018-3824","authenticated-orcid":false,"given":"Haofen","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Junyu","family":"Lin","sequence":"additional","affiliation":[]},{"given":"Jiajie","family":"Ma","sequence":"additional","affiliation":[]},{"given":"Jin","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Yingchun","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,12,20]]},"reference":[{"key":"9818_CR1","unstructured":"Ames, R. T., & Rosemont, H. (2010). The analects of confucius: A philosophical translation. Random House Publishing Group."},{"issue":"4047","key":"9818_CR2","doi-asserted-by":"publisher","first-page":"393","DOI":"10.1126\/science.177.4047.393","volume":"177","author":"PW Anderson","year":"1972","unstructured":"Anderson, P. W. (1972). More Is Different: Broken symmetry and the nature of the hierarchical structure of science. Science, 177(4047), 393\u2013396.","journal-title":"Science"},{"key":"9818_CR3","unstructured":"Bai, Y., Kadavath, S., Kundu, S., Askell, A., Kernion, J., Jones, A., Chen, A., Goldie, A., Mirhoseini, A., & McKinnon, C. (2022). Constitutional ai: Harmlessness from ai feedback. arXiv:2212.08073"},{"key":"9818_CR4","unstructured":"Bommasani, R., Hudson, D. A., Adeli, E., Altman, R., Arora, S., von Arx, S., Bernstein, M. S., Bohg, J., Bosselut, A., & Brunskill, E. (2021). On the opportunities and risks of foundation models. arXiv:2108.07258"},{"issue":"1\u20133","key":"9818_CR5","doi-asserted-by":"publisher","first-page":"139","DOI":"10.1016\/0004-3702(91)90053-M","volume":"47","author":"RA Brooks","year":"1991","unstructured":"Brooks, R. A. (1991). Intelligence without representation. Artificial Intelligence, 47(1\u20133), 139\u2013159.","journal-title":"Artificial Intelligence"},{"key":"9818_CR6","first-page":"1877","volume":"33","author":"T Brown","year":"2020","unstructured":"Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J. D., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., & Askell, A. (2020). Language models are few-shot learners. Advances in Neural Information Processing Systems, 33, 1877\u20131901.","journal-title":"Advances in Neural Information Processing Systems"},{"issue":"1","key":"9818_CR7","doi-asserted-by":"publisher","first-page":"15","DOI":"10.1007\/s10676-018-9448-6","volume":"20","author":"JJ Bryson","year":"2018","unstructured":"Bryson, J. J. (2018). Patiency is not a virtue: The design of intelligent systems and systems of ethics. Ethics and Information Technology, 20(1), 15\u201326. https:\/\/doi.org\/10.1007\/s10676-018-9448-6","journal-title":"Ethics and Information Technology"},{"issue":"1","key":"9818_CR8","doi-asserted-by":"publisher","first-page":"91","DOI":"10.1007\/s13218-021-00705-x","volume":"35","author":"MV Butz","year":"2021","unstructured":"Butz, M. V. (2021). Towards strong AI. KI\u2014K\u00fcnstliche Intelligenz, 35(1), 91\u2013101. https:\/\/doi.org\/10.1007\/s13218-021-00705-x","journal-title":"KI\u2014K\u00fcnstliche Intelligenz"},{"issue":"1","key":"9818_CR9","doi-asserted-by":"publisher","first-page":"55","DOI":"10.1007\/s43681-020-00005-4","volume":"1","author":"H Cai","year":"2021","unstructured":"Cai, H. (2021). Reaching consensus with human beings through blockchain as an ethical rule of strong artificial intelligence. AI and Ethics, 1(1), 55\u201359. https:\/\/doi.org\/10.1007\/s43681-020-00005-4","journal-title":"AI and Ethics"},{"key":"9818_CR10","unstructured":"Chowdhery, A., Narang, S., Devlin, J., Bosma, M., Mishra, G., Roberts, A., Barham, P., Chung, H. W., Sutton, C., & Gehrmann, S. (2022). Palm: Scaling language modeling with pathways. arXiv:2204.02311"},{"key":"9818_CR11","unstructured":"Christiano, P., Shlegeris, B., & Amodei, D. (2018). Supervising strong learners by amplifying weak experts. arXiv:1810.08575"},{"key":"9818_CR12","unstructured":"Dewey, J. (1981). The later works, 1925\u20131953. Southern Illinois University Press."},{"key":"9818_CR14","unstructured":"Dewey, J. (2008a). The middle works of John Dewey, volume 5, 1899\u20131924: Ethics 1908. Southern Illinois University Press."},{"key":"9818_CR13","unstructured":"Dewey, J. (2008b). The middle works of John Dewey, volume 14, 1899\u20131924: Human nature and conduct 1922. Southern Illinois University Press."},{"key":"9818_CR15","unstructured":"Dewey, J. (2008). Experience and nature. McCutchen Press."},{"key":"9818_CR16","volume-title":"Democracy and education","author":"J Dewey","year":"1997","unstructured":"Dewey, J. (1997). Democracy and education. Free Press."},{"key":"9818_CR17","unstructured":"Docherty, B. (2016, June 16). Losing control: The dangers of killer robots. The Conversation. Retrieved October 19, 2023, from https:\/\/theconversation.com\/losing-control-the-dangers-of-killer-robots-58262"},{"key":"9818_CR18","first-page":"5","volume":"5","author":"P Foot","year":"1967","unstructured":"Foot, P. (1967). The problem of abortion and the doctrine of the double effect. Oxford Review, 5, 5\u201315.","journal-title":"Oxford Review"},{"issue":"3","key":"9818_CR19","doi-asserted-by":"publisher","first-page":"411","DOI":"10.1007\/s11023-020-09539-2","volume":"30","author":"I Gabriel","year":"2020","unstructured":"Gabriel, I. (2020). Artificial intelligence, values, and alignment. Minds and Machines, 30(3), 411\u2013437.","journal-title":"Minds and Machines"},{"key":"9818_CR20","doi-asserted-by":"crossref","unstructured":"Greene, J. D., Sommerville, R. B., Nystrom, L. E., Darley, J. M., & Cohen, J. D. (2001). An fMRI investigation of emotional engagement in moral judgment. Science, 293(5537), 2105\u20132108. http:\/\/www.jstor.org\/stable\/3084564","DOI":"10.1126\/science.1062872"},{"key":"9818_CR21","unstructured":"Knox, W. B., & Stone, P. (2011). Augmenting reinforcement learning with human feedback. In ICML 2011 workshop on new developments in imitation learning (July 2011)."},{"key":"9818_CR22","unstructured":"Kohlberg, L. (1958). The development of modes of moral thinking and choice in the years 10 to 16 The University of Chicago."},{"key":"9818_CR23","unstructured":"Lambert, N., Castricato, L., von Werra, L., & Havrilla, A. (2022). Illustrating reinforcement learning from human feedback (rlhf). Hugging Face Blog. https:\/\/huggingface.co\/blog\/rlhf"},{"key":"9818_CR24","unstructured":"Maslej, N., Fattorini, L., Brynjolfsson, E., Etchemendy, J., Ligett, K., Lyons, T., Manyika, J., Ngo, H., Niebles, J. C., Parli, V., Shoham, Y., Wald, R., Clark, J., & Perrault, R. (2023). The AI Index 2023 Annual Report. AI Index Steering Committee, Institute for Human-Centered AI, Stanford University."},{"key":"9818_CR25","volume-title":"Artificial intelligence: A guide for thinking humans","author":"M Mitchell","year":"2019","unstructured":"Mitchell, M. (2019). Artificial intelligence: A guide for thinking humans. Farrar."},{"key":"9818_CR40","unstructured":"Murphey, M. G. (2008). Introduction. In J. Dewey, The middle works of John Dewey, volume 14, 1899\u20131924: Human nature and conduct 1922 (pp. ix\u2013xxiii). Southern Illinois University Press."},{"key":"9818_CR26","unstructured":"Ouyang, L., Wu, J., Jiang, X., Almeida, D., Wainwright, C. L., Mishkin, P., Zhang, C., Agarwal, S., Slama, K., Ray, A., Schulman, J., Hilton, J., Kelton, F., Miller, L., Simens, M., Askell, A., Welinder, P., Christiano, P., Leike, J., & Lowe, R. (2022). Training language models to follow instructions with human feedback. arXiv:2203.02155"},{"issue":"4","key":"9818_CR27","doi-asserted-by":"publisher","first-page":"46","DOI":"10.1109\/MIS.2006.77","volume":"21","author":"TM Powers","year":"2006","unstructured":"Powers, T. M. (2006). Prospects for a Kantian machine. IEEE Intelligent Systems, 21(4), 46\u201351.","journal-title":"IEEE Intelligent Systems"},{"issue":"4","key":"9818_CR28","doi-asserted-by":"publisher","first-page":"851","DOI":"10.1007\/s10677-015-9563-y","volume":"18","author":"D Purves","year":"2015","unstructured":"Purves, D., Jenkins, R., & Strawser, B. J. (2015). Autonomous machines, moral judgment, and acting for the right reasons. Ethical Theory and Moral Practice, 18(4), 851\u2013872. https:\/\/doi.org\/10.1007\/s10677-015-9563-y","journal-title":"Ethical Theory and Moral Practice"},{"key":"9818_CR29","unstructured":"Radford, A., Narasimhan, K., Salimans, T., & Sutskever, I. (2018). Improving language understanding by generative pre-training. OpenAI. https:\/\/cdn.openai.com\/research-covers\/language-unsupervised\/language_understanding_paper.pdf"},{"issue":"8","key":"9818_CR30","first-page":"9","volume":"1","author":"A Radford","year":"2019","unstructured":"Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., & Sutskever, I. (2019). Language models are unsupervised multitask learners. OpenAI Blog, 1(8), 9.","journal-title":"OpenAI Blog"},{"key":"9818_CR31","unstructured":"Schulman, J., Wolski, F., Dhariwal, P., Radford, A., & Klimov, O. (2017). Proximal policy optimization algorithms. arXiv:1707.06347"},{"issue":"4","key":"9818_CR32","doi-asserted-by":"publisher","first-page":"283","DOI":"10.1007\/s10676-017-9425-5","volume":"22","author":"A Sharkey","year":"2020","unstructured":"Sharkey, A. (2020). Can we program or train robots to be good? Ethics and Information Technology, 22(4), 283\u2013295. https:\/\/doi.org\/10.1007\/s10676-017-9425-5","journal-title":"Ethics and Information Technology"},{"key":"9818_CR33","doi-asserted-by":"crossref","unstructured":"Sun H., Zhang Z., Mi F., Wang Y., Liu W., Cui J., Wang B., Liu Q., & Huang M. (2023). Moraldial: A framework to train and evaluate moral dialogue systems via moral discussions. In Proceedings of the 61st annual meeting of the association for computational linguistics (vol. 1: Long Papers, pp. 2213\u20132230).","DOI":"10.18653\/v1\/2023.acl-long.123"},{"key":"9818_CR34","volume-title":"Reinforcement learning","author":"RS Sutton","year":"2018","unstructured":"Sutton, R. S., & Barto, A. G. (2018). Reinforcement learning (2nd ed.). MIT Press.","edition":"2"},{"issue":"6","key":"9818_CR35","doi-asserted-by":"publisher","first-page":"1395","DOI":"10.2307\/796133","volume":"94","author":"JJ Thomson","year":"1985","unstructured":"Thomson, J. J. (1985). The Trolley problem. The Yale Law Journal, 94(6), 1395\u20131415. https:\/\/doi.org\/10.2307\/796133","journal-title":"The Yale Law Journal"},{"key":"9818_CR36","unstructured":"Wallach, W., & Allen, C. (2008). Moral machines: Teaching robots right from wrong. Oxford University Press, Inc."},{"issue":"18","key":"9818_CR37","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.2213709120","volume":"120","author":"L Weidinger","year":"2023","unstructured":"Weidinger, L., McKee, K. R., Everett, R., Huang, S., Zhu, T. O., Chadwick, M. J., Summerfield, C., & Gabriel, I. (2023). Using the Veil of Ignorance to align AI systems with principles of justice. Proceedings of the National Academy of Sciences of the United States of America, 120(18), e2213709120.","journal-title":"Proceedings of the National Academy of Sciences of the United States of America"},{"issue":"68","key":"9818_CR38","doi-asserted-by":"publisher","first-page":"eabm4183","DOI":"10.1126\/scirobotics.abm4183","volume":"7","author":"L Yuan","year":"2022","unstructured":"Yuan, L., Gao, X., Zheng, Z., Edmonds, M., Wu, Y. N., Rossano, F., Lu, H., Zhu, Y., & Zhu, S.-C. (2022). In situ bidirectional human-robot value alignment. Science Robotics, 7(68), eabm4183.","journal-title":"Science Robotics"},{"key":"9818_CR39","doi-asserted-by":"crossref","unstructured":"Ziems C., Yu J. A., Wang Y.-C., Halevy A., & Yang D. (2022). The moral integrity corpus: A benchmark for ethical dialogue systems. arXiv:2204.03021","DOI":"10.18653\/v1\/2022.acl-long.261"}],"container-title":["Ethics and Information Technology"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10676-024-09818-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10676-024-09818-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10676-024-09818-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,4,5]],"date-time":"2025-04-05T19:45:19Z","timestamp":1743882319000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10676-024-09818-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,20]]},"references-count":40,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2025,3]]}},"alternative-id":["9818"],"URL":"https:\/\/doi.org\/10.1007\/s10676-024-09818-x","relation":{},"ISSN":["1388-1957","1572-8439"],"issn-type":[{"value":"1388-1957","type":"print"},{"value":"1572-8439","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,12,20]]},"assertion":[{"value":"5 December 2024","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 December 2024","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no competing interests to declare that are relevant to the content of this article.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"This article does not contain any studies with human participants performed by any of the authors.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}],"article-number":"9"}}