{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,23]],"date-time":"2026-02-23T12:05:01Z","timestamp":1771848301158,"version":"3.50.1"},"reference-count":24,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2025,5,18]],"date-time":"2025-05-18T00:00:00Z","timestamp":1747526400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2025,5,18]],"date-time":"2025-05-18T00:00:00Z","timestamp":1747526400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100005357","name":"Agent\u00fara na Podporu V\u00fdskumu a V\u00fdvoja","doi-asserted-by":"publisher","award":["APVV-22-0323"],"award-info":[{"award-number":["APVV-22-0323"]}],"id":[{"id":"10.13039\/501100005357","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100006109","name":"Vedeck\u00e1 Grantov\u00e1 Agent\u00fara M\u0160VVa\u0160 SR a SAV","doi-asserted-by":"publisher","award":["VEGA 2\/0125\/22"],"award-info":[{"award-number":["VEGA 2\/0125\/22"]}],"id":[{"id":"10.13039\/501100006109","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000925","name":"John Templeton Foundation","doi-asserted-by":"publisher","award":["University of Oxford subgrant"],"award-info":[{"award-number":["University of Oxford subgrant"]}],"id":[{"id":"10.13039\/100000925","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100010783","name":"Slovak Academy of Sciences","doi-asserted-by":"crossref","id":[{"id":"10.13039\/100010783","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Ethics Inf Technol"],"published-print":{"date-parts":[[2025,6]]},"abstract":"<jats:title>Abstract<\/jats:title>\n          <jats:p>The present paper tackles what we might call <jats:italic>the AI achievement challenge<\/jats:italic>, which has been the subject of recent debate in AI ethics. The challenge concerns the question of whether there are any achievement gaps due to artificial intelligence and what we should do if there are: how to fill them, or what policies are needed to decrease their impact on us if they cannot be filled. This paper argues that none of the proposed views is entirely satisfactory, even though they all have certain merits. The paper will provide two mutually compatible answers to the AI achievement challenge: one in terms of collective achievement and the other in terms of vicarious achievement.<\/jats:p>","DOI":"10.1007\/s10676-025-09836-3","type":"journal-article","created":{"date-parts":[[2025,5,18]],"date-time":"2025-05-18T02:13:41Z","timestamp":1747534421000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Meeting the AI achievement challenge: collective and vicarious achievements"],"prefix":"10.1007","volume":"27","author":[{"given":"Daniela","family":"Vacek","sequence":"first","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,5,18]]},"reference":[{"key":"9836_CR1","doi-asserted-by":"publisher","DOI":"10.1093\/acprof:oso\/9780198714026.001.0001","volume-title":"Achievement","author":"G Bradford","year":"2015","unstructured":"Bradford, G. (2015). Achievement. Oxford University Press."},{"key":"9836_CR2","doi-asserted-by":"publisher","first-page":"63","DOI":"10.1075\/nlp.8.11bry","volume":"8","author":"JJ Bryson","year":"2010","unstructured":"Bryson, J. J. (2010). Robots should be slaves. Close engagements with artificial companions: Key social, psychological, ethical and design issues, 8, 63\u201374.","journal-title":"Close engagements with artificial companions: Key social, psychological, ethical and design issues"},{"issue":"1","key":"9836_CR3","doi-asserted-by":"publisher","first-page":"15","DOI":"10.1007\/s10676-018-9448-6","volume":"20","author":"JJ Bryson","year":"2018","unstructured":"Bryson, J. J. (2018). Patiency is not a virtue: The design of intelligent systems and systems of ethics. Ethics and Information Technology, 20(1), 15\u201326.","journal-title":"Ethics and Information Technology"},{"issue":"4","key":"9836_CR4","doi-asserted-by":"publisher","first-page":"299","DOI":"10.1007\/s10676-016-9403-3","volume":"18","author":"J Danaher","year":"2016","unstructured":"Danaher, J. (2016). Robots, law and the retribution gap. Ethics and Information Technology, 18(4), 299\u2013309.","journal-title":"Ethics and Information Technology"},{"issue":"3","key":"9836_CR5","doi-asserted-by":"publisher","first-page":"227","DOI":"10.1007\/s43681-020-00028-x","volume":"1","author":"J Danaher","year":"2021","unstructured":"Danaher, J., & Nyholm, S. (2021). Automation, work and the achievement gap. AI and Ethics, 1(3), 227\u2013237.","journal-title":"AI and Ethics"},{"key":"9836_CR6","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-30371-6","volume-title":"Responsible artificial intelligence: How to develop and use AI in a responsible way","author":"V Dignum","year":"2019","unstructured":"Dignum, V. (2019). Responsible artificial intelligence: How to develop and use AI in a responsible way (Vol. 2156). Springer."},{"issue":"3","key":"9836_CR7","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1007\/s10676-022-09657-8","volume":"24","author":"D Glavani\u010dov\u00e1","year":"2022","unstructured":"Glavani\u010dov\u00e1, D., & Pascucci, M. (2022). Vicarious liability: A solution to a problem of AI responsibility? Ethics and Information Technology, 24(3), 28.","journal-title":"Ethics and Information Technology"},{"issue":"1","key":"9836_CR8","doi-asserted-by":"publisher","first-page":"107","DOI":"10.1007\/s10670-022-00525-x","volume":"89","author":"D Glavani\u010dov\u00e1","year":"2024","unstructured":"Glavani\u010dov\u00e1, D., & Pascucci, M. (2024). Making sense of vicarious responsibility: Moral philosophy meets legal theory. Erkenntnis, 89(1), 107\u2013128.","journal-title":"Erkenntnis"},{"key":"9836_CR9","doi-asserted-by":"publisher","first-page":"307","DOI":"10.1007\/s10676-017-9428-2","volume":"22","author":"DJ Gunkel","year":"2020","unstructured":"Gunkel, D. J. (2020). Mind the gap: Responsible robotics and the problem of responsibility. Ethics and Information Technology, 22, 307\u2013320.","journal-title":"Ethics and Information Technology"},{"issue":"2","key":"9836_CR10","doi-asserted-by":"publisher","first-page":"259","DOI":"10.1093\/monist\/onz009","volume":"102","author":"R Hakli","year":"2019","unstructured":"Hakli, R., & M\u00e4kel\u00e4, P. (2019). Moral responsibility of robots and hybrid agents. The Monist, 102(2), 259\u2013275.","journal-title":"The Monist"},{"issue":"3","key":"9836_CR11","doi-asserted-by":"publisher","first-page":"40","DOI":"10.1007\/s10676-023-09713-x","volume":"25","author":"B Karlan","year":"2023","unstructured":"Karlan, B. (2023). Human achievement and artificial intelligence. Ethics and Information Technology, 25(3), 40.","journal-title":"Ethics and Information Technology"},{"issue":"4","key":"9836_CR12","doi-asserted-by":"publisher","first-page":"575","DOI":"10.1007\/s10677-022-10313-9","volume":"25","author":"M Kiener","year":"2022","unstructured":"Kiener, M. (2022). Can we bridge AI\u2019s responsibility gap at Will? Ethical Theory and Moral Practice, 25(4), 575\u2013593.","journal-title":"Ethical Theory and Moral Practice"},{"issue":"1","key":"9836_CR13","doi-asserted-by":"publisher","first-page":"32","DOI":"10.1093\/analys\/anad052","volume":"84","author":"PH Kieval","year":"2024","unstructured":"Kieval, P. H. (2024). Artificial achievements. Analysis, 84(1), 32\u201341.","journal-title":"Analysis"},{"issue":"1","key":"9836_CR14","doi-asserted-by":"publisher","first-page":"25","DOI":"10.1007\/s13347-024-00708-0","volume":"37","author":"A Kuchtov\u00e1","year":"2024","unstructured":"Kuchtov\u00e1, A. (2024). The incalculability of the generated text. Philosophy & Technology, 37(1), 25.","journal-title":"Philosophy & Technology"},{"key":"9836_CR15","doi-asserted-by":"publisher","first-page":"175","DOI":"10.1007\/s10676-004-3422-1","volume":"6","author":"A Matthias","year":"2004","unstructured":"Matthias, A. (2004). The responsibility gap: Ascribing responsibility for the actions of learning automata. Ethics and Information Technology, 6, 175\u2013183.","journal-title":"Ethics and Information Technology"},{"issue":"4","key":"9836_CR16","doi-asserted-by":"publisher","first-page":"1201","DOI":"10.1007\/s11948-017-9943-x","volume":"24","author":"S Nyholm","year":"2018","unstructured":"Nyholm, S. (2018). Attributing agency to automated systems: Reflections on human-robot collaborations and responsibility-loci. Science and Engineering Ethics, 24(4), 1201\u20131219.","journal-title":"Science and Engineering Ethics"},{"key":"9836_CR17","first-page":"191","volume-title":"Risk and responsibility in context","author":"S Nyholm","year":"2024","unstructured":"Nyholm, S. (2024). Responsibility gaps, value alignment, and meaningful human control over artificial intelligence. In A. Placani & S. Broadhead (Eds.), Risk and responsibility in context (pp. 191\u2013213). Routledge."},{"issue":"2","key":"9836_CR18","doi-asserted-by":"publisher","first-page":"20","DOI":"10.1007\/s13347-023-00620-z","volume":"36","author":"S Nyholm","year":"2023","unstructured":"Nyholm, S., & R\u00fcther, M. (2023). Meaning in life in AI ethics-some trends and perspectives. Philosophy & Technology, 36(2), 20.","journal-title":"Philosophy & Technology"},{"issue":"5","key":"9836_CR19","doi-asserted-by":"publisher","first-page":"472","DOI":"10.1038\/s42256-023-00653-1","volume":"5","author":"S Porsdam Mann","year":"2023","unstructured":"Porsdam Mann, S., Earp, B. D., Nyholm, S., Danaher, J., M\u00f8ller, N., Bowman-Smart, H., Hatherley, J., Koplin, J., Plozza, M., Rodger, D., Treit, P. V., Renard, G., McMillan, J., & Savulescu, J. (2023). Generative AI entails a credit-blame asymmetry. Nature Machine Intelligence, 5(5), 472\u2013475.","journal-title":"Nature Machine Intelligence"},{"issue":"4","key":"9836_CR20","doi-asserted-by":"publisher","first-page":"1057","DOI":"10.1007\/s13347-021-00450-x","volume":"34","author":"F Santoni de Sio","year":"2021","unstructured":"Santoni de Sio, F., & Mecacci, G. (2021). Four responsibility gaps with artificial intelligence: Why they matter and how to address them. Philosophy & Technology, 34(4), 1057\u20131084.","journal-title":"Philosophy & Technology"},{"key":"9836_CR21","doi-asserted-by":"crossref","unstructured":"Scripter, L. (2024). The achievement gap thesis reconsidered: Artificial intelligence, automation, and meaningful work. Online first in AI & Society, 1-14.","DOI":"10.1007\/s00146-023-01828-5"},{"issue":"1","key":"9836_CR22","doi-asserted-by":"publisher","first-page":"27","DOI":"10.1007\/s13347-024-00718-y","volume":"37","author":"I Taylor","year":"2024","unstructured":"Taylor, I. (2024). Collective responsibility and artificial intelligence. Philosophy & Technology, 37(1), 27.","journal-title":"Philosophy & Technology"},{"issue":"3","key":"9836_CR23","doi-asserted-by":"publisher","first-page":"589","DOI":"10.1007\/s13347-020-00414-7","volume":"34","author":"DW Tigard","year":"2021","unstructured":"Tigard, D. W. (2021). There is no techno-responsibility gap. Philosophy & Technology, 34(3), 589\u2013607.","journal-title":"Philosophy & Technology"},{"issue":"4","key":"9836_CR24","doi-asserted-by":"publisher","first-page":"611","DOI":"10.1007\/s43681-021-00064-1","volume":"1","author":"DW Tigard","year":"2021","unstructured":"Tigard, D. W. (2021). Workplace automation without achievement gaps: A reply to Danaher and Nyholm. AI and Ethics, 1(4), 611\u2013617.","journal-title":"AI and Ethics"}],"container-title":["Ethics and Information Technology"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10676-025-09836-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10676-025-09836-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10676-025-09836-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,6]],"date-time":"2025-09-06T15:02:29Z","timestamp":1757170949000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10676-025-09836-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,18]]},"references-count":24,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2025,6]]}},"alternative-id":["9836"],"URL":"https:\/\/doi.org\/10.1007\/s10676-025-09836-3","relation":{},"ISSN":["1388-1957","1572-8439"],"issn-type":[{"value":"1388-1957","type":"print"},{"value":"1572-8439","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,5,18]]},"assertion":[{"value":"18 May 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"This work was supported by the Slovak Research and Development Agency under the Contract no. APVV-22-0323 \u2018Philosophical and methodological challenges of intelligent technologies\u2019 and by VEGA 2\/0125\/22 \u2018Responsibility and Modal Logic\u2019. This research was also supported by the University of Oxford project \u2018New Horizons for Science and Religion in Central and Eastern Europe\u2019 funded by the John Templeton Foundation (subgrant \u2018Persons of Responsibility: Human, Animal, Artificial, Divine\u2019). The opinions expressed in the publication are those of the author and do not necessarily reflect the view of the John Templeton Foundation. The author has no other relevant financial or non-financial interests to disclose.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"25"}}