{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T12:23:12Z","timestamp":1776082992836,"version":"3.50.1"},"reference-count":27,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2023,12,12]],"date-time":"2023-12-12T00:00:00Z","timestamp":1702339200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2023,12,12]],"date-time":"2023-12-12T00:00:00Z","timestamp":1702339200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"name":"European Union\u2019s Horizon 2020 Research and Innovation Programme","award":["739578"],"award-info":[{"award-number":["739578"]}]},{"name":"European Union\u2019s Horizon 2020 Research and Innovation Programme","award":["739578"],"award-info":[{"award-number":["739578"]}]},{"name":"The Government of the Republic of Cyprus","award":["739578"],"award-info":[{"award-number":["739578"]}]},{"name":"The Government of the Republic of Cyprus","award":["739578"],"award-info":[{"award-number":["739578"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Discov Artif Intell"],"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Algorithms have greatly advanced and become integrated into our everyday lives. Although they support humans in daily functions, they often exhibit unwanted behaviors perpetuating social stereotypes, discrimination, and other forms of biases. Regardless of their accuracy on task, many algorithms do not get scrutinized for unintended behaviors in a systematic way. This phenomenon can propagate and amplify existing societal issues or even create new ones. Many have called for human supervision (human oversight) of algorithmic processes. Oversight is often presented as a way of monitoring algorithmic behavior, as to then address identified issues, by initiating a fix or even correcting the final decision. Unfortunately, a common consensus is missing in the scientific community as to what all human oversight entails. Most importantly, the requirements for a successful application of a human oversight process are only vaguely defined. To address this, we present a critical synthesis of five key articles from different domains, which discuss requirements for human oversight. We use the concept of the Society-in-the-Loop (SITL) [1] as the baseline for understanding and mapping these requirements. In addition, we comment on the requirements and the overall multidisciplinary trend around the topic. Then, we present the concept of a Modular Oversight Methodology (MOM) following the SITL viewpoint, by also considering the requirements identified from the selected literature. Finally, we present a set of suggestions and future work required for a successful application of a human oversight process in a SITL approach.<\/jats:p>","DOI":"10.1007\/s44163-023-00092-2","type":"journal-article","created":{"date-parts":[[2023,12,12]],"date-time":"2023-12-12T18:04:16Z","timestamp":1702404256000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":11,"title":["In humans, we trust"],"prefix":"10.1007","volume":"3","author":[{"given":"Kyriakos","family":"Kyriakou","sequence":"first","affiliation":[]},{"given":"Jahna","family":"Otterbacher","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,12,12]]},"reference":[{"issue":"1","key":"92_CR1","doi-asserted-by":"publisher","first-page":"5","DOI":"10.1007\/s10676-017-9430-8","volume":"20","author":"I Rahwan","year":"2018","unstructured":"Rahwan I. Society-in-the-loop: programming the algorithmic social contract. Ethics Inf Technol. 2018;20(1):5\u201314. https:\/\/doi.org\/10.1007\/s10676-017-9430-8.","journal-title":"Ethics Inf Technol"},{"issue":"1","key":"92_CR2","doi-asserted-by":"publisher","first-page":"4","DOI":"10.1016\/j.intcom.2010.07.003","volume":"23","author":"G Baxter","year":"2011","unstructured":"Baxter G, Sommerville I. Socio-technical systems: from design methods to systems engineering. Interact Comput. 2011;23(1):4\u201317. https:\/\/doi.org\/10.1016\/j.intcom.2010.07.003.","journal-title":"Interact Comput"},{"key":"92_CR3","doi-asserted-by":"publisher","DOI":"10.1145\/3236009","author":"R Guidotti","year":"2018","unstructured":"Guidotti R, Monreale A, Ruggieri S, Turini F, Giannotti F, Pedreschi D. A survey of methods for explaining black box models. ACM Comput Surv. 2018. https:\/\/doi.org\/10.1145\/3236009.","journal-title":"ACM Comput Surv"},{"issue":"3","key":"92_CR4","doi-asserted-by":"publisher","first-page":"252","DOI":"10.1108\/JICES-06-2018-0056","volume":"16","author":"NT Lee","year":"2018","unstructured":"Lee NT. Detecting racial bias in algorithms and machine learning. J Inf Commun Ethics Soc. 2018;16(3):252\u201360.","journal-title":"J Inf Commun Ethics Soc"},{"key":"92_CR5","unstructured":"Kodiyan AA. An overview of ethical issues in using AI systems in hiring with a case study of Amazon\u2019s AI based hiring tool. Researchgate Preprint, 2019;1\u201319."},{"key":"92_CR6","first-page":"131","volume":"17","author":"AL Washington","year":"2018","unstructured":"Washington AL. How to argue with an algorithm: lessons from the compas-propublica debate. Colo Tech LJ. 2018;17:131.","journal-title":"Colo Tech LJ"},{"key":"92_CR7","doi-asserted-by":"publisher","DOI":"10.1145\/3527152","author":"K Orphanou","year":"2022","unstructured":"Orphanou K, Otterbacher J, Kleanthous S, Batsuren K, Giunchiglia F, Bogina V, Tal AS, Hartman A, Kuflik T. Mitigating bias in algorithmic systems\u2014a fish-eye view. ACM Comput Surv. 2022. https:\/\/doi.org\/10.1145\/3527152.","journal-title":"ACM Comput Surv."},{"key":"92_CR8","unstructured":"Buolamwini J, Gebru T. Gender shades: intersectional accuracy disparities in commercial gender classification. In: Friedler SA, Wilson C. editors. Proceedings of the 1st conference on fairness, accountability and transparency. Proceedings of Machine Learning Research, 81, PMLR, 2018. p. 77\u201391. https:\/\/proceedings.mlr.press\/v81\/buolamwini18a.html."},{"issue":"01","key":"92_CR9","doi-asserted-by":"publisher","first-page":"313","DOI":"10.1609\/icwsm.v13i01.3232","volume":"13","author":"K Kyriakou","year":"2019","unstructured":"Kyriakou K, Barlas P, Kleanthous S, Otterbacher J. Fairness in proprietary image tagging algorithms: a cross-platform audit on people images. Proc Int AAAI Conf Web Soc Media. 2019;13(01):313\u201322.","journal-title":"Proc Int AAAI Conf Web Soc Media"},{"key":"92_CR10","doi-asserted-by":"publisher","unstructured":"Kyriakou K, Kleanthous S, Otterbacher J, Papadopoulos GA. Emotion-based stereotypes in image analysis services. In: Adjunct Publication of the 28th ACM Conference on User Modeling, Adaptation and Personalization. UMAP \u201920 Adjunct. Association for Computing Machinery, New York, NY, USA: 2020. p. 252\u2013259. https:\/\/doi.org\/10.1145\/3386392.3399567. event-place: Genoa, Italy.","DOI":"10.1145\/3386392.3399567"},{"issue":"1","key":"92_CR11","first-page":"32","volume":"7","author":"P Hase","year":"2019","unstructured":"Hase P, Chen C, Li O, Rudin C. Interpretable image recognition with hierarchical prototypes. Proc AAAI Conf Hum Comput Crowdsourc. 2019;7(1):32\u201340.","journal-title":"Proc AAAI Conf Hum Comput Crowdsourc"},{"key":"92_CR12","doi-asserted-by":"crossref","unstructured":"Page MJ, Moher D, Bossuyt PM, Boutron I, Hoffmann TC, Mulrow CD, Shamseer L, Tetzlaff JM, Akl EA, Brennan SE, et al. Prisma 2020 explanation and elaboration: updated guidance and exemplars for reporting systematic reviews. BMJ. 2021;372.","DOI":"10.1136\/bmj.n160"},{"key":"92_CR13","doi-asserted-by":"crossref","unstructured":"Tal AS, Batsuren K, Bogina V, Giunchiglia F, Hartman A, Loizou SK, Kuflik T, Otterbacher J. \u201cend to end\u201d towards a framework for reducing biases and promoting transparency of algorithmic systems. In: 2019 14th International Workshop on Semantic and Social Media Adaptation and Personalization (SMAP), 2019. p. 1\u20136. IEEE.","DOI":"10.1109\/SMAP.2019.8864914"},{"key":"92_CR14","doi-asserted-by":"publisher","unstructured":"Bandy, J. Problematic machine behavior: a systematic literature review of algorithm audits. Proc. ACM Hum.-Comput. Interact. 5(CSCW1). New York, NY: Association for Computing. Machinery. 2021. https:\/\/doi.org\/10.1145\/3449148.","DOI":"10.1145\/3449148"},{"issue":"3","key":"92_CR15","doi-asserted-by":"publisher","first-page":"388","DOI":"10.1080\/0960085X.2021.1927212","volume":"31","author":"N Kordzadeh","year":"2022","unstructured":"Kordzadeh N, Ghasemaghaei M. Algorithmic bias: review, synthesis, and future research directions. Eur J Inf Syst. 2022;31(3):388\u2013409.","journal-title":"Eur J Inf Syst"},{"key":"92_CR16","doi-asserted-by":"publisher","unstructured":"Wieringa M. What to account for when accounting for algorithms: a systematic literature review on algorithmic accountability. In: Proceedings of the 2020 Conference on fairness, accountability, and transparency. FAT* \u201920. Association for Computing Machinery, New York, NY, USA: 2020. p. 1\u201318. https:\/\/doi.org\/10.1145\/3351095.3372833.","DOI":"10.1145\/3351095.3372833"},{"key":"92_CR17","doi-asserted-by":"publisher","first-page":"141","DOI":"10.1146\/annurev-statistics-042720-125902","volume":"8","author":"S Mitchell","year":"2021","unstructured":"Mitchell S, Potash E, Barocas S, D\u2019Amour A, Lum K. Algorithmic fairness: choices, assumptions, and definitions. Annu Rev Stat Its Appl. 2021;8:141\u201363.","journal-title":"Annu Rev Stat Its Appl"},{"key":"92_CR18","doi-asserted-by":"crossref","unstructured":"Xivuri K, Twinomurinzi H. A systematic review of fairness in artificial intelligence algorithms. In: Responsible AI and analytics for an ethical and inclusive digitized society: 20th IFIP WG 6.11 Conference on e-Business, e-Services and e-Society, I3E 2021, Galway, Ireland, September 1\u20133, 2021, Proceedings 20, Springer: 2021. p. 271\u2013284.","DOI":"10.1007\/978-3-030-85447-8_24"},{"key":"92_CR19","doi-asserted-by":"publisher","first-page":"82","DOI":"10.1016\/j.inffus.2019.12.012","volume":"58","author":"AB Arrieta","year":"2020","unstructured":"Arrieta AB, D\u00edaz-Rodr\u00edguez N, Del Ser J, Bennetot A, Tabik S, Barbado A, Garc\u00eda S, Gil-L\u00f3pez S, Molina D, Benjamins R, et al. Explainable artificial intelligence (XAI): concepts, taxonomies, opportunities and challenges toward responsible AI. Inf Fusion. 2020;58:82\u2013115.","journal-title":"Inf Fusion"},{"issue":"6","key":"92_CR20","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3457607","volume":"54","author":"N Mehrabi","year":"2021","unstructured":"Mehrabi N, Morstatter F, Saxena N, Lerman K, Galstyan A. A survey on bias and fairness in machine learning. ACM Comput Surv (CSUR). 2021;54(6):1\u201335.","journal-title":"ACM Comput Surv (CSUR)"},{"key":"92_CR21","doi-asserted-by":"publisher","first-page":"13","DOI":"10.3389\/fdata.2019.00013","volume":"2","author":"A Olteanu","year":"2019","unstructured":"Olteanu A, Castillo C, Diaz F, K\u0131c\u0131man E. Social data: biases, methodological pitfalls, and ethical boundaries. Front Big Data. 2019;2:13.","journal-title":"Front Big Data"},{"issue":"3","key":"92_CR22","doi-asserted-by":"publisher","first-page":"1356","DOI":"10.1002\/widm.1356","volume":"10","author":"E Ntoutsi","year":"2020","unstructured":"Ntoutsi E, Fafalios P, Gadiraju U, Iosifidis V, Nejdl W, Vidal M-E, Ruggieri S, Turini F, Papadopoulos S, Krasanakis E, et al. Bias in data-driven artificial intelligence systems\u2014an introductory survey. Wiley Interdiscipl Rev Data Mining Knowl Discov. 2020;10(3):1356.","journal-title":"Wiley Interdiscipl Rev Data Mining Knowl Discov"},{"key":"92_CR23","first-page":"105","volume-title":"Adv Netw-Based Inf Syst","author":"D Kaur","year":"2021","unstructured":"Kaur D, Uslu S, Durresi A. Requirements for trustworthy artificial intelligence\u2014a review. In: Barolli L, Li KF, Enokido T, Takizawa M, editors. Adv Netw-Based Inf Syst. Cham: Springer; 2021. p. 105\u201315."},{"issue":"2","key":"92_CR24","doi-asserted-by":"publisher","first-page":"205395171667967","DOI":"10.1177\/2053951716679679","volume":"3","author":"BD Mittelstadt","year":"2016","unstructured":"Mittelstadt BD, Allo P, Taddeo M, Wachter S, Floridi L. The ethics of algorithms: mapping the debate. Big Data Soc. 2016;3(2):2053951716679679. https:\/\/doi.org\/10.1177\/2053951716679679.","journal-title":"Big Data Soc"},{"issue":"4","key":"92_CR25","doi-asserted-by":"publisher","first-page":"689","DOI":"10.1007\/s11023-018-9482-5","volume":"28","author":"L Floridi","year":"2018","unstructured":"Floridi L, Cowls J, Beltrametti M, Chatila R, Chazerand P, Dignum V, Luetge C, Madelin R, Pagallo U, Rossi F, Schafer B, Valcke P, Vayena E. AI4People\u2014an ethical framework for a good AI society: opportunities, risks, principles, and recommendations. Mind Mach. 2018;28(4):689\u2013707. https:\/\/doi.org\/10.1007\/s11023-018-9482-5.","journal-title":"Mind Mach"},{"issue":"6404","key":"92_CR26","doi-asserted-by":"publisher","first-page":"751","DOI":"10.1126\/science.aat5991","volume":"361","author":"M Taddeo","year":"2018","unstructured":"Taddeo M, Floridi L. How AI can be a force for good. Science. 2018;361(6404):751\u20132. https:\/\/doi.org\/10.1126\/science.aat5991.","journal-title":"Science"},{"key":"92_CR27","doi-asserted-by":"publisher","unstructured":"Jacobs AZ, Wallach H. Measurement and fairness. In: Proceedings of the 2021 ACM conference on fairness, accountability, and transparency. FAccT \u201921. Association for Computing Machinery, New York, NY, USA: 2021. p. 375\u2013385. https:\/\/doi.org\/10.1145\/3442188.3445901. event-place: Virtual Event, Canada.","DOI":"10.1145\/3442188.3445901"}],"container-title":["Discover Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44163-023-00092-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s44163-023-00092-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44163-023-00092-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,12,12]],"date-time":"2023-12-12T18:04:49Z","timestamp":1702404289000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s44163-023-00092-2"}},"subtitle":["Multidisciplinary perspectives on the requirements for human oversight in algorithmic processes"],"short-title":[],"issued":{"date-parts":[[2023,12,12]]},"references-count":27,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2023,12]]}},"alternative-id":["92"],"URL":"https:\/\/doi.org\/10.1007\/s44163-023-00092-2","relation":{},"ISSN":["2731-0809"],"issn-type":[{"value":"2731-0809","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,12,12]]},"assertion":[{"value":"19 July 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 November 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 December 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"44"}}