{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,6]],"date-time":"2026-05-06T08:30:06Z","timestamp":1778056206173,"version":"3.51.4"},"reference-count":64,"publisher":"Walter de Gruyter GmbH","issue":"9","license":[{"start":{"date-parts":[[2022,9,1]],"date-time":"2022-09-01T00:00:00Z","timestamp":1661990400000},"content-version":"unspecified","delay-in-days":0,"URL":"http:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100016377","name":"Ministerium f\u00fcr Wirtschaft, Innovation, Digitalisierung und Energie des Landes Nordrhein-Westfalen","doi-asserted-by":"publisher","award":["005-2011-0048"],"award-info":[{"award-number":["005-2011-0048"]}],"id":[{"id":"10.13039\/501100016377","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022,9,27]]},"abstract":"<jats:title>Abstract<\/jats:title>\n               <jats:p>Artificial intelligence is increasingly penetrating industrial applications as well as areas that affect our daily lives. As a consequence, there is a need for criteria to validate whether the quality of AI applications is sufficient for their intended use. Both in the academic community and societal debate, an agreement has emerged under the term \u201ctrustworthiness\u201d as the set of essential quality requirements that should be placed on an AI application. At the same time, the question of how these quality requirements can be operationalized is to a large extent still open.<\/jats:p>\n               <jats:p>In this paper, we consider trustworthy AI from two perspectives: the product and organizational perspective. For the former, we present an AI-specific risk analysis and outline how verifiable arguments for the trustworthiness of an AI application can be developed. For the second perspective, we explore how an AI management system can be employed to assure the trustworthiness of an organization with respect to its handling of AI. Finally, we argue that in order to achieve AI trustworthiness, coordinated measures from both product and organizational perspectives are required.<\/jats:p>","DOI":"10.1515\/auto-2022-0012","type":"journal-article","created":{"date-parts":[[2022,9,2]],"date-time":"2022-09-02T07:34:11Z","timestamp":1662104051000},"page":"793-804","source":"Crossref","is-referenced-by-count":14,"title":["The why and how of trustworthy AI"],"prefix":"10.1515","volume":"70","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8801-3700","authenticated-orcid":false,"given":"Anna","family":"Schmitz","sequence":"first","affiliation":[{"name":"Fraunhofer-Institut f\u00fcr Intelligente Analyse- und Informationssysteme IAIS , Sankt Augustin , Germany"}]},{"given":"Maram","family":"Akila","sequence":"additional","affiliation":[{"name":"Fraunhofer-Institut f\u00fcr Intelligente Analyse- und Informationssysteme IAIS , Sankt Augustin , Germany"},{"name":"ML2R \u2013 Kompetenzzentrum Maschinelles Lernen Rhein-Ruhr , Dortmund , Germany"}]},{"given":"Dirk","family":"Hecker","sequence":"additional","affiliation":[{"name":"Fraunhofer-Institut f\u00fcr Intelligente Analyse- und Informationssysteme IAIS , Sankt Augustin , Germany"},{"name":"ML2R \u2013 Kompetenzzentrum Maschinelles Lernen Rhein-Ruhr , Dortmund , Germany"}]},{"given":"Maximilian","family":"Poretschkin","sequence":"additional","affiliation":[{"name":"Fraunhofer-Institut f\u00fcr Intelligente Analyse- und Informationssysteme IAIS , Sankt Augustin , Germany"},{"name":"ML2R \u2013 Kompetenzzentrum Maschinelles Lernen Rhein-Ruhr , Dortmund , Germany"},{"name":"Institut f\u00fcr Informatik III , Rheinische-Friedrich-Wilhelms-Universit\u00e4t Bonn , Bonn , Germany"}]},{"given":"Stefan","family":"Wrobel","sequence":"additional","affiliation":[{"name":"Fraunhofer-Institut f\u00fcr Intelligente Analyse- und Informationssysteme IAIS , Sankt Augustin , Germany"},{"name":"ML2R \u2013 Kompetenzzentrum Maschinelles Lernen Rhein-Ruhr , Dortmund , Germany"},{"name":"Institut f\u00fcr Informatik III , Rheinische-Friedrich-Wilhelms-Universit\u00e4t Bonn , Bonn , Germany"}]}],"member":"374","published-online":{"date-parts":[[2022,9,3]]},"reference":[{"key":"2023033111053262424_j_auto-2022-0012_ref_001","unstructured":"Angwin, J., J. Larson, S. Mattu and L. Kirchner. 2016. Machine bias: There\u2019s software used across the country to predict future criminals. And it\u2019s biased against blacks. ProPublica."},{"key":"2023033111053262424_j_auto-2022-0012_ref_002","unstructured":"Bogen, M. and A. Rieke. 2018. Help Wanted: An Examination of Hiring Algorithms, Equity and Bias. Technical Report. Upturn."},{"key":"2023033111053262424_j_auto-2022-0012_ref_003","doi-asserted-by":"crossref","unstructured":"Bouti, A. and D.\u2009A. Kadi. 1994. A state-of-the-art review of FMEA\/FMECA. International Journal of reliability, quality and safety engineering, 1(04): 515\u2013543.","DOI":"10.1142\/S0218539394000362"},{"key":"2023033111053262424_j_auto-2022-0012_ref_004","unstructured":"Brown, T.\u2009B., B. Mann, N. Ryder, M. Subbiah, J. Kaplan, P. Dhariwal, et\u2009al. 2020. Language models are few-shot learners. arXiv preprint arXiv:2005.14165."},{"key":"2023033111053262424_j_auto-2022-0012_ref_005","unstructured":"Brundage, M., S. Avin, J. Wang, H. Belfield, G. Krueger, G. Hadfield, et\u2009al. 2020. Toward trustworthy AI development: mechanisms for supporting verifiable claims. arXiv preprint arXiv:2004.07213."},{"key":"2023033111053262424_j_auto-2022-0012_ref_006","unstructured":"Buolamwini, J. and T. Gebru. 2018. Gender shades: Intersectional accuracy disparities in commercial gender classification. Conference on fairness, accountability and transparency (pp.\u200977\u201391). PMLR."},{"key":"2023033111053262424_j_auto-2022-0012_ref_007","doi-asserted-by":"crossref","unstructured":"Burrell, J. 2016. How the machine \u2018thinks\u2019: Understanding opacity in machine learning algorithms. Big Data & Society, 3(1): 2053951715622512.","DOI":"10.1177\/2053951715622512"},{"key":"2023033111053262424_j_auto-2022-0012_ref_008","doi-asserted-by":"crossref","unstructured":"\u00c7\u0131nar, Z.\u2009M., A. Abdussalam Nuhu, Q. Zeeshan, O. Korhan, M. Asmael and B. Safaei. 2020. Machine learning in predictive maintenance towards sustainable smart manufacturing in industry 4.0. Sustainability, 12(19): 8211.","DOI":"10.3390\/su12198211"},{"key":"2023033111053262424_j_auto-2022-0012_ref_009","unstructured":"Crawford, K., R. Dobbe, T. Dryer, G. Fried, B. Green, et al. 2019. \u200bAI Now 2019 Report. \u200bNew York: AI Now Institute."},{"key":"2023033111053262424_j_auto-2022-0012_ref_010","unstructured":"Cremers, A.\u2009B., A. Englander, M. Gabriel, D. Hecker, M. Mock, et al. 2019. Trustworthy use of AI. Priorities from a philosophical, ethical, legal and technological viewpoint as a basis for certification of Artificial Intelligence. Fraunhofer-Institut f\u00fcr Intelligente Analyse und Informationssysteme IAIS, Sankt Augustin."},{"key":"2023033111053262424_j_auto-2022-0012_ref_011","unstructured":"DIN e. V. and DKE 2020. Deutsche Normungsroadmap K\u00fcnstliche Intelligenz."},{"key":"2023033111053262424_j_auto-2022-0012_ref_012","doi-asserted-by":"crossref","unstructured":"Druzhkov, P.\u2009N. and V.\u2009D. Kustikova. 2016. A survey of deep learning methods and software tools for image classification and object detection. Pattern Recognition and Image Analysis, 26(1): 9\u201315.","DOI":"10.1134\/S1054661816010065"},{"key":"2023033111053262424_j_auto-2022-0012_ref_013","unstructured":"Dvijotham, K., R. Stanforth, S. Gowal, T.\u2009A. Mann and P. Kohli. 2018. A Dual Approach to Scalable Verification of Deep Networks. In UAI (Vol.\u20091, No.\u20092, p.\u20093)."},{"key":"2023033111053262424_j_auto-2022-0012_ref_014","unstructured":"European Commission. 2018. Communication from the Commission, Artificial Intelligence for Europe (COM\/2018\/237 final)."},{"key":"2023033111053262424_j_auto-2022-0012_ref_015","unstructured":"European Commission. 2021. Proposal for a Regulation of the European Parliament and of the Council laying down Harmonized Rules on Artificial Intelligence (Artificial Intelligence Act) and amending certain Union Legislative Acts."},{"key":"2023033111053262424_j_auto-2022-0012_ref_016","doi-asserted-by":"crossref","unstructured":"Floridi, L. and M. Taddeo. 2016. What is data ethics? Philosophical Transactions of the Royal Society: A.3742016036020160360. http:\/\/doi.org\/10.1098\/rsta.2016.0360.","DOI":"10.1098\/rsta.2016.0360"},{"key":"2023033111053262424_j_auto-2022-0012_ref_017","doi-asserted-by":"crossref","unstructured":"Floridi, L., J. Cowls, M. Beltrametti, R. Chatila, P. Chazerand, et al. 2018. AI4People\u2014an ethical framework for a good AI society: opportunities, risks, principles, and recommendations. Minds and Machines, 28(4): 689\u2013707.","DOI":"10.1007\/s11023-018-9482-5"},{"key":"2023033111053262424_j_auto-2022-0012_ref_018","doi-asserted-by":"crossref","unstructured":"Fredrikson, M., S. Jha and T. Ristenpart. 2015. Model inversion attacks that exploit confidence information and basic countermeasures. Proceedings of the 22nd ACM SIGSAC Conference on Computer and Communications Security. ACM, 2015, pp.\u20091322\u20131333.","DOI":"10.1145\/2810103.2813677"},{"key":"2023033111053262424_j_auto-2022-0012_ref_019","doi-asserted-by":"crossref","unstructured":"Gladisch, C., C. Heinzemann, M. Herrmann and M. Woehrle. 2020. Leveraging Combinatorial Testing for Safety-Critical Computer Vision Datasets. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, pp.\u2009324\u2013325.","DOI":"10.1109\/CVPRW50498.2020.00170"},{"key":"2023033111053262424_j_auto-2022-0012_ref_020","doi-asserted-by":"crossref","unstructured":"Guegan, D. and B. Hassani. 2018. Regulatory learning: How to supervise machine learning models? An application to credit scoring. The Journal of Finance and Data Science, 4(3): 157\u2013171.","DOI":"10.1016\/j.jfds.2018.04.001"},{"key":"2023033111053262424_j_auto-2022-0012_ref_021","doi-asserted-by":"crossref","unstructured":"Hagendorff, T. 2020. The ethics of AI ethics: An evaluation of guidelines. Minds and Machines, 30(1): 99\u2013120.","DOI":"10.1007\/s11023-020-09517-8"},{"key":"2023033111053262424_j_auto-2022-0012_ref_022","unstructured":"Hendrycks, D., N. Carlini, J. Schulman, and J. Steinhardt. 2021. Unsolved problems in ml safety. arXiv preprint arXiv:2109.13916."},{"key":"2023033111053262424_j_auto-2022-0012_ref_023","unstructured":"High-level Expert Group on AI. 2019. Ethics Guidelines on Trustworthy AI. European Commission."},{"key":"2023033111053262424_j_auto-2022-0012_ref_024","doi-asserted-by":"crossref","unstructured":"Houben, S., S. Abrecht, M. Akila, A. B\u00e4r, Brockherde, F., et\u2009al. 2021. Inspect, understand, overcome: A survey of practical methods for ai safety. arXiv preprint arXiv:2104.14235.","DOI":"10.1007\/978-3-031-01233-4_1"},{"key":"2023033111053262424_j_auto-2022-0012_ref_025","unstructured":"International Organization for Standardization. Standard ISO\/IEC CD 23894. Information Technology \u2013 Artificial intelligence \u2013 Risk Management, under development."},{"key":"2023033111053262424_j_auto-2022-0012_ref_026","unstructured":"International Organization for Standardization. Standard ISO\/IEC CD 42001. Information Technology \u2013 Artificial intelligence \u2013 Management system, under development."},{"key":"2023033111053262424_j_auto-2022-0012_ref_027","unstructured":"International Organization for Standardization. 2013. Standard ISO\/IEC 27001:2013. Information technology \u2013 Security techniques \u2013 Information security management systems \u2014 Requirements."},{"key":"2023033111053262424_j_auto-2022-0012_ref_028","unstructured":"International Organization for Standardization. 2015. Standard ISO 9001:2015. Quality management systems \u2013 Requirements."},{"key":"2023033111053262424_j_auto-2022-0012_ref_029","unstructured":"International Organization for Standardization. 2015. Standard ISO 14001:2015. Environmental management systems \u2014 Requirements with guidance for use."},{"key":"2023033111053262424_j_auto-2022-0012_ref_030","unstructured":"International Organization for Standardization. 2015. Standard ISO\/IEC 15408-1:2009. Information technology \u2014 Security techniques \u2014 Evaluation criteria for IT security \u2014 Part 1: Introduction and general model."},{"key":"2023033111053262424_j_auto-2022-0012_ref_031","unstructured":"International Organization for Standardization. 2018. Standard ISO 31000:2018. Risk management \u2013 Guidelines."},{"key":"2023033111053262424_j_auto-2022-0012_ref_032","unstructured":"ISO. 2021. The ISO Survey of Management System Standard Certifications \u2013 2020 \u2013 Explanatory Note."},{"key":"2023033111053262424_j_auto-2022-0012_ref_033","unstructured":"ISO. Management System Standards, n. d. URL: https:\/\/www.iso.org\/management-system-standards.html (Accessed on 11.01.2022)."},{"key":"2023033111053262424_j_auto-2022-0012_ref_034","doi-asserted-by":"crossref","unstructured":"Jobin, A., M. Iencaand E. Vayena. 2019. The global landscape of AI ethics guidelines. Nat. Mach. Intell. 1: 389\u2013399.","DOI":"10.1038\/s42256-019-0088-2"},{"key":"2023033111053262424_j_auto-2022-0012_ref_035","doi-asserted-by":"crossref","unstructured":"Katz, G., C. Barrett, D.\u2009L. Dill, K. Julian and M.\u2009J. Kochenderfer. 2017. Reluplex: An efficient SMT solver for verifying deep neural networks. In International conference on computer aided verification, pp.\u200997\u2013117. Springer, Cham.","DOI":"10.1007\/978-3-319-63387-9_5"},{"key":"2023033111053262424_j_auto-2022-0012_ref_036","doi-asserted-by":"crossref","unstructured":"Kersten, H., G. Klett, J. Reuter and K.-W. Schr\u00f6der. 2020. IT-Sicherheitsmanagement nach der neuen ISO 27001. Springer Vieweg. ISBN 978-3-658-27691-1.","DOI":"10.1007\/978-3-658-27692-8"},{"key":"2023033111053262424_j_auto-2022-0012_ref_037","doi-asserted-by":"crossref","unstructured":"Killourhy, K.\u2009S. and R.\u2009A. Maxion. 2009. Comparing anomaly-detection algorithms for keystroke dynamics. 2009 IEEE\/IFIP International Conference on Dependable Systems & Networks, pp.\u2009125\u2013134.","DOI":"10.1109\/DSN.2009.5270346"},{"key":"2023033111053262424_j_auto-2022-0012_ref_038","unstructured":"Lambert, G. 2017. A stroll down Quality Street. ISOfocus 123 July-August 2017. pp.\u200937\u201340."},{"key":"2023033111053262424_j_auto-2022-0012_ref_039","doi-asserted-by":"crossref","unstructured":"Li, Z., X. Ma, C. Xu, and C. Cao. 2019. Structural coverage criteria for neural networks could be misleading. 2019 IEEE\/ACM 41st International Conference on Software Engineering: New Ideas and Emerging Results, pp.\u200989\u201392.","DOI":"10.1109\/ICSE-NIER.2019.00031"},{"key":"2023033111053262424_j_auto-2022-0012_ref_040","doi-asserted-by":"crossref","unstructured":"Liu, C., T. Arnon, C. Lazarus, C. Strong, C. Barrett and M.\u2009J. Kochenderfer. 2021. Algorithms for verifying deep neural networks. Foundations and Trends\u00ae in Optimization, 4(3\u20134): 244\u2013404.","DOI":"10.1561\/2400000035"},{"key":"2023033111053262424_j_auto-2022-0012_ref_041","doi-asserted-by":"crossref","unstructured":"Mata, F.\u2009J., W.\u2009L. Fuerst and J.\u2009B. Barney. 1995. Information technology and sustained competitive advantage: A resource-based analysis. MIS quarterly, 487\u2013505.","DOI":"10.2307\/249630"},{"key":"2023033111053262424_j_auto-2022-0012_ref_042","doi-asserted-by":"crossref","unstructured":"Mehrabi, N., F. Morstatter, N. Saxena, K. Lerman and A. Galstyan. 2021. A survey on bias and fairness in machine learning. ACM Computing Surveys (CSUR), 54(6): 1\u201335.","DOI":"10.1145\/3457607"},{"key":"2023033111053262424_j_auto-2022-0012_ref_043","doi-asserted-by":"crossref","unstructured":"Mittelstadt, B. 2019. Principles alone cannot guarantee ethical AI. Nature Machine Intelligence, 1(11): 501\u2013507.","DOI":"10.1038\/s42256-019-0114-4"},{"key":"2023033111053262424_j_auto-2022-0012_ref_044","unstructured":"Mock, M., A. Schmitz, L. Adilova, D. Becker, A.\u2009B. Cremers and M. Poretschkin. 2021. Management System Support for Trustworthy Artificial Intelligence. Fraunhofer-Institut f\u00fcr Intelligente Analyse und Informationssysteme. IAIS, Sankt Augustin."},{"key":"2023033111053262424_j_auto-2022-0012_ref_045","doi-asserted-by":"crossref","unstructured":"Morley, J., L. Floridi, L. Kinsey and A. Elhalal. 2021. From what to how: an initial review of publicly available AI ethics tools, methods and research to translate principles into practices. Ethics, Governance, and Policies in Artificial Intelligence. pp.\u2009153\u2013183. Springer, Cham.","DOI":"10.1007\/978-3-030-81907-1_10"},{"key":"2023033111053262424_j_auto-2022-0012_ref_046","unstructured":"OECD Council. 2019. Recommendation of the Council on Artificial Intelligence (OECD\/LEGAL\/0449)."},{"key":"2023033111053262424_j_auto-2022-0012_ref_047","unstructured":"OECD. AI, powered by EC OECD. 2021. Database of national AI policies. https:\/\/oecd.ai (Accessed on 29.03.2022)."},{"key":"2023033111053262424_j_auto-2022-0012_ref_048","unstructured":"Papernot, N., P. McDaniel, A. Sinha and M. Wellman. 2016. Towards the science of security and privacy in machine learning. arXiv preprint arXiv:1611.03814."},{"key":"2023033111053262424_j_auto-2022-0012_ref_049","unstructured":"Poretschkin, M., A. Schmitz, L. Adilova, M. Akila, D. Becker, et al. 2021. KI-Pr\u00fcfkatalog: Leitfaden zur Gestaltung vertrauensw\u00fcrdiger K\u00fcnstlicher Intelligenz. Fraunhofer-Institut f\u00fcr Intelligente Analyse und Informationssysteme IAIS, Sankt Augustin."},{"key":"2023033111053262424_j_auto-2022-0012_ref_050","doi-asserted-by":"crossref","unstructured":"Poretschkin, M., M. Mock and S. Wrobel. 2021. Zur Systematischen Bewertung der Vertrauensw\u00fcrdigkeit von KI-Systemen. D. Zimmer (Hrsg.), Regulierung f\u00fcr Algorithmen und K\u00fcnstliche Intelligenz.","DOI":"10.5771\/9783748927990-175"},{"key":"2023033111053262424_j_auto-2022-0012_ref_051","unstructured":"Rostalski, F., and E. Weiss. 2021. Der KI-Verordnungsentwurf der Europ\u00e4ischen Kommission \u2013 Eine erste Analyse unter besonderer Ber\u00fccksichtigung der Rolle von Zertifizierung. Zeitschrift f\u00fcr Digitalisierung und Recht, 4\/2021."},{"key":"2023033111053262424_j_auto-2022-0012_ref_052","unstructured":"Department of International Cooperation Ministry of Science and Technology (MOST), P.R.China. 2017. Next Generation Artificial Intelligence Development Plan. China Science & Technology Newsletter No.\u200917, Issued by State Council."},{"key":"2023033111053262424_j_auto-2022-0012_ref_053","doi-asserted-by":"crossref","unstructured":"Sharma, R. and P. Yetton. 2003. The contingent effects of management support and task interdependence on successful information systems implementation. MIS quarterly, 533\u2013556.","DOI":"10.2307\/30036548"},{"key":"2023033111053262424_j_auto-2022-0012_ref_054","doi-asserted-by":"crossref","unstructured":"Toreini, E., M. Aitken, K. Coopamootoo, K. Elliott, C.\u2009G. Zelaya and A. Van Moorsel. 2020. The relationship between trust in AI and trustworthy machine learning technologies. Proceedings of the 2020 conference on fairness, accountability, and transparency, pp.\u2009272\u2013283.","DOI":"10.1145\/3351095.3372834"},{"key":"2023033111053262424_j_auto-2022-0012_ref_055","unstructured":"Select Committee on Artificial Intelligence of the National Science & Technology Council. 2019. The National Artificial Intelligence Research and Development Strategic Plan: 2019 Update."},{"key":"2023033111053262424_j_auto-2022-0012_ref_056","doi-asserted-by":"crossref","unstructured":"Verma, S. and J. Rubin. 2018. Fairness definitions explained. 2018 IEEE\/ACM international workshop on software fairness (fairware), pp.\u20091\u20137.","DOI":"10.1145\/3194770.3194776"},{"key":"2023033111053262424_j_auto-2022-0012_ref_057","doi-asserted-by":"crossref","unstructured":"Wachter, S., B. Mittelstadt and C. Russell. 2020. Bias preservation in machine learning: the legality of fairness metrics under EU non-discrimination law. West Virginia Law Review, 123: 735.","DOI":"10.2139\/ssrn.3792772"},{"key":"2023033111053262424_j_auto-2022-0012_ref_058","unstructured":"Whitman, M.\u2009E. and H.\u2009J. Mattord. 2021. Principles of information security. Cengage learning."},{"key":"2023033111053262424_j_auto-2022-0012_ref_059","doi-asserted-by":"crossref","unstructured":"Yu, K.\u2009H., A.\u2009L. Beam and I.\u2009S. Kohane. 2018. Artificial intelligence in healthcare. Nature biomedical engineering, 2(10): 719\u2013731.","DOI":"10.1038\/s41551-018-0305-z"},{"key":"2023033111053262424_j_auto-2022-0012_ref_060","doi-asserted-by":"crossref","unstructured":"Yurtsever, E., J. Lambert, A. Carballo and K. Takeda. 2020. A survey of autonomous driving: Common practices and emerging technologies. IEEE access, 8: 58443\u201358469.","DOI":"10.1109\/ACCESS.2020.2983149"},{"key":"2023033111053262424_j_auto-2022-0012_ref_061","unstructured":"Zhang, H., T.\u2009W. Weng, P.\u2009Y. Chen, C.\u2009J. Hsieh and L. Daniel. 2018. Efficient neural network robustness certification with general activation functions. Advances in neural information processing systems 31."},{"key":"2023033111053262424_j_auto-2022-0012_ref_062","unstructured":"Zhang, J.\u2009M., M. Harman, L. Ma and Y. Liu. 2020. Machine learning testing: Survey, landscapes and horizons. IEEE Transactions on Software Engineering."},{"key":"2023033111053262424_j_auto-2022-0012_ref_063","doi-asserted-by":"crossref","unstructured":"Zhang, Y., P. Ti\u0148o, A. Leonardis and K. Tang. 2021. A survey on neural network interpretability. IEEE Transactions on Emerging Topics in Computational Intelligence.","DOI":"10.1109\/TETCI.2021.3100641"},{"key":"2023033111053262424_j_auto-2022-0012_ref_064","doi-asserted-by":"crossref","unstructured":"Zhou, J., F. Chen, A. Berry, M. Reed, S. Zhang and S. Savage. 2020. A Survey on Ethical Principles of AI and Implementations. 2020 IEEE Symposium Series on Computational Intelligence (SSCI), 3010\u20133017.","DOI":"10.1109\/SSCI47803.2020.9308437"}],"container-title":["at - Automatisierungstechnik"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.degruyter.com\/document\/doi\/10.1515\/auto-2022-0012\/xml","content-type":"application\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.degruyter.com\/document\/doi\/10.1515\/auto-2022-0012\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,3,31]],"date-time":"2023-03-31T13:54:52Z","timestamp":1680270892000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.degruyter.com\/document\/doi\/10.1515\/auto-2022-0012\/html"}},"subtitle":["An approach for systematic quality assurance when working with ML components"],"short-title":[],"issued":{"date-parts":[[2022,9,1]]},"references-count":64,"journal-issue":{"issue":"9","published-online":{"date-parts":[[2022,9,3]]},"published-print":{"date-parts":[[2022,9,27]]}},"alternative-id":["10.1515\/auto-2022-0012"],"URL":"https:\/\/doi.org\/10.1515\/auto-2022-0012","relation":{},"ISSN":["2196-677X","0178-2312"],"issn-type":[{"value":"2196-677X","type":"electronic"},{"value":"0178-2312","type":"print"}],"subject":[],"published":{"date-parts":[[2022,9,1]]}}}