{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T05:01:00Z","timestamp":1750309260895,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":36,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,6,3]],"date-time":"2024-06-03T00:00:00Z","timestamp":1717372800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,6,3]]},"DOI":"10.1145\/3630106.3659035","type":"proceedings-article","created":{"date-parts":[[2024,6,5]],"date-time":"2024-06-05T13:14:21Z","timestamp":1717593261000},"page":"2222-2233","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Trust Issues: Discrepancies in Trustworthy AI Keywords Use in Policy and Research"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3247-1248","authenticated-orcid":false,"given":"Autumn","family":"Toney","sequence":"first","affiliation":[{"name":"Georgetown University, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-0039-0975","authenticated-orcid":false,"given":"Kathleen","family":"Curlee","sequence":"additional","affiliation":[{"name":"Georgetown University, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-2043-8644","authenticated-orcid":false,"given":"Emelia","family":"Probasco","sequence":"additional","affiliation":[{"name":"Georgetown University, USA"}]}],"member":"320","published-online":{"date-parts":[[2024,6,5]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"crossref","unstructured":"NIST AI. 2023. Artificial Intelligence Risk Management Framework (AI RMF 1.0). (2023).","DOI":"10.6028\/NIST.AI.100-1.jpn"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2023.101805"},{"key":"e_1_3_2_1_3_1","unstructured":"IEEE\u00a0Standards Association. 2017. Ethically Aligned Design: A Vision for Prioritizing Human Well-being with Autonomous and Intelligent Systems. https:\/\/standards.ieee.org\/wp-content\/uploads\/import\/documents\/other\/ead_v2.pdf. Accessed: 2023-11-20."},{"key":"e_1_3_2_1_4_1","volume-title":"The Governance of Artificial Intelligence in Canada: Findings and Opportunities from a Review of 84 AI Governance Initiatives. Available at SSRN","author":"Attard-Frost Blair","year":"2023","unstructured":"Blair Attard-Frost, Ana Brandusescu, and Kelly Lyons. 2023. The Governance of Artificial Intelligence in Canada: Findings and Opportunities from a Review of 84 AI Governance Initiatives. Available at SSRN (2023)."},{"key":"e_1_3_2_1_5_1","volume-title":"Accessed","author":"Science Australia Department","year":"2023","unstructured":"Science Australia Department\u00a0of Industry and Resources. 2023. Australia\u2019s AI Ethics Principles. https:\/\/legalinstruments.oecd.org\/en\/instruments\/oecd-legal-0449. (Canberra:Australian Government, Accessed March 17, 2023."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"crossref","unstructured":"Iz Beltagy Kyle Lo and Arman Cohan. 2019. SciBERT: Pretrained Language Model for Scientific Text. In EMNLP. arXiv:arXiv:1903.10676","DOI":"10.18653\/v1\/D19-1371"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3533150"},{"key":"e_1_3_2_1_8_1","volume-title":"Identifying the development and application of artificial intelligence in scientific text. arXiv preprint arXiv:2002.07143","author":"Dunham James","year":"2020","unstructured":"James Dunham, Jennifer Melot, and Dewey Murdick. 2020. Identifying the development and application of artificial intelligence in scientific text. arXiv preprint arXiv:2002.07143 (2020)."},{"key":"e_1_3_2_1_9_1","volume-title":"Investing in AI for social good: an analysis of European national strategies. AI & society 38, 2","author":"Foffano Francesca","year":"2023","unstructured":"Francesca Foffano, Teresa Scantamburlo, and Atia Cort\u00e9s. 2023. Investing in AI for social good: an analysis of European national strategies. AI & society 38, 2 (2023), 479\u2013500."},{"key":"e_1_3_2_1_10_1","unstructured":"Japan\u00a0Council for Social Principles of Human-centric AI. 2019. Social Principles of Human-Centric AI. https:\/\/ai.bsa.org\/wp-content\/uploads\/2019\/09\/humancentricai.pdf. Accessed: 2023-11-20."},{"key":"e_1_3_2_1_11_1","unstructured":"The\u00a0White House. 2021. The United States Joined by Allies and Partners Attributes Malicious Cyber Activity and Irresponsible State Behavior to the People\u2019s Republic of China. https:\/\/www.whitehouse.gov\/briefing-room\/statements-releases\/2021\/07\/19\/the-united-states-joined-by-allies-and-partners-attributes-malicious-cyber-activity-and-irresponsible-state-behavior-to-the-peoples-republic-of-china\/. . Accessed: 2023-11-20."},{"key":"e_1_3_2_1_12_1","volume-title":"The global landscape of AI ethics guidelines. Nature machine intelligence 1, 9","author":"Jobin Anna","year":"2019","unstructured":"Anna Jobin, Marcello Ienca, and Effy Vayena. 2019. The global landscape of AI ethics guidelines. Nature machine intelligence 1, 9 (2019), 389\u2013399."},{"key":"e_1_3_2_1_13_1","unstructured":"Joe King Betty Zhang Hanif Mahboobi and Shantu Roy. 2021. Model Explainability with AWS Artificial Intelligence and Machine Learning Solutions. https:\/\/docs.aws.amazon.com\/whitepapers\/latest\/model-explainability-aws-ai-ml\/interpretability-versus-explainability.html. Accessed: 2023-11-20."},{"key":"e_1_3_2_1_14_1","unstructured":"United Kingdom. 2018. Data Protection Act. https:\/\/www.legislation.gov.uk\/ukpga\/2018\/12\/contents\/enacted. Accessed: 2023-11-20."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1111\/rego.12512"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/1242572.1242592"},{"key":"e_1_3_2_1_17_1","unstructured":"OECD. 2023. Recommendation of the Council on OECD Legal Instruments Artificial Intelligence. https:\/\/legalinstruments.oecd.org\/en\/instruments\/oecd-legal-0449. OECD\/LEGAL\/0449 Accessed: 2023-11-20."},{"key":"e_1_3_2_1_18_1","unstructured":"The\u00a0Government of Canada. 2019. Policy on Service and Digital. https:\/\/www.tbs-sct.canada.ca\/pol\/doc-eng.aspx?id=32603. Accessed: 2023-11-20."},{"key":"e_1_3_2_1_19_1","unstructured":"The\u00a0Government of Canada. 2023. Directive on Automated Decision-Making. https:\/\/www.tbs-sct.canada.ca\/pol\/doc-eng.aspx?id=32592. Accessed: 2023-11-20."},{"key":"e_1_3_2_1_20_1","unstructured":"The\u00a0Government of Canada. 2023. Privacy Act. https:\/\/laws-lois.justice.gc.ca. Accessed: 2023-11-20."},{"key":"e_1_3_2_1_21_1","unstructured":"The\u00a0Government of Canada. 2023. Responsible Use of Artificial Intelligence Guiding Principles. https:\/\/www.canada.ca\/en\/government\/system\/digital-government\/digital-government-innovations\/responsible-use-ai.html. Accessed: 2023-11-20."},{"key":"e_1_3_2_1_22_1","unstructured":"Executive\u00a0Office of\u00a0the President. 2019. E.O. 13859;84 FR 3967;2019-02544. 3967\u20133972 pages."},{"key":"e_1_3_2_1_23_1","unstructured":"Executive\u00a0Office of\u00a0the President. 2020. E.O. 13960;85 FR 78939;2020-27065. 78939\u201378943 pages."},{"key":"e_1_3_2_1_24_1","unstructured":"United Kingdom Information\u00a0Commissioner\u2019s Office. 2018. UK GDPR guidance and resources. https:\/\/ico.org.uk\/for-organisations\/uk-gdpr-guidance-and-resources\/. Accessed: 2023-11-20."},{"key":"e_1_3_2_1_25_1","unstructured":"United Kingdom Information\u00a0Commissioner\u2019s Office. 2023. Guidance on AI and data protection. https:\/\/ico.org.uk\/for-organisations\/uk-gdpr-guidance-and-resources\/artificial-intelligence\/guidance-on-ai-and-data-protection\/. Accessed: 2023-11-20."},{"key":"e_1_3_2_1_26_1","unstructured":"United Kingdom Information\u00a0Commissioner\u2019s Office and The Alan\u00a0Turing Institute. 2022. Explaining decisions made with AI\u2019s. https:\/\/ico.org.uk\/for-organisations\/uk-gdpr-guidance-and-resources\/artificial-intelligence\/explaining-decisions-made-with-artificial-intelligence\/. Accessed: 2023-11-20."},{"key":"e_1_3_2_1_27_1","unstructured":"Japan Expert\u00a0Group on\u00a0How AI Principles Should\u00a0be Implemented. 2022. Integrated Innovation Strategy Promotion Council Governance Guidelines for Implementation of AI Principles. https:\/\/www.meti.go.jp\/shingikai\/mono_info_service\/ai_shakai_jisso\/pdf\/20220128_2.pdf. Accessed: 2023-11-20."},{"key":"e_1_3_2_1_28_1","unstructured":"International\u00a0Standards Organization. 2022. Trustworthiness \u2014 Vocabulary. https:\/\/www.iso.org\/obp\/ui\/##iso:std:iso-iec:ts:5723:ed-1:v1:en. Accessed: 2023-11-20."},{"key":"e_1_3_2_1_29_1","volume-title":"Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency. 1139\u20131150","author":"Panigutti Cecilia","year":"2023","unstructured":"Cecilia Panigutti, Ronan Hamon, Isabelle Hupont, David Fernandez\u00a0Llorca, Delia Fano\u00a0Yela, Henrik Junklewitz, Salvatore Scalzo, Gabriele Mazzini, Ignacio Sanchez, Josep Soler\u00a0Garrido, 2023. The role of explainable AI in the context of the AI Act. In Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency. 1139\u20131150."},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1145\/3593013.3594002"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"crossref","first-page":"1020592","DOI":"10.3389\/frai.2023.1020592","article-title":"The assessment list for trustworthy artificial intelligence: A review and recommendations","volume":"6","author":"Radclyffe Charles","year":"2023","unstructured":"Charles Radclyffe, Mafalda Ribeiro, and Robert\u00a0H Wortham. 2023. The assessment list for trustworthy artificial intelligence: A review and recommendations. Frontiers in Artificial Intelligence 6 (2023), 1020592.","journal-title":"Frontiers in Artificial Intelligence"},{"key":"e_1_3_2_1_32_1","volume-title":"The Chinese approach to artificial intelligence: an analysis of policy, ethics, and regulation. AI & society 36","author":"Roberts Huw","year":"2021","unstructured":"Huw Roberts, Josh Cowls, Jessica Morley, Mariarosaria Taddeo, Vincent Wang, and Luciano Floridi. 2021. The Chinese approach to artificial intelligence: an analysis of policy, ethics, and regulation. AI & society 36 (2021), 59\u201377."},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.techsoc.2020.101421"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1145\/3375627.3375804"},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2021.05.009"},{"key":"e_1_3_2_1_36_1","unstructured":"T.\u00a0Russell Vought. 2020. OMB Memorandum M-21-06: Guidance for Regulation of Artificial Intelligence Applications. https:\/\/www.whitehouse.gov\/wp-content\/uploads\/2020\/11\/M-21-06.pdf. Accessed: 2023-11-20."}],"event":{"name":"FAccT '24: The 2024 ACM Conference on Fairness, Accountability, and Transparency","acronym":"FAccT '24","location":"Rio de Janeiro Brazil"},"container-title":["The 2024 ACM Conference on Fairness, Accountability, and Transparency"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3630106.3659035","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3630106.3659035","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T23:57:07Z","timestamp":1750291027000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3630106.3659035"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,6,3]]},"references-count":36,"alternative-id":["10.1145\/3630106.3659035","10.1145\/3630106"],"URL":"https:\/\/doi.org\/10.1145\/3630106.3659035","relation":{},"subject":[],"published":{"date-parts":[[2024,6,3]]},"assertion":[{"value":"2024-06-05","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}