{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,8]],"date-time":"2026-04-08T03:00:51Z","timestamp":1775617251772,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":34,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,8,4]],"date-time":"2023-08-04T00:00:00Z","timestamp":1691107200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,8,6]]},"DOI":"10.1145\/3580305.3599557","type":"proceedings-article","created":{"date-parts":[[2023,8,4]],"date-time":"2023-08-04T18:13:58Z","timestamp":1691172838000},"page":"5805-5806","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":74,"title":["Generative AI meets Responsible AI: Practical Challenges and Opportunities"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1237-087X","authenticated-orcid":false,"given":"Krishnaram","family":"Kenthapadi","sequence":"first","affiliation":[{"name":"Fiddler AI, Palo Alto, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7922-6544","authenticated-orcid":false,"given":"Himabindu","family":"Lakkaraju","sequence":"additional","affiliation":[{"name":"Harvard University, Cambridge, MA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6301-1960","authenticated-orcid":false,"given":"Nazneen","family":"Rajani","sequence":"additional","affiliation":[{"name":"Hugging Face, Palo Alto, CA, USA"}]}],"member":"320","published-online":{"date-parts":[[2023,8,4]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-021-00359-2"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445922"},{"key":"e_1_3_2_1_3_1","unstructured":"Tolga Bolukbasi Kai-Wei Chang James Y Zou Venkatesh Saligrama and Adam T Kalai. 2016. Man is to computer programmer as woman is to homemaker? Debiasing word embeddings. In NeurIPS.  Tolga Bolukbasi Kai-Wei Chang James Y Zou Venkatesh Saligrama and Adam T Kalai. 2016. Man is to computer programmer as woman is to homemaker? Debiasing word embeddings. In NeurIPS."},{"key":"e_1_3_2_1_4_1","unstructured":"Rishi Bommasani et al. 2021. On the Opportunities and Risks of Foundation Models. ArXiv (2021). https:\/\/crfm.stanford.edu\/assets\/report.pdf  Rishi Bommasani et al. 2021. On the Opportunities and Risks of Foundation Models. ArXiv (2021). https:\/\/crfm.stanford.edu\/assets\/report.pdf"},{"key":"e_1_3_2_1_5_1","volume-title":"Improving Transparency in AI Language Models: A Holistic Evaluation. Stanford HAI Policy Brief","author":"Bommasani Rishi","year":"2023","unstructured":"Rishi Bommasani , Daniel Zhang , Tony Lee , and Percy Liang . 2023. Improving Transparency in AI Language Models: A Holistic Evaluation. Stanford HAI Policy Brief ( 2023 ). Rishi Bommasani, Daniel Zhang, Tony Lee, and Percy Liang. 2023. Improving Transparency in AI Language Models: A Holistic Evaluation. Stanford HAI Policy Brief (2023)."},{"key":"e_1_3_2_1_6_1","volume-title":"Extracting training data from diffusion models. arXiv preprint arXiv:2301.13188","author":"Carlini Nicholas","year":"2023","unstructured":"Nicholas Carlini , Jamie Hayes , Milad Nasr , Matthew Jagielski , Vikash Sehwag , Florian Tram\u00e8r , Borja Balle , Daphne Ippolito , and Eric Wallace . 2023. Extracting training data from diffusion models. arXiv preprint arXiv:2301.13188 ( 2023 ). Nicholas Carlini, Jamie Hayes, Milad Nasr, Matthew Jagielski, Vikash Sehwag, Florian Tram\u00e8r, Borja Balle, Daphne Ippolito, and Eric Wallace. 2023. Extracting training data from diffusion models. arXiv preprint arXiv:2301.13188 (2023)."},{"key":"e_1_3_2_1_7_1","volume-title":"USENIX Security Symposium","volume":"6","author":"Carlini Nicholas","year":"2021","unstructured":"Nicholas Carlini , Florian Tramer , Eric Wallace , Matthew Jagielski , Ariel Herbert-Voss , Katherine Lee , Adam Roberts , Tom Brown , Dawn Song , Ulfar Erlingsson , Alina Oprea , and Colin Raffel . 2021 . Extracting Training Data from Large Language Models . In USENIX Security Symposium , Vol. 6 . Nicholas Carlini, Florian Tramer, Eric Wallace, Matthew Jagielski, Ariel Herbert-Voss, Katherine Lee, Adam Roberts, Tom Brown, Dawn Song, Ulfar Erlingsson, Alina Oprea, and Colin Raffel. 2021. Extracting Training Data from Large Language Models. In USENIX Security Symposium, Vol. 6."},{"key":"e_1_3_2_1_8_1","unstructured":"Paul F Christiano Jan Leike Tom Brown Miljan Martic Shane Legg and Dario Amodei. 2017. Deep reinforcement learning from human preferences. In NeurIPS.  Paul F Christiano Jan Leike Tom Brown Miljan Martic Shane Legg and Dario Amodei. 2017. Deep reinforcement learning from human preferences. In NeurIPS."},{"key":"e_1_3_2_1_9_1","unstructured":"Hyung Won Chung Le Hou Shayne Longpre Barret Zoph Yi Tay William Fedus Eric Li Xuezhi Wang Mostafa Dehghani Siddhartha Brahma etal 2022. Scaling instruction-finetuned language models. arXiv preprint arXiv:2210.11416 (2022).  Hyung Won Chung Le Hou Shayne Longpre Barret Zoph Yi Tay William Fedus Eric Li Xuezhi Wang Mostafa Dehghani Siddhartha Brahma et al. 2022. Scaling instruction-finetuned language models. arXiv preprint arXiv:2210.11416 (2022)."},{"key":"e_1_3_2_1_10_1","unstructured":"Deep Ganguli Liane Lovitt Jackson Kernion Amanda Askell Yuntao Bai Saurav Kadavath Ben Mann Ethan Perez Nicholas Schiefer Kamal Ndousse etal 2022. Red teaming language models to reduce harms: Methods scaling behaviors and lessons learned. arXiv preprint arXiv:2209.07858 (2022).  Deep Ganguli Liane Lovitt Jackson Kernion Amanda Askell Yuntao Bai Saurav Kadavath Ben Mann Ethan Perez Nicholas Schiefer Kamal Ndousse et al. 2022. Red teaming language models to reduce harms: Methods scaling behaviors and lessons learned. arXiv preprint arXiv:2209.07858 (2022)."},{"key":"e_1_3_2_1_11_1","volume-title":"ICML Workshop on Knowledge Retrieval and Language Models.","author":"Huang Jie","year":"2022","unstructured":"Jie Huang , Hanyin Shao , and Kevin Chang . 2022 b. Are Large Pre-Trained Language Models Leaking Your Personal Information? . In ICML Workshop on Knowledge Retrieval and Language Models. Jie Huang, Hanyin Shao, and Kevin Chang. 2022b. Are Large Pre-Trained Language Models Leaking Your Personal Information?. In ICML Workshop on Knowledge Retrieval and Language Models."},{"key":"e_1_3_2_1_12_1","volume-title":"Generative AI: A Creative New World. https:\/\/www.sequoiacap.com\/article\/generative-ai-a-creative-new-world\/","author":"Huang Sonya","year":"2022","unstructured":"Sonya Huang , Pat Grady , and GPT-3. 2022 a. Generative AI: A Creative New World. https:\/\/www.sequoiacap.com\/article\/generative-ai-a-creative-new-world\/ Sonya Huang, Pat Grady, and GPT-3. 2022a. Generative AI: A Creative New World. https:\/\/www.sequoiacap.com\/article\/generative-ai-a-creative-new-world\/"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"crossref","unstructured":"Ben Hutchinson Vinodkumar Prabhakaran Emily Denton Kellie Webster Yu Zhong and Stephen Denuyl. 2020. Social Biases in NLP Models as Barriers for Persons with Disabilities. In ACL. 5491--5501.  Ben Hutchinson Vinodkumar Prabhakaran Emily Denton Kellie Webster Yu Zhong and Stephen Denuyl. 2020. Social Biases in NLP Models as Barriers for Persons with Disabilities. In ACL. 5491--5501.","DOI":"10.18653\/v1\/2020.acl-main.487"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/3546954"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.nuse-1.5"},{"key":"e_1_3_2_1_16_1","unstructured":"James MacGlashan Mark K Ho Robert Loftin Bei Peng Guan Wang David L Roberts Matthew E Taylor and Michael L Littman. 2017. Interactive learning from policy-dependent human feedback. In ICML.  James MacGlashan Mark K Ho Robert Loftin Bei Peng Guan Wang David L Roberts Matthew E Taylor and Michael L Littman. 2017. Interactive learning from policy-dependent human feedback. In ICML."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/3476415.3476428"},{"key":"e_1_3_2_1_18_1","volume-title":"Hannah Rose Kirk, and Luciano Floridi","author":"M\u00f6kander Jakob","year":"2023","unstructured":"Jakob M\u00f6kander , Jonas Schuett , Hannah Rose Kirk, and Luciano Floridi . 2023 . Auditing large language models: A three-layered approach. arXiv preprint arXiv:2302.08500 (2023). Jakob M\u00f6kander, Jonas Schuett, Hannah Rose Kirk, and Luciano Floridi. 2023. Auditing large language models: A three-layered approach. arXiv preprint arXiv:2302.08500 (2023)."},{"key":"e_1_3_2_1_19_1","unstructured":"Andrew Ng. 2022. ChatGPT Mania! Crypto Fiasco Defunds AI Safety Alexa Tells Bedtime Stories. https:\/\/www.deeplearning.ai\/the-batch\/issue-174\/ The Batch - Deeplearning.ai newsletter.  Andrew Ng. 2022. ChatGPT Mania! Crypto Fiasco Defunds AI Safety Alexa Tells Bedtime Stories. https:\/\/www.deeplearning.ai\/the-batch\/issue-174\/ The Batch - Deeplearning.ai newsletter."},{"key":"e_1_3_2_1_20_1","unstructured":"Long Ouyang et al. 2022. Training language models to follow instructions with human feedback. arXiv preprint arXiv:2203.02155 (2022).  Long Ouyang et al. 2022. Training language models to follow instructions with human feedback. arXiv preprint arXiv:2203.02155 (2022)."},{"key":"e_1_3_2_1_21_1","volume-title":"Carbon emissions and large neural network training. arXiv preprint arXiv:2104.10350","author":"Patterson David","year":"2021","unstructured":"David Patterson , Joseph Gonzalez , Quoc Le , Chen Liang , Lluis-Miquel Munguia , Daniel Rothchild , David So , Maud Texier , and Jeff Dean . 2021. Carbon emissions and large neural network training. arXiv preprint arXiv:2104.10350 ( 2021 ). David Patterson, Joseph Gonzalez, Quoc Le, Chen Liang, Lluis-Miquel Munguia, Daniel Rothchild, David So, Maud Texier, and Jeff Dean. 2021. Carbon emissions and large neural network training. arXiv preprint arXiv:2104.10350 (2021)."},{"key":"e_1_3_2_1_22_1","volume-title":"Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al.","author":"Radford Alec","year":"2021","unstructured":"Alec Radford , Jong Wook Kim , Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. 2021 . Learning transferable visual models from natural language supervision. In ICML. Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. 2021. Learning transferable visual models from natural language supervision. In ICML."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"crossref","unstructured":"Robin Rombach Andreas Blattmann Dominik Lorenz Patrick Esser and Bj\u00f6rn Ommer. 2022. High-resolution image synthesis with latent diffusion models. In CVPR.  Robin Rombach Andreas Blattmann Dominik Lorenz Patrick Esser and Bj\u00f6rn Ommer. 2022. High-resolution image synthesis with latent diffusion models. In CVPR.","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"e_1_3_2_1_24_1","volume-title":"BayesFormer: Transformer with Uncertainty Estimation. arXiv preprint arXiv:2206.00826","author":"Sankararaman Karthik Abinav","year":"2022","unstructured":"Karthik Abinav Sankararaman , Sinong Wang , and Han Fang . 2022. BayesFormer: Transformer with Uncertainty Estimation. arXiv preprint arXiv:2206.00826 ( 2022 ). Karthik Abinav Sankararaman, Sinong Wang, and Han Fang. 2022. BayesFormer: Transformer with Uncertainty Estimation. arXiv preprint arXiv:2206.00826 (2022)."},{"key":"e_1_3_2_1_25_1","volume-title":"Francc ois Yvon, Matthias Gall\u00e9, et al.","author":"Scao Teven Le","year":"2022","unstructured":"Teven Le Scao , Angela Fan , Christopher Akiki , Ellie Pavlick , Suzana Ili\u0107 , Daniel Hesslow , Roman Castagn\u00e9 , Alexandra Sasha Luccioni , Francc ois Yvon, Matthias Gall\u00e9, et al. 2022 . BLOOM : A 176B-Parameter Open-Access Multilingual Language Model . arXiv preprint arXiv:2211.05100 (2022). Teven Le Scao, Angela Fan, Christopher Akiki, Ellie Pavlick, Suzana Ili\u0107, Daniel Hesslow, Roman Castagn\u00e9, Alexandra Sasha Luccioni, Francc ois Yvon, Matthias Gall\u00e9, et al. 2022. BLOOM: A 176B-Parameter Open-Access Multilingual Language Model. arXiv preprint arXiv:2211.05100 (2022)."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1145\/3498366.3505816"},{"key":"e_1_3_2_1_27_1","volume-title":"The cost of training NLP models: A concise overview. arXiv preprint arXiv:2004.08900","author":"Sharir Or","year":"2020","unstructured":"Or Sharir , Barak Peleg , and Yoav Shoham . 2020. The cost of training NLP models: A concise overview. arXiv preprint arXiv:2004.08900 ( 2020 ). Or Sharir, Barak Peleg, and Yoav Shoham. 2020. The cost of training NLP models: A concise overview. arXiv preprint arXiv:2004.08900 (2020)."},{"key":"e_1_3_2_1_28_1","volume-title":"The Gradient of Generative AI Release: Methods and Considerations. arXiv preprint arXiv:2302.04844","author":"Solaiman Irene","year":"2023","unstructured":"Irene Solaiman . 2023. The Gradient of Generative AI Release: Methods and Considerations. arXiv preprint arXiv:2302.04844 ( 2023 ). Irene Solaiman. 2023. The Gradient of Generative AI Release: Methods and Considerations. arXiv preprint arXiv:2302.04844 (2023)."},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"crossref","unstructured":"Eric Wallace Tony Zhao Shi Feng and Sameer Singh. 2021. Concealed Data Poisoning Attacks on NLP Models. In NAACL-HLT.  Eric Wallace Tony Zhao Shi Feng and Sameer Singh. 2021. Concealed Data Poisoning Attacks on NLP Models. In NAACL-HLT.","DOI":"10.18653\/v1\/2021.naacl-main.13"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"crossref","unstructured":"Yizhong Wang et al. 2022a. Super-NaturalInstructions: Generalization via Declarative Instructions on 1600 NLP Tasks. In EMNLP.  Yizhong Wang et al. 2022a. Super-NaturalInstructions: Generalization via Declarative Instructions on 1600 NLP Tasks. In EMNLP.","DOI":"10.18653\/v1\/2022.emnlp-main.340"},{"key":"e_1_3_2_1_31_1","volume-title":"Self-Instruct: Aligning Language Model with Self Generated Instructions. arXiv preprint arXiv:2212.10560","author":"Wang Yizhong","year":"2022","unstructured":"Yizhong Wang , Yeganeh Kordi , Swaroop Mishra , Alisa Liu , Noah A Smith , Daniel Khashabi , and Hannaneh Hajishirzi . 2022b. Self-Instruct: Aligning Language Model with Self Generated Instructions. arXiv preprint arXiv:2212.10560 ( 2022 ). Yizhong Wang, Yeganeh Kordi, Swaroop Mishra, Alisa Liu, Noah A Smith, Daniel Khashabi, and Hannaneh Hajishirzi. 2022b. Self-Instruct: Aligning Language Model with Self Generated Instructions. arXiv preprint arXiv:2212.10560 (2022)."},{"key":"e_1_3_2_1_32_1","volume-title":"Contrastive Language-Vision AI Models Pretrained on Web-Scraped Multimodal Data Exhibit Sexual Objectification Bias. arXiv preprint arXiv:2212.11261","author":"Wolfe Robert","year":"2022","unstructured":"Robert Wolfe , Yiwei Yang , Bill Howe , and Aylin Caliskan . 2022. Contrastive Language-Vision AI Models Pretrained on Web-Scraped Multimodal Data Exhibit Sexual Objectification Bias. arXiv preprint arXiv:2212.11261 ( 2022 ). Robert Wolfe, Yiwei Yang, Bill Howe, and Aylin Caliskan. 2022. Contrastive Language-Vision AI Models Pretrained on Web-Scraped Multimodal Data Exhibit Sexual Objectification Bias. arXiv preprint arXiv:2212.11261 (2022)."},{"key":"e_1_3_2_1_33_1","volume-title":"Xi Victoria Lin, et al","author":"Zhang Susan","year":"2022","unstructured":"Susan Zhang , Stephen Roller , Naman Goyal , Mikel Artetxe , Moya Chen , Shuohui Chen , Christopher Dewan , Mona Diab , Xian Li , Xi Victoria Lin, et al . 2022 . OPT : Open pre-trained transformer language models. arXiv preprint arXiv:2205.01068 (2022). Susan Zhang, Stephen Roller, Naman Goyal, Mikel Artetxe, Moya Chen, Shuohui Chen, Christopher Dewan, Mona Diab, Xian Li, Xi Victoria Lin, et al. 2022. OPT: Open pre-trained transformer language models. arXiv preprint arXiv:2205.01068 (2022)."},{"key":"e_1_3_2_1_34_1","volume-title":"Fine-tuning language models from human preferences. arXiv preprint arXiv:1909.08593","author":"Ziegler Daniel M","year":"2019","unstructured":"Daniel M Ziegler , Nisan Stiennon , Jeffrey Wu , Tom B Brown , Alec Radford , Dario Amodei , Paul Christiano , and Geoffrey Irving . 2019. Fine-tuning language models from human preferences. arXiv preprint arXiv:1909.08593 ( 2019 ). Daniel M Ziegler, Nisan Stiennon, Jeffrey Wu, Tom B Brown, Alec Radford, Dario Amodei, Paul Christiano, and Geoffrey Irving. 2019. Fine-tuning language models from human preferences. arXiv preprint arXiv:1909.08593 (2019)."}],"event":{"name":"KDD '23: The 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining","location":"Long Beach CA USA","acronym":"KDD '23","sponsor":["SIGMOD ACM Special Interest Group on Management of Data","SIGKDD ACM Special Interest Group on Knowledge Discovery in Data"]},"container-title":["Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3580305.3599557","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3580305.3599557","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:37:52Z","timestamp":1750178272000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3580305.3599557"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,8,4]]},"references-count":34,"alternative-id":["10.1145\/3580305.3599557","10.1145\/3580305"],"URL":"https:\/\/doi.org\/10.1145\/3580305.3599557","relation":{},"subject":[],"published":{"date-parts":[[2023,8,4]]},"assertion":[{"value":"2023-08-04","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}