{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T21:48:53Z","timestamp":1776116933226,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":88,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,6,3]],"date-time":"2024-06-03T00:00:00Z","timestamp":1717372800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/https:\/\/doi.org\/10.13039\/100006754","name":"Army Research Laboratory","doi-asserted-by":"publisher","award":["ARL DCIST CRA W911NF-17-2-0181"],"award-info":[{"award-number":["ARL DCIST CRA W911NF-17-2-0181"]}],"id":[{"id":"10.13039\/https:\/\/doi.org\/10.13039\/100006754","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,6,3]]},"DOI":"10.1145\/3630106.3659032","type":"proceedings-article","created":{"date-parts":[[2024,6,5]],"date-time":"2024-06-05T13:14:21Z","timestamp":1717593261000},"page":"2175-2198","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":9,"title":["An Information Bottleneck Characterization of the Understanding-Workload Tradeoff in Human-Centered Explainable AI"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-4247-3004","authenticated-orcid":false,"given":"Lindsay","family":"Sanneman","sequence":"first","affiliation":[{"name":"CSAIL, Massachusetts Institute of Technology, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1160-9789","authenticated-orcid":false,"given":"Mycal","family":"Tucker","sequence":"additional","affiliation":[{"name":"CSAIL, Massachusetts Institute of Technology, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1338-8107","authenticated-orcid":false,"given":"Julie A.","family":"Shah","sequence":"additional","affiliation":[{"name":"CSAIL, Massachusetts Institute of Technology, USA"}]}],"member":"320","published-online":{"date-parts":[[2024,6,5]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Sanity checks for saliency maps. Advances in neural information processing systems 31","author":"Adebayo Julius","year":"2018","unstructured":"Julius Adebayo, Justin Gilmer, Michael Muelly, Ian Goodfellow, Moritz Hardt, and Been Kim. 2018. Sanity checks for saliency maps. Advances in neural information processing systems 31 (2018)."},{"key":"e_1_3_2_1_2_1","volume-title":"Deep Variational Information Bottleneck. In 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings. OpenReview.net.","author":"Alemi A.","year":"2017","unstructured":"Alexander\u00a0A. Alemi, Ian Fischer, Joshua\u00a0V. Dillon, and Kevin Murphy. 2017. Deep Variational Information Bottleneck. In 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings. OpenReview.net."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290605.3300233"},{"key":"e_1_3_2_1_4_1","volume-title":"Proceedings of the 17th International Conference on Autonomous Agents and MultiAgent Systems. 1168\u20131176","author":"Amir Dan","year":"2018","unstructured":"Dan Amir and Ofra Amir. 2018. Highlights: Summarizing agent behavior to people. In Proceedings of the 17th International Conference on Autonomous Agents and MultiAgent Systems. 1168\u20131176."},{"key":"e_1_3_2_1_5_1","volume-title":"Proceedings of the 17th International Conference on Autonomous Agents and MultiAgent Systems. 1203\u20131207","author":"Amir Ofra","year":"2018","unstructured":"Ofra Amir, Finale Doshi-Velez, and David Sarne. 2018. Agent strategy summarization. In Proceedings of the 17th International Conference on Autonomous Agents and MultiAgent Systems. 1203\u20131207."},{"key":"e_1_3_2_1_6_1","volume-title":"ASQ-IT: Interactive Explanations for Reinforcement-Learning Agents. arXiv preprint arXiv:2301.09941","author":"Amitai Yotam","year":"2023","unstructured":"Yotam Amitai, Guy Avni, and Ofra Amir. 2023. ASQ-IT: Interactive Explanations for Reinforcement-Learning Agents. arXiv preprint arXiv:2301.09941 (2023)."},{"key":"e_1_3_2_1_7_1","volume-title":"Explaining reinforcement learning to mere mortals: An empirical study. arXiv preprint arXiv:1903.09708","author":"Anderson Andrew","year":"2019","unstructured":"Andrew Anderson, Jonathan Dodge, Amrita Sadarangani, Zoe Juozapaitis, Evan Newman, Jed Irvine, Souti Chattopadhyay, Alan Fern, and Margaret Burnett. 2019. Explaining reinforcement learning to mere mortals: An empirical study. arXiv preprint arXiv:1903.09708 (2019)."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"crossref","unstructured":"Guy Aridor Rava\u00a0Azeredo da Silveira and Michael Woodford. 2023. Information-Constrained Coordination of Economic Behavior. (2023).","DOI":"10.2139\/ssrn.4723728"},{"key":"e_1_3_2_1_9_1","volume-title":"Training a helpful and harmless assistant with reinforcement learning from human feedback. arXiv preprint arXiv:2204.05862","author":"Bai Yuntao","year":"2022","unstructured":"Yuntao Bai, Andy Jones, Kamal Ndousse, Amanda Askell, Anna Chen, Nova DasSarma, Dawn Drain, Stanislav Fort, Deep Ganguli, Tom Henighan, 2022. Training a helpful and harmless assistant with reinforcement learning from human feedback. arXiv preprint arXiv:2204.05862 (2022)."},{"key":"e_1_3_2_1_10_1","volume-title":"Proceedings of the AAAI Conference on Artificial Intelligence, Vol.\u00a035","author":"Bang Seojin","year":"2021","unstructured":"Seojin Bang, Pengtao Xie, Heewook Lee, Wei Wu, and Eric Xing. 2021. Explaining a black-box by using a deep variational information bottleneck approach. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol.\u00a035. 11396\u201311404."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1609\/hcomp.v7i1.5285"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1007\/s42454-020-00018-8"},{"key":"e_1_3_2_1_13_1","volume-title":"2019 14th ACM\/IEEE International Conference on Human-Robot Interaction (HRI). IEEE, 258\u2013266","author":"Chakraborti Tathagata","year":"2019","unstructured":"Tathagata Chakraborti, Sarath Sreedharan, Sachin Grover, and Subbarao Kambhampati. 2019. Plan explanations as model reconciliation. In 2019 14th ACM\/IEEE International Conference on Human-Robot Interaction (HRI). IEEE, 258\u2013266."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1109\/SMC.2016.7844505"},{"key":"e_1_3_2_1_15_1","volume-title":"Micro-and nanotechnology sensors, systems, and applications IX, Vol.\u00a010194","author":"Chen YC","unstructured":"Jessie\u00a0YC Chen, Michael\u00a0J Barnes, Julia\u00a0L Wright, Kimberly Stowers, and Shan\u00a0G Lakhmani. 2017. Situation awareness-based agent transparency for human-autonomy teaming effectiveness. In Micro-and nanotechnology sensors, systems, and applications IX, Vol.\u00a010194. SPIE, 362\u2013367."},{"key":"e_1_3_2_1_16_1","volume-title":"Situation awareness-based agent transparency and human-autonomy teaming effectiveness. Theoretical issues in ergonomics science 19, 3","author":"Chen YC","year":"2018","unstructured":"Jessie\u00a0YC Chen, Shan\u00a0G Lakhmani, Kimberly Stowers, Anthony\u00a0R Selkowitz, Julia\u00a0L Wright, and Michael Barnes. 2018. Situation awareness-based agent transparency and human-autonomy teaming effectiveness. Theoretical issues in ergonomics science 19, 3 (2018), 259\u2013282."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"crossref","DOI":"10.21236\/ADA600351","volume-title":"Situation awareness-based agent transparency. US Army Research Laboratory","author":"Chen Y","year":"2014","unstructured":"Jessie\u00a0Y Chen, Katelyn Procci, Michael Boyce, Julia Wright, Andre Garcia, and Michael Barnes. 2014. Situation awareness-based agent transparency. US Army Research Laboratory (2014), 1\u201329."},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/3278721.3278736"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3579467"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/3301275.3302316"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1518\/001872095779049499"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1109\/NAECON.1988.195097"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1177\/1555343415572631"},{"key":"e_1_3_2_1_24_1","volume-title":"Situational Awareness","author":"Endsley R","unstructured":"Mica\u00a0R Endsley. 2017. Direct measurement of situation awareness: Validity and use of SAGAT. In Situational Awareness. Routledge, 129\u2013156."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.chb.2022.107574"},{"key":"e_1_3_2_1_26_1","volume-title":"Proceedings of the ACM on Human-Computer Interaction 4, CSCW3","author":"Ghai Bhavya","year":"2021","unstructured":"Bhavya Ghai, Q\u00a0Vera Liao, Yunfeng Zhang, Rachel Bellamy, and Klaus Mueller. 2021. Explainable active learning (xal) toward ai explanations as interfaces for machine teachers. Proceedings of the ACM on Human-Computer Interaction 4, CSCW3 (2021), 1\u201328."},{"key":"e_1_3_2_1_27_1","volume-title":"DARPA\u2019s explainable artificial intelligence (XAI) program. AI magazine 40, 2","author":"Gunning David","year":"2019","unstructured":"David Gunning and David Aha. 2019. DARPA\u2019s explainable artificial intelligence (XAI) program. AI magazine 40, 2 (2019), 44\u201358."},{"key":"e_1_3_2_1_28_1","volume-title":"Advances in psychology. Vol.\u00a052","author":"Hart G","unstructured":"Sandra\u00a0G Hart and Lowell\u00a0E Staveland. 1988. Development of NASA-TLX (Task Load Index): Results of empirical and theoretical research. In Advances in psychology. Vol.\u00a052. Elsevier, 139\u2013183."},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/2909824.3020233"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1609\/hcomp.v10i1.21990"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1145\/3610067"},{"key":"e_1_3_2_1_32_1","volume-title":"The value of abstraction. Current opinion in behavioral sciences 29","author":"K Ho.","year":"2019","unstructured":"Mark\u00a0K Ho. 2019. The value of abstraction. Current opinion in behavioral sciences 29 (2019)."},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1109\/THMS.2019.2904558"},{"key":"e_1_3_2_1_34_1","volume-title":"Metrics for explainable AI: Challenges and prospects. arXiv preprint arXiv:1812.04608","author":"Hoffman R","year":"2018","unstructured":"Robert\u00a0R Hoffman, Shane\u00a0T Mueller, Gary Klein, and Jordan Litman. 2018. Metrics for explainable AI: Challenges and prospects. arXiv preprint arXiv:1812.04608 (2018)."},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10514-018-9771-0"},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.artint.2021.103571"},{"key":"e_1_3_2_1_37_1","volume-title":"C","author":"Huey Holly","year":"2023","unstructured":"Holly Huey, Xuanchen Lu, Caren\u00a0M. Walker, and Judith\u00a0E. Fan. 2023. Visual Explanations Prioritize Functional Properties at the Expense of Visual Fidelity. Cognition 236, C (2023), 105414."},{"key":"e_1_3_2_1_38_1","volume-title":"IJCAI\/ECAI Workshop on explainable artificial intelligence.","author":"Juozapaitis Zoe","year":"2019","unstructured":"Zoe Juozapaitis, Anurag Koul, Alan Fern, Martin Erwig, and Finale Doshi-Velez. 2019. Explainable reinforcement learning via reward decomposition. In IJCAI\/ECAI Workshop on explainable artificial intelligence."},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/3544549.3585742"},{"key":"e_1_3_2_1_40_1","volume-title":"The world color survey","author":"Kay Paul","unstructured":"Paul Kay, Brent Berlin, Luisa Maffi, William\u00a0R Merrifield, and Richard Cook. 2009. The world color survey. CSLI Publications Stanford, CA."},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1145\/3544548.3581001"},{"key":"e_1_3_2_1_42_1","volume-title":"An evaluation of the human-interpretability of explanation. arXiv preprint arXiv:1902.00006","author":"Lage Isaac","year":"2019","unstructured":"Isaac Lage, Emily Chen, Jeffrey He, Menaka Narayanan, Been Kim, Sam Gershman, and Finale Doshi-Velez. 2019. An evaluation of the human-interpretability of explanation. arXiv preprint arXiv:1902.00006 (2019)."},{"key":"e_1_3_2_1_43_1","volume-title":"Selective explanations: Leveraging human input to align explainable ai. arXiv preprint arXiv:2301.09656","author":"Lai Vivian","year":"2023","unstructured":"Vivian Lai, Yiming Zhang, Chacha Chen, Q\u00a0Vera Liao, and Chenhao Tan. 2023. Selective explanations: Leveraging human input to align explainable ai. arXiv preprint arXiv:2301.09656 (2023)."},{"key":"e_1_3_2_1_44_1","volume-title":"2022 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 9140\u20139147","author":"Lee S","year":"2022","unstructured":"Michael\u00a0S Lee, Henny Admoni, and Reid Simmons. 2022. Reasoning about Counterfactuals to Improve Human Inverse Reinforcement Learning. In 2022 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 9140\u20139147."},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376590"},{"key":"e_1_3_2_1_46_1","volume-title":"AI Transparency in the Age of LLMs: A Human-Centered Research Roadmap. arXiv preprint arXiv:2306.01941","author":"Liao Q\u00a0Vera","year":"2023","unstructured":"Q\u00a0Vera Liao and Jennifer\u00a0Wortman Vaughan. 2023. AI Transparency in the Age of LLMs: A Human-Centered Research Roadmap. arXiv preprint arXiv:2306.01941 (2023)."},{"key":"e_1_3_2_1_47_1","volume-title":"IUI Workshops.","author":"Lim Y","year":"2019","unstructured":"Brian\u00a0Y Lim, Qian Yang, Ashraf\u00a0M Abdul, and Danding Wang. 2019. Why these explanations? Selecting intelligibility types for explanation goals.. In IUI Workshops."},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0199661"},{"key":"e_1_3_2_1_49_1","volume-title":"Human mental workload: A survey and a novel inclusive definition. Frontiers in psychology 13","author":"Longo Luca","year":"2022","unstructured":"Luca Longo, Christopher\u00a0D Wickens, Gabriella Hancock, and Peter\u00a0A Hancock. 2022. Human mental workload: A survey and a novel inclusive definition. Frontiers in psychology 13 (2022), 883321."},{"key":"e_1_3_2_1_50_1","volume-title":"A unified approach to interpreting model predictions. Advances in neural information processing systems 30","author":"Lundberg M","year":"2017","unstructured":"Scott\u00a0M Lundberg and Su-In Lee. 2017. A unified approach to interpreting model predictions. Advances in neural information processing systems 30 (2017)."},{"key":"e_1_3_2_1_51_1","volume-title":"Understanding learned reward functions. arXiv preprint arXiv:2012.05862","author":"Michaud J","year":"2020","unstructured":"Eric\u00a0J Michaud, Adam Gleave, and Stuart Russell. 2020. Understanding learned reward functions. arXiv preprint arXiv:2012.05862 (2020)."},{"key":"e_1_3_2_1_52_1","volume-title":"Explanation in artificial intelligence: Insights from the social sciences. Artificial intelligence 267","author":"Miller Tim","year":"2019","unstructured":"Tim Miller. 2019. Explanation in artificial intelligence: Insights from the social sciences. Artificial intelligence 267 (2019), 1\u201338."},{"key":"e_1_3_2_1_53_1","volume-title":"Explainable AI: Beware of inmates running the asylum or: How I learnt to stop worrying and love the social and behavioural sciences. arXiv preprint arXiv:1712.00547","author":"Miller Tim","year":"2017","unstructured":"Tim Miller, Piers Howe, and Liz Sonenberg. 2017. Explainable AI: Beware of inmates running the asylum or: How I learnt to stop worrying and love the social and behavioural sciences. arXiv preprint arXiv:1712.00547 (2017)."},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"crossref","first-page":"e2025993118","DOI":"10.1073\/pnas.2025993118","article-title":"The forms and meanings of grammatical markers support efficient communication","volume":"118","author":"Mollica Francis","year":"2021","unstructured":"Francis Mollica, Geoff Bacon, Noga Zaslavsky, Yang Xu, Terry Regier, and Charles Kemp. 2021. The forms and meanings of grammatical markers support efficient communication. Proceedings of the National Academy of Sciences 118, 49 (2021), e2025993118.","journal-title":"Proceedings of the National Academy of Sciences"},{"key":"e_1_3_2_1_55_1","volume-title":"The utility of explainable ai in ad hoc human-machine teaming. Advances in neural information processing systems 34","author":"Paleja Rohan","year":"2021","unstructured":"Rohan Paleja, Muyleng Ghuy, Nadun Ranawaka\u00a0Arachchige, Reed Jensen, and Matthew Gombolay. 2021. The utility of explainable ai in ad hoc human-machine teaming. Advances in neural information processing systems 34 (2021), 610\u2013623."},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1518\/155534308X284417"},{"key":"e_1_3_2_1_57_1","article-title":"Embo: a Python package for empirical data analysis using the Information Bottleneck","volume":"9","author":"Piasini Eugenio","year":"2021","unstructured":"Eugenio Piasini, Alexandre\u00a0LS Filipowicz, Jonathan Levine, and Joshua\u00a0I Gold. 2021. Embo: a Python package for empirical data analysis using the Information Bottleneck. Journal of open research software 9, 1 (2021).","journal-title":"Journal of open research software"},{"key":"e_1_3_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.1145\/3411764.3445315"},{"key":"e_1_3_2_1_59_1","volume-title":"Thirty-seventh Conference on Neural Information Processing Systems. https:\/\/openreview.net\/forum?id=HPuSIXJaa9","author":"Rafailov Rafael","year":"2023","unstructured":"Rafael Rafailov, Archit Sharma, Eric Mitchell, Christopher\u00a0D Manning, Stefano Ermon, and Chelsea Finn. 2023. Direct Preference Optimization: Your Language Model is Secretly a Reward Model. In Thirty-seventh Conference on Neural Information Processing Systems. https:\/\/openreview.net\/forum?id=HPuSIXJaa9"},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1145\/2939672.2939778"},{"key":"e_1_3_2_1_61_1","volume-title":"Artificial intelligence a modern approach. Pearson Education","author":"Russell J","unstructured":"Stuart\u00a0J Russell. 2010. Artificial intelligence a modern approach. Pearson Education, Inc."},{"key":"e_1_3_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.2196\/19472"},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.1145\/3565472.3592959"},{"key":"e_1_3_2_1_64_1","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2022.3189441"},{"key":"e_1_3_2_1_65_1","doi-asserted-by":"publisher","DOI":"10.1080\/10447318.2022.2081282"},{"key":"e_1_3_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.chb.2023.107809"},{"key":"e_1_3_2_1_67_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.cogsys.2017.02.002"},{"key":"e_1_3_2_1_68_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.artint.2021.103535"},{"key":"e_1_3_2_1_69_1","doi-asserted-by":"crossref","unstructured":"Sarath Sreedharan Siddharth Srivastava and Subbarao Kambhampati. 2018. Hierarchical Expertise Level Modeling for User Specific Contrastive Explanations.. In IJCAI. 4829\u20134836.","DOI":"10.24963\/ijcai.2018\/671"},{"key":"e_1_3_2_1_70_1","volume-title":"The deterministic information bottleneck. Neural computation 29, 6","author":"Strouse DJ","year":"2017","unstructured":"DJ Strouse and David\u00a0J Schwab. 2017. The deterministic information bottleneck. Neural computation 29, 6 (2017), 1611\u20131630."},{"key":"e_1_3_2_1_71_1","volume-title":"Reinforcement learning: An introduction","author":"Sutton S","unstructured":"Richard\u00a0S Sutton and Andrew\u00a0G Barto. 2018. Reinforcement learning: An introduction. MIT press."},{"key":"e_1_3_2_1_72_1","volume-title":"2019 14th ACM\/IEEE International Conference on Human-Robot Interaction (HRI). IEEE, 249\u2013257","author":"Tabrez Aaquib","year":"2019","unstructured":"Aaquib Tabrez, Shivendra Agrawal, and Bradley Hayes. 2019. Explanation-based reward coaching to improve human performance via reinforcement learning. In 2019 14th ACM\/IEEE International Conference on Human-Robot Interaction (HRI). IEEE, 249\u2013257."},{"key":"e_1_3_2_1_73_1","volume-title":"Proceedings of the 37th Annual Allerton Conference on Communication, Control and Computing.","author":"Tishby Naftali","year":"1999","unstructured":"Naftali Tishby, Fernando\u00a0C. Pereira, and William Bialek. 1999. The Information Bottleneck Method. In Proceedings of the 37th Annual Allerton Conference on Communication, Control and Computing."},{"key":"e_1_3_2_1_74_1","doi-asserted-by":"crossref","first-page":"64","DOI":"10.1016\/j.iccn.2018.01.004","article-title":"The NASA Task Load Index as a measure of overall workload among neonatal, paediatric and adult intensive care nurses","volume":"46","author":"Tubbs-Cooley L","year":"2018","unstructured":"Heather\u00a0L Tubbs-Cooley, Constance\u00a0A Mara, Adam\u00a0C Carle, and Ayse\u00a0P Gurses. 2018. The NASA Task Load Index as a measure of overall workload among neonatal, paediatric and adult intensive care nurses. Intensive and Critical Care Nursing 46 (2018), 64\u201369.","journal-title":"Intensive and Critical Care Nursing"},{"key":"e_1_3_2_1_75_1","unstructured":"Mycal Tucker Roger\u00a0P. Levy Julie Shah and Noga Zaslavsky. 2022. Trading off Utility Informativeness and Complexity in Emergent Communication. In Advances in Neural Information Processing Systems Alice\u00a0H. Oh Alekh Agarwal Danielle Belgrave and Kyunghyun Cho (Eds.)."},{"key":"e_1_3_2_1_76_1","volume-title":"A human-centered agenda for intelligible machine learning. Machines We Trust: Getting Along with Artificial Intelligence","author":"Vaughan Jennifer\u00a0Wortman","year":"2020","unstructured":"Jennifer\u00a0Wortman Vaughan and Hanna Wallach. 2020. A human-centered agenda for intelligible machine learning. Machines We Trust: Getting Along with Artificial Intelligence (2020)."},{"key":"e_1_3_2_1_77_1","volume-title":"Multiple resources and mental workload. Human factors 50, 3","author":"Wickens D","year":"2008","unstructured":"Christopher\u00a0D Wickens. 2008. Multiple resources and mental workload. Human factors 50, 3 (2008), 449\u2013455."},{"key":"e_1_3_2_1_78_1","doi-asserted-by":"publisher","DOI":"10.2307\/2346786"},{"key":"e_1_3_2_1_79_1","volume-title":"Proceedings of the human factors and ergonomics society annual meeting, Vol.\u00a060","author":"Wright L","year":"2016","unstructured":"Julia\u00a0L Wright, Jessie\u00a0YC Chen, Michael\u00a0J Barnes, and Peter\u00a0A Hancock. 2016. Agent reasoning transparency\u2019s effect on operator workload. In Proceedings of the human factors and ergonomics society annual meeting, Vol.\u00a060. SAGE Publications Sage CA: Los Angeles, CA, 249\u2013253."},{"key":"e_1_3_2_1_80_1","volume-title":"Proceedings of the Human Factors and Ergonomics Society Annual Meeting, Vol.\u00a058","author":"Wright L","year":"2014","unstructured":"Julia\u00a0L Wright, Stephanie\u00a0A Quinn, Jessie\u00a0YC Chen, and Michael\u00a0J Barnes. 2014. Individual differences in human-agent teaming: An analysis of workload and situation awareness through eye movements. In Proceedings of the Human Factors and Ergonomics Society Annual Meeting, Vol.\u00a058. SAGE Publications Sage CA: Los Angeles, CA, 1410\u20131414."},{"key":"e_1_3_2_1_81_1","volume-title":"Individual differences in working memory capacity and workload capacity. Frontiers in psychology 5","author":"Yu Ju-Chi","year":"2014","unstructured":"Ju-Chi Yu, Ting-Yun Chang, and Cheng-Ta Yang. 2014. Individual differences in working memory capacity and workload capacity. Frontiers in psychology 5 (2014), 1465."},{"key":"e_1_3_2_1_82_1","volume-title":"Online explanation generation for human-robot teaming. arXiv preprint arXiv:1903.06418","author":"Zakershahrak Mehrdad","year":"2019","unstructured":"Mehrdad Zakershahrak, Ze Gong, Nikhillesh Sadassivam, and Yu Zhang. 2019. Online explanation generation for human-robot teaming. arXiv preprint arXiv:1903.06418 (2019)."},{"key":"e_1_3_2_1_83_1","volume-title":"2020 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 6304\u20136310","author":"Zakershahrak Mehrdad","year":"2020","unstructured":"Mehrdad Zakershahrak, Ze Gong, Nikhillesh Sadassivam, and Yu Zhang. 2020. Online explanation generation for planning tasks in human-robot teaming. In 2020 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 6304\u20136310."},{"key":"e_1_3_2_1_84_1","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.1800521115"},{"key":"e_1_3_2_1_85_1","volume-title":"Proceedings of the 43rd Annual Meeting of the Cognitive Science Society.","author":"Zaslavsky Noga","year":"2021","unstructured":"Noga Zaslavsky, Mora Maldonado, and Jennifer Culbertson. 2021. Let\u2019s talk (efficiently) about us: Person systems achieve near-optimal compression. In Proceedings of the 43rd Annual Meeting of the Cognitive Science Society."},{"key":"e_1_3_2_1_86_1","volume-title":"Proceedings of the 41st Annual Meeting of the Cognitive Science Society.","author":"Zaslavsky Noga","year":"2019","unstructured":"Noga Zaslavsky, Terry Regier, Naftali Tishby, and Charles Kemp. 2019. Semantic categories of artifacts and animals reflect efficient coding. In Proceedings of the 41st Annual Meeting of the Cognitive Science Society."},{"key":"e_1_3_2_1_87_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372852"},{"key":"e_1_3_2_1_88_1","volume-title":"Fine-tuning language models from human preferences. arXiv preprint arXiv:1909.08593","author":"Ziegler M","year":"2019","unstructured":"Daniel\u00a0M Ziegler, Nisan Stiennon, Jeffrey Wu, Tom\u00a0B Brown, Alec Radford, Dario Amodei, Paul Christiano, and Geoffrey Irving. 2019. Fine-tuning language models from human preferences. arXiv preprint arXiv:1909.08593 (2019)."}],"event":{"name":"FAccT '24: The 2024 ACM Conference on Fairness, Accountability, and Transparency","location":"Rio de Janeiro Brazil","acronym":"FAccT '24"},"container-title":["The 2024 ACM Conference on Fairness, Accountability, and Transparency"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3630106.3659032","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3630106.3659032","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T23:57:07Z","timestamp":1750291027000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3630106.3659032"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,6,3]]},"references-count":88,"alternative-id":["10.1145\/3630106.3659032","10.1145\/3630106"],"URL":"https:\/\/doi.org\/10.1145\/3630106.3659032","relation":{},"subject":[],"published":{"date-parts":[[2024,6,3]]},"assertion":[{"value":"2024-06-05","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}