{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,23]],"date-time":"2025-08-23T00:08:42Z","timestamp":1755907722692,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":20,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,12,15]],"date-time":"2023-12-15T00:00:00Z","timestamp":1702598400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,12,15]]},"DOI":"10.1145\/3639233.3639352","type":"proceedings-article","created":{"date-parts":[[2024,3,5]],"date-time":"2024-03-05T11:02:10Z","timestamp":1709636530000},"page":"57-62","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Explaining Prediction Uncertainty in Text Classification: The DUX Approach"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8606-9743","authenticated-orcid":false,"given":"Jakob Smedegaard","family":"Andersen","sequence":"first","affiliation":[{"name":"Department of Computer Science, Hamburg University of Applied Sciences, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8398-6132","authenticated-orcid":false,"given":"Olaf","family":"Zukunft","sequence":"additional","affiliation":[{"name":"Department of Computer Science, Hamburg University of Applied Sciences, Germany"}]}],"member":"320","published-online":{"date-parts":[[2024,3,5]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Peeking inside the black-box: a survey on explainable artificial intelligence (XAI)","author":"Adadi Amina","year":"2018","unstructured":"Amina Adadi and Mohammed Berrada. 2018. Peeking inside the black-box: a survey on explainable artificial intelligence (XAI). IEEE access 6 (2018), 52138\u201352160."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.findings-acl.121"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.coling-main.484"},{"key":"e_1_3_2_1_4_1","volume-title":"Getting a clue: A method for explaining uncertainty estimates. arXiv preprint arXiv:2006.06848","author":"Antor\u00e1n Javier","year":"2020","unstructured":"Javier Antor\u00e1n, Umang Bhatt, Tameem Adel, Adrian Weller, and Jos\u00e9\u00a0Miguel Hern\u00e1ndez-Lobato. 2020. Getting a clue: A method for explaining uncertainty estimates. arXiv preprint arXiv:2006.06848 (2020)."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.5555\/1756006.1859912"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1145\/3461702.3462571"},{"key":"e_1_3_2_1_7_1","volume-title":"Can we open the black box of AI?Nature News 538, 7623","author":"Castelvecchi Davide","year":"2016","unstructured":"Davide Castelvecchi. 2016. Can we open the black box of AI?Nature News 538, 7623 (2016), 20."},{"key":"e_1_3_2_1_8_1","volume-title":"Aleatory or epistemic? Does it matter?Structural safety 31, 2","author":"Der\u00a0Kiureghian Armen","year":"2009","unstructured":"Armen Der\u00a0Kiureghian and Ove Ditlevsen. 2009. Aleatory or epistemic? Does it matter?Structural safety 31, 2 (2009), 105\u2013112."},{"key":"e_1_3_2_1_9_1","volume-title":"international conference on machine learning. PMLR, 1050\u20131059","author":"Gal Yarin","year":"2016","unstructured":"Yarin Gal and Zoubin Ghahramani. 2016. Dropout as a bayesian approximation: Representing model uncertainty in deep learning. In international conference on machine learning. PMLR, 1050\u20131059."},{"key":"e_1_3_2_1_10_1","volume-title":"IBreakDown: Uncertainty of model explanations for non-additive predictive models. arXiv preprint arXiv:1903.11420","author":"Gosiewska Alicja","year":"2019","unstructured":"Alicja Gosiewska and Przemyslaw Biecek. 2019. IBreakDown: Uncertainty of model explanations for non-additive predictive models. arXiv preprint arXiv:1903.11420 (2019)."},{"key":"e_1_3_2_1_11_1","volume-title":"5th International Conference on Learning Representations, ICLR.","author":"Hendrycks Dan","year":"2017","unstructured":"Dan Hendrycks and Kevin Gimpel. 2017. A baseline for detecting misclassified and out-of-distribution examples in neural networks. 5th International Conference on Learning Representations, ICLR."},{"key":"e_1_3_2_1_12_1","volume-title":"Long short-term memory. Neural computation 9, 8","author":"Hochreiter Sepp","year":"1997","unstructured":"Sepp Hochreiter and J\u00fcrgen Schmidhuber. 1997. Long short-term memory. Neural computation 9, 8 (1997), 1735\u20131780."},{"key":"e_1_3_2_1_13_1","volume-title":"Towards Benchmarking the Utility of Explanations for Model Debugging. In TrustNLP: First Workshop on Trustworthy Natural Language Processing. 68","author":"Idahl Maximilian","year":"2021","unstructured":"Maximilian Idahl, Lijun Lyu, Ujwal Gadiraju, and Avishek Anand. 2021. Towards Benchmarking the Utility of Explanations for Model Debugging. In TrustNLP: First Workshop on Trustworthy Natural Language Processing. 68."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.csda.2019.106816"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2023\/749"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.5555\/2002472.2002491"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/2939672.2939778"},{"key":"e_1_3_2_1_18_1","volume-title":"Reliable post hoc explanations: Modeling uncertainty in explainability. Advances in neural information processing systems 34","author":"Slack Dylan","year":"2021","unstructured":"Dylan Slack, Anna Hilgard, Sameer Singh, and Himabindu Lakkaraju. 2021. Reliable post hoc explanations: Modeling uncertainty in explainability. Advances in neural information processing systems 34 (2021), 9391\u20139404."},{"key":"e_1_3_2_1_19_1","volume-title":"Dropout: a simple way to prevent neural networks from overfitting. The journal of machine learning research 15, 1","author":"Srivastava Nitish","year":"2014","unstructured":"Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov. 2014. Dropout: a simple way to prevent neural networks from overfitting. The journal of machine learning research 15, 1 (2014), 1929\u20131958."},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIV.2022.3188662"}],"event":{"name":"NLPIR 2023: 2023 7th International Conference on Natural Language Processing and Information Retrieval","acronym":"NLPIR 2023","location":"Seoul Republic of Korea"},"container-title":["Proceedings of the 2023 7th International Conference on Natural Language Processing and Information Retrieval"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3639233.3639352","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3639233.3639352","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T19:56:24Z","timestamp":1755892584000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3639233.3639352"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,15]]},"references-count":20,"alternative-id":["10.1145\/3639233.3639352","10.1145\/3639233"],"URL":"https:\/\/doi.org\/10.1145\/3639233.3639352","relation":{},"subject":[],"published":{"date-parts":[[2023,12,15]]},"assertion":[{"value":"2024-03-05","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}