{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,27]],"date-time":"2025-08-27T16:18:23Z","timestamp":1756311503044,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":14,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,5,16]],"date-time":"2022-05-16T00:00:00Z","timestamp":1652659200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,5,16]]},"DOI":"10.1145\/3522664.3528617","type":"proceedings-article","created":{"date-parts":[[2022,10,17]],"date-time":"2022-10-17T16:30:14Z","timestamp":1666024214000},"page":"39-40","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Identification of out-of-distribution cases of CNN using class-based surprise adequacy"],"prefix":"10.1145","author":[{"given":"Mira","family":"Marhaba","sequence":"first","affiliation":[{"name":"Polytechnique Montreal, Montreal, Canada"}]},{"given":"Ettore","family":"Merlo","sequence":"additional","affiliation":[{"name":"Polytechnique Montreal, Montreal, Canada"}]},{"given":"Foutse","family":"Khomh","sequence":"additional","affiliation":[{"name":"Polytechnique Montreal, Montreal, Canada"}]},{"given":"Giuliano","family":"Antoniol","sequence":"additional","affiliation":[{"name":"Polytechnique Montreal, Montreal, Canada"}]}],"member":"320","published-online":{"date-parts":[[2022,10,17]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"crossref","unstructured":"Nicholas Carlini and David Wagner. 2017. Towards Evaluating the Robustness of Neural Networks. arXiv:1608.04644 [cs.CR]","DOI":"10.1109\/SP.2017.49"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1145\/3143561"},{"key":"e_1_3_2_1_3_1","volume-title":"RAID: Randomized Adversarial-Input Detection for Neural Networks. arXiv e-prints, Article arXiv:2002.02776 (Feb.","author":"Eniser Hasan Ferit","year":"2020","unstructured":"Hasan Ferit Eniser, Maria Christakis, and Valentin W\u00fcstholz. 2020. RAID: Randomized Adversarial-Input Detection for Neural Networks. arXiv e-prints, Article arXiv:2002.02776 (Feb. 2020), arXiv:2002.02776 pages. arXiv:2002.02776 [cs.LG]"},{"key":"e_1_3_2_1_4_1","unstructured":"Ian J. Goodfellow Jonathon Shlens and Christian Szegedy. 2015. Explaining and Harnessing Adversarial Examples. arXiv:1412.6572 [stat.ML]"},{"key":"e_1_3_2_1_5_1","unstructured":"IBM-ART. 2022. Adversarial Robustness 360 Toolbox. https:\/\/developer.ibm.com\/technologies\/analytics\/projects\/adversarial-robustness-toolbox"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICSE.2019.00108"},{"key":"e_1_3_2_1_7_1","unstructured":"MNIST. 2022. MNIST-fashion. https:\/\/github.com\/zalandoresearch\/fashion-mnist."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"crossref","unstructured":"Seyed-Mohsen Moosavi-Dezfooli Alhussein Fawzi and Pascal Frossard. 2016. DeepFool: a simple and accurate method to fool deep neural networks. arXiv:1511.04599 [cs.LG]","DOI":"10.1109\/CVPR.2016.282"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"crossref","unstructured":"Nicolas Papernot Patrick McDaniel Somesh Jha Matt Fredrikson Z. Berkay Celik and Ananthram Swami. 2015. The Limitations of Deep Learning in Adversarial Settings. arXiv:1511.07528 [cs.CR]","DOI":"10.1109\/EuroSP.2016.36"},{"key":"e_1_3_2_1_10_1","volume-title":"Distillation as a Defense to Adversarial Perturbations against Deep Neural Networks. CoRR abs\/1511.04508","author":"Papernot Nicolas","year":"2015","unstructured":"Nicolas Papernot, Patrick D. McDaniel, Xi Wu, Somesh Jha, and Ananthram Swami. 2015. Distillation as a Defense to Adversarial Perturbations against Deep Neural Networks. CoRR abs\/1511.04508 (2015). arXiv:1511.04508 http:\/\/arxiv.org\/abs\/1511.04508"},{"key":"e_1_3_2_1_11_1","unstructured":"TensorFlow. 2022. TensorFlow - LeNet. https:\/\/github.com\/tensorflow\/models\/blob\/master\/research\/slim\/nets\/lenet.py"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.5555\/3524938.3525824"},{"volume-title":"Proceedings of the 6th International Conference on Learning Representations (ICLR).","author":"Tram\u00e8r Florian","key":"e_1_3_2_1_13_1","unstructured":"Florian Tram\u00e8r, A. Kurakin, Nicolas Papernot, D. Boneh, and P. McDaniel. 2018. Ensemble Adversarial Training: Attacks and Defenses. In Proceedings of the 6th International Conference on Learning Representations (ICLR)."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2018.2886017"}],"event":{"name":"CAIN '22: 1st Conference on AI Engineering - Software Engineering for AI","sponsor":["SIGSOFT ACM Special Interest Group on Software Engineering","IEEE TCSC IEEE Technical Committee on Scalable Computing"],"location":"Pittsburgh Pennsylvania","acronym":"CAIN '22"},"container-title":["Proceedings of the 1st International Conference on AI Engineering: Software Engineering for AI"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3522664.3528617","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3522664.3528617","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T18:09:34Z","timestamp":1750183774000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3522664.3528617"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,5,16]]},"references-count":14,"alternative-id":["10.1145\/3522664.3528617","10.1145\/3522664"],"URL":"https:\/\/doi.org\/10.1145\/3522664.3528617","relation":{},"subject":[],"published":{"date-parts":[[2022,5,16]]},"assertion":[{"value":"2022-10-17","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}