{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:31:25Z","timestamp":1750221085923,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":10,"publisher":"ACM","license":[{"start":{"date-parts":[[2018,10,15]],"date-time":"2018-10-15T00:00:00Z","timestamp":1539561600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2018,10,15]]},"DOI":"10.1145\/3243734.3278486","type":"proceedings-article","created":{"date-parts":[[2018,11,13]],"date-time":"2018-11-13T13:39:38Z","timestamp":1542116378000},"page":"2246-2248","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":2,"title":["Spartan Networks"],"prefix":"10.1145","author":[{"given":"Fran\u00e7ois","family":"Menet","sequence":"first","affiliation":[{"name":"Polytechnique Montr\u00e9al, Montr\u00e9al, PQ, Canada"}]},{"given":"Paul","family":"Berthier","sequence":"additional","affiliation":[{"name":"Polytechnique Montr\u00e9al, Montr\u00e9al, PQ, Canada"}]},{"given":"Michel","family":"Gagnon","sequence":"additional","affiliation":[{"name":"Polytechnique Montr\u00e9al, Montr\u00e9al, PQ, Canada"}]},{"given":"Jos\u00e9 M.","family":"Fernandez","sequence":"additional","affiliation":[{"name":"Polytechnique Montr\u00e9al, Montr\u00e9al, PQ, Canada"}]}],"member":"320","published-online":{"date-parts":[[2018,10,15]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"Jacob Buckman Aurko Roy Colin Raffel and Ian Goodfellow. 2018. Thermometer Encoding: one hot way to resist adversarial examples. (2018) 22.  Jacob Buckman Aurko Roy Colin Raffel and Ian Goodfellow. 2018. Thermometer Encoding: one hot way to resist adversarial examples. (2018) 22."},{"key":"e_1_3_2_1_2_1","volume-title":"Binarized Neural Networks: Training Deep Neural Networks with Weights and Activations Constrained to +1 or -1. arXiv:1602.02830 {cs} (Feb","author":"Courbariaux Matthieu","year":"2016","unstructured":"Matthieu Courbariaux , Itay Hubara , Daniel Soudry , Ran El-Yaniv , and Yoshua Bengio . 2016. Binarized Neural Networks: Training Deep Neural Networks with Weights and Activations Constrained to +1 or -1. arXiv:1602.02830 {cs} (Feb . 2016 ). arXiv: 1602.02830. Matthieu Courbariaux, Itay Hubara, Daniel Soudry, Ran El-Yaniv, and Yoshua Bengio. 2016. Binarized Neural Networks: Training Deep Neural Networks with Weights and Activations Constrained to +1 or -1. arXiv:1602.02830 {cs} (Feb. 2016). arXiv: 1602.02830."},{"key":"e_1_3_2_1_3_1","unstructured":"Ian J. Goodfellow Jonathon Shlens and Christian Szegedy. 2014. Explaining and Harnessing Adversarial Examples. In arXiv:1412.6572 {cs stat}. arXiv: 1412.6572.  Ian J. Goodfellow Jonathon Shlens and Christian Szegedy. 2014. Explaining and Harnessing Adversarial Examples. In arXiv:1412.6572 {cs stat}. arXiv: 1412.6572."},{"key":"e_1_3_2_1_4_1","volume-title":"Safetynet: Detecting and rejecting adversarial examples robustly. CoRR, abs\/1704.00103","author":"Lu Jiajun","year":"2017","unstructured":"Jiajun Lu , Theerasit Issaranon , and David Forsyth . 2017 . Safetynet: Detecting and rejecting adversarial examples robustly. CoRR, abs\/1704.00103 (2017). Jiajun Lu, Theerasit Issaranon, and David Forsyth. 2017. Safetynet: Detecting and rejecting adversarial examples robustly. CoRR, abs\/1704.00103 (2017)."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1145\/3133956.3134057"},{"key":"e_1_3_2_1_6_1","unstructured":"A. Nayebi and S. Ganguli. 2017. Biologically inspired protection of deep networks from adversarial attacks. ArXiv e-prints (March 2017). arXiv:stat.ML\/1703.09202  A. Nayebi and S. Ganguli. 2017. Biologically inspired protection of deep networks from adversarial attacks. ArXiv e-prints (March 2017). arXiv:stat.ML\/1703.09202"},{"key":"e_1_3_2_1_7_1","volume-title":"Distillation as a Defense to Adversarial Perturbations against Deep Neural Networks. arXiv:1511.04508 {cs, stat} (Nov","author":"Papernot Nicolas","year":"2015","unstructured":"Nicolas Papernot , Patrick McDaniel , Xi Wu , Somesh Jha , and Ananthram Swami . 2015. Distillation as a Defense to Adversarial Perturbations against Deep Neural Networks. arXiv:1511.04508 {cs, stat} (Nov . 2015 ). http:\/\/arxiv.org\/abs\/1511.04508 arXiv: 1511.04508. Nicolas Papernot, Patrick McDaniel, Xi Wu, Somesh Jha, and Ananthram Swami. 2015. Distillation as a Defense to Adversarial Perturbations against Deep Neural Networks. arXiv:1511.04508 {cs, stat} (Nov. 2015). http:\/\/arxiv.org\/abs\/1511.04508 arXiv: 1511.04508."},{"key":"e_1_3_2_1_8_1","volume-title":"Improving the Adversarial Robustness and Interpretability of Deep Neural Networks by Regularizing their Input Gradients. arXiv:1711.09404 {cs} (Nov","author":"Ross Andrew Slavin","year":"2017","unstructured":"Andrew Slavin Ross and Finale Doshi-Velez . 2017. Improving the Adversarial Robustness and Interpretability of Deep Neural Networks by Regularizing their Input Gradients. arXiv:1711.09404 {cs} (Nov . 2017 ). arXiv: 1711.09404. Andrew Slavin Ross and Finale Doshi-Velez. 2017. Improving the Adversarial Robustness and Interpretability of Deep Neural Networks by Regularizing their Input Gradients. arXiv:1711.09404 {cs} (Nov. 2017). arXiv: 1711.09404."},{"key":"e_1_3_2_1_9_1","volume-title":"Intriguing properties of neural networks. arXiv:1312.6199 {cs} (Dec","author":"Szegedy Christian","year":"2013","unstructured":"Christian Szegedy , Wojciech Zaremba , Ilya Sutskever , Joan Bruna , Dumitru Erhan , Ian Goodfellow , and Rob Fergus . 2013. Intriguing properties of neural networks. arXiv:1312.6199 {cs} (Dec . 2013 ). arXiv: 1312.6199. Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, and Rob Fergus. 2013. Intriguing properties of neural networks. arXiv:1312.6199 {cs} (Dec. 2013). arXiv: 1312.6199."},{"key":"e_1_3_2_1_10_1","volume-title":"Feature squeezing: Detecting adversarial examples in deep neural networks. arXiv preprint arXiv:1704.01155","author":"Xu Weilin","year":"2017","unstructured":"Weilin Xu , David Evans , and Yanjun Qi. 2017. Feature squeezing: Detecting adversarial examples in deep neural networks. arXiv preprint arXiv:1704.01155 ( 2017 ). Weilin Xu, David Evans, and Yanjun Qi. 2017. Feature squeezing: Detecting adversarial examples in deep neural networks. arXiv preprint arXiv:1704.01155 (2017)."}],"event":{"name":"CCS '18: 2018 ACM SIGSAC Conference on Computer and Communications Security","sponsor":["SIGSAC ACM Special Interest Group on Security, Audit, and Control"],"location":"Toronto Canada","acronym":"CCS '18"},"container-title":["Proceedings of the 2018 ACM SIGSAC Conference on Computer and Communications Security"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3243734.3278486","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3243734.3278486","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T00:57:47Z","timestamp":1750208267000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3243734.3278486"}},"subtitle":["Self-Feature-Squeezing Networks for Increased Robustness in Adversarial Settings"],"short-title":[],"issued":{"date-parts":[[2018,10,15]]},"references-count":10,"alternative-id":["10.1145\/3243734.3278486","10.1145\/3243734"],"URL":"https:\/\/doi.org\/10.1145\/3243734.3278486","relation":{},"subject":[],"published":{"date-parts":[[2018,10,15]]},"assertion":[{"value":"2018-10-15","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}