{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:26:09Z","timestamp":1750220769545,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":22,"publisher":"ACM","license":[{"start":{"date-parts":[[2020,2,19]],"date-time":"2020-02-19T00:00:00Z","timestamp":1582070400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/501100010418","name":"Institute for Information and Communications Technology Promotion","doi-asserted-by":"publisher","award":["2018-0-00420, 2019-0-00426"],"award-info":[{"award-number":["2018-0-00420, 2019-0-00426"]}],"id":[{"id":"10.13039\/501100010418","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003725","name":"National Research Foundation of Korea","doi-asserted-by":"publisher","award":["2017R1C1B2003957, 2017R1A2B4006026"],"award-info":[{"award-number":["2017R1C1B2003957, 2017R1A2B4006026"]}],"id":[{"id":"10.13039\/501100003725","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2020,2,19]]},"DOI":"10.1145\/3385209.3385216","type":"proceedings-article","created":{"date-parts":[[2020,6,7]],"date-time":"2020-06-07T00:35:48Z","timestamp":1591490148000},"page":"140-145","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":4,"title":["TargetNet Backdoor"],"prefix":"10.1145","author":[{"given":"Hyun","family":"Kwon","sequence":"first","affiliation":[{"name":"Korea Advanced Institute of Science and Technology, School of Computing, Daejeon, South Korea"}]},{"given":"Jungmin","family":"Roh","sequence":"additional","affiliation":[{"name":"Ministry of National Defense, ROK Army Training &amp; Doctrine Command, Daejeon, South Korea"}]},{"given":"Hyunsoo","family":"Yoon","sequence":"additional","affiliation":[{"name":"Korea Advanced Institute of Science and Technology, School of Computing, Daejeon, South Korea"}]},{"given":"Ki-Woong","family":"Park","sequence":"additional","affiliation":[{"name":"Sejong University Computer &amp; Information Security, Seoul, South Korea"}]}],"member":"320","published-online":{"date-parts":[[2020,6,6]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2014.09.003"},{"key":"e_1_3_2_1_2_1","volume-title":"Intriguing properties of neural networks,\" in International Conference on Learning Representations","author":"Szegedy C.","year":"2014","unstructured":"C. Szegedy, W. Zaremba, I. Sutskever, J. Bruna, D. Erhan, I. Goodfellow, and R. Fergus, \"Intriguing properties of neural networks,\" in International Conference on Learning Representations, 2014."},{"key":"e_1_3_2_1_3_1","first-page":"1467","volume-title":"Poisoning attacks against support vector machines,\" in Proceedings of the 29th International Coference on International Conference on Machine Learning","author":"Biggio B.","year":"2012","unstructured":"B. Biggio, B. Nelson, and P. Laskov, \"Poisoning attacks against support vector machines,\" in Proceedings of the 29th International Coference on International Conference on Machine Learning, pp. 1467--1474, Omnipress, 2012."},{"key":"e_1_3_2_1_4_1","volume-title":"Badnets: Identifying vulnerabilities in the machine learning model supply chain,\" arXiv preprint arXiv:1708.06733","author":"Gu T.","year":"2017","unstructured":"T. Gu, B. Dolan-Gavitt, and S. Garg, \"Badnets: Identifying vulnerabilities in the machine learning model supply chain,\" arXiv preprint arXiv:1708.06733, 2017."},{"key":"e_1_3_2_1_5_1","volume-title":"Mnist handwritten digit database,\" AT&T Labs [Online]. Available: http:\/\/yann.lecun.com\/exdb\/mnist","author":"LeCun Y.","year":"2010","unstructured":"Y. LeCun, C. Cortes, and C. J. Burges, \"Mnist handwritten digit database,\" AT&T Labs [Online]. Available: http:\/\/yann.lecun.com\/exdb\/mnist, vol. 2, 2010."},{"key":"e_1_3_2_1_6_1","volume-title":"Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms,\" arXiv preprint arXiv:1708.07747","author":"Xiao H.","year":"2017","unstructured":"H. Xiao, K. Rasul, and R. Vollgraf, \"Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms,\" arXiv preprint arXiv:1708.07747, 2017."},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10994-010-5188-5"},{"key":"e_1_3_2_1_8_1","volume-title":"Explaining and harnessing adversarial examples,\" in International Conference on Learning Representations","author":"Goodfellow I.","year":"2015","unstructured":"I. Goodfellow, J. Shlens, and C. Szegedy, \"Explaining and harnessing adversarial examples,\" in International Conference on Learning Representations, 2015."},{"key":"e_1_3_2_1_9_1","volume-title":"Adversarial examples in the physical world,\" ICLR Workshop","author":"Kurakin A.","year":"2017","unstructured":"A. Kurakin, I. Goodfellow, and S. Bengio, \"Adversarial examples in the physical world,\" ICLR Workshop, 2017."},{"key":"e_1_3_2_1_10_1","first-page":"2574","article-title":"Deepfool: a simple and accurate method to fool deep neural networks","author":"Moosavi-Dezfooli S.-M.","year":"2016","unstructured":"S.-M. Moosavi-Dezfooli, A. Fawzi, and P. Frossard, \"Deepfool: a simple and accurate method to fool deep neural networks,\" in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2574--2582, 2016.","journal-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition"},{"key":"e_1_3_2_1_11_1","first-page":"372","volume-title":"2016 IEEE European Symposium on","author":"Papernot N.","year":"2016","unstructured":"N. Papernot, P. McDaniel, S. Jha, M. Fredrikson, Z. B. Celik, and A. Swami, \"The limitations of deep learning in adversarial settings,\" in Security and Privacy (EuroS&P), 2016 IEEE European Symposium on, pp. 372--387, IEEE, 2016."},{"key":"e_1_3_2_1_12_1","first-page":"39","volume-title":"2017 IEEE Symposium on","author":"Carlini N.","year":"2017","unstructured":"N. Carlini and D. Wagner, \"Towards evaluating the robustness of neural networks,\" in Security and Privacy (SP), 2017 IEEE Symposium on, pp. 39--57, IEEE, 2017."},{"key":"e_1_3_2_1_13_1","volume-title":"Generative poisoning attack method against neural networks,\" arXiv preprint arXiv:1703.01340","author":"Yang C.","year":"2017","unstructured":"C. Yang, Q. Wu, H. Li, and Y. Chen, \"Generative poisoning attack method against neural networks,\" arXiv preprint arXiv:1703.01340, 2017."},{"volume-title":"Systematic poisoning attacks on and defenses for machine learning in healthcare,\" IEEE journal of biomedical and health informatics","author":"Mozaffari-Kermani M.","key":"e_1_3_2_1_14_1","unstructured":"M. Mozaffari-Kermani, S. Sur-Kolay, A. Raghunathan, and N. K. Jha, \"Systematic poisoning attacks on and defenses for machine learning in healthcare,\" IEEE journal of biomedical and health informatics, vol. 19, no. 6, pp. 1893--1905, 2015."},{"key":"e_1_3_2_1_15_1","volume-title":"Trojaning attack on neural networks,\" NDSS","author":"Liu Y.","year":"2018","unstructured":"Y. Liu, S. Ma, Y. Aafer, W.-C. Lee, J. Zhai, W. Wang, and X. Zhang, \"Trojaning attack on neural networks,\" NDSS, 2018."},{"key":"e_1_3_2_1_16_1","first-page":"0","article-title":"Neural cleanse: Identifying and mitigating backdoor attacks in neural networks","author":"Wang B.","year":"2019","unstructured":"B. Wang, Y. Yao, S. Shan, H. Li, B. Viswanath, H. Zheng, and B. Y. Zhao, \"Neural cleanse: Identifying and mitigating backdoor attacks in neural networks,\" Neural Cleanse: Identifying and Mitigating Backdoor Attacks in Neural Networks, p. 0, 2019.","journal-title":"Neural Cleanse: Identifying and Mitigating Backdoor Attacks in Neural Networks"},{"key":"e_1_3_2_1_17_1","volume-title":"Hardware trojan attacks on neural networks,\" arXiv preprint arXiv:1806.05768","author":"Clements J.","year":"2018","unstructured":"J. Clements and Y. Lao, \"Hardware trojan attacks on neural networks,\" arXiv preprint arXiv:1806.05768, 2018."},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1587\/transinf.2019EDL8170"},{"key":"e_1_3_2_1_19_1","first-page":"53","volume-title":"Indentifying backdoor attack that is safe for friendly deep neural network,\" in Proceedings of the 3rd International Conference on Software Engineering and Information Management","author":"Kwon H.","year":"2020","unstructured":"H. Kwon, H. Yoon, and K.-W. Park, \"Friendnet backdoor: Indentifying backdoor attack that is safe for friendly deep neural network,\" in Proceedings of the 3rd International Conference on Software Engineering and Information Management, pp. 53--57, 2020."},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1109\/5.726791"},{"key":"e_1_3_2_1_21_1","volume-title":"Adam: A method for stochastic optimization,\" The International Conference on Learning Representations (ICLR)","author":"Kingma D.","year":"2015","unstructured":"D. Kingma and J. Ba, \"Adam: A method for stochastic optimization,\" The International Conference on Learning Representations (ICLR), 2015."},{"key":"e_1_3_2_1_22_1","first-page":"265","article-title":"Tensorflow: A system for large-scale machine learning","volume":"16","author":"Abadi M.","year":"2016","unstructured":"M. Abadi, P. Barham, J. Chen, Z. Chen, A. Davis, J. Dean, M. Devin, S. Ghemawat, G. Irving, M. Isard, et al., \"Tensorflow: A system for large-scale machine learning.,\" in OSDI, vol. 16, pp. 265--283, 2016.","journal-title":"OSDI"}],"event":{"name":"ICIIT 2020: 2020 5th International Conference on Intelligent Information Technology","acronym":"ICIIT 2020","location":"Hanoi Viet Nam"},"container-title":["Proceedings of the 2020 5th International Conference on Intelligent Information Technology"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3385209.3385216","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3385209.3385216","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T22:41:14Z","timestamp":1750200074000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3385209.3385216"}},"subtitle":["Attack on Deep Neural Network with Use of Different Triggers"],"short-title":[],"issued":{"date-parts":[[2020,2,19]]},"references-count":22,"alternative-id":["10.1145\/3385209.3385216","10.1145\/3385209"],"URL":"https:\/\/doi.org\/10.1145\/3385209.3385216","relation":{},"subject":[],"published":{"date-parts":[[2020,2,19]]},"assertion":[{"value":"2020-06-06","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}