{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,9,7]],"date-time":"2024-09-07T16:01:19Z","timestamp":1725724879837},"reference-count":36,"publisher":"IEEE","license":[{"start":{"date-parts":[[2021,7,18]],"date-time":"2021-07-18T00:00:00Z","timestamp":1626566400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2021,7,18]],"date-time":"2021-07-18T00:00:00Z","timestamp":1626566400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2021,7,18]],"date-time":"2021-07-18T00:00:00Z","timestamp":1626566400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021,7,18]]},"DOI":"10.1109\/ijcnn52387.2021.9534397","type":"proceedings-article","created":{"date-parts":[[2021,9,23]],"date-time":"2021-09-23T22:32:08Z","timestamp":1632436328000},"page":"1-8","source":"Crossref","is-referenced-by-count":2,"title":["Luring Transferable Adversarial Perturbations for Deep Neural Networks"],"prefix":"10.1109","author":[{"given":"Remi","family":"Bernhard","sequence":"first","affiliation":[]},{"given":"Pierre-Alain","family":"Moellic","sequence":"additional","affiliation":[]},{"given":"Jean-Max","family":"Dutertre","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00444"},{"key":"ref32","article-title":"Improving transferability of adversarial examples with input diversity","author":"xie","year":"2018","journal-title":"2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition"},{"key":"ref31","first-page":"4636","article-title":"Parsimonious black-box adversarial attacks via efficient combinatorial optimization","author":"moon","year":"2019","journal-title":"Proceedings of the 36th International Conference on Machine Learning"},{"key":"ref30","article-title":"Very deep convolutional networks for large-scale image recognition","author":"simonyan","year":"2015","journal-title":"International Conference on Learning Representations"},{"key":"ref36","article-title":"Using honeypots to catch adversarial attacks on neural networks","author":"shan","year":"2019","journal-title":"Proceedings of ACM Conference on Computer and Communications Security (CCS)"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00474"},{"key":"ref34","article-title":"Obfuscated gradients give a false sense of security: Circumventing defenses to adversarial examples","author":"athalye","year":"2018","journal-title":"Proceedings of the 35th International Conference on Machine Learning ICML 2018"},{"key":"ref10","article-title":"Unlabeled data improves adversarial robustness","author":"carmon","year":"2019","journal-title":"Advances in neural information processing systems"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1145\/3052973.3053009"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1145\/3133956.3134057"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2019.2939352"},{"key":"ref14","article-title":"Explaining and harnessing adversarial examples","author":"goodfellow","year":"2015","journal-title":"International Conference on Learning Representations"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/SP.2017.49"},{"key":"ref16","article-title":"Ead: elastic-net attacks to deep neural networks via adversarial examples","author":"chen","year":"2018","journal-title":"Thirty-Second AAAI Conf Artif Intell"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00957"},{"key":"ref18","article-title":"Enhancing gradient-based attacks with symbolic intervals","author":"wang","year":"2019","journal-title":"Proceedings of the 36th International Conference on Machine Learning ICML 2019"},{"key":"ref19","first-page":"5025","article-title":"Adversarial risk and the dangers of evaluating against weak attacks","author":"uesato","year":"2018","journal-title":"Proceedings of the 35th International Conference on Machine Learning ICML 2018"},{"key":"ref28","article-title":"Reading digits in natural images with unsupervised feature learning","author":"netzer","year":"2011","journal-title":"NIPS 2011 Workshop on Deep Learning and Unsupervised Feature Learning"},{"key":"ref4","first-page":"125","article-title":"Adversarial examples are not bugs, they are features","author":"ilyas","year":"2019","journal-title":"Advances in neural information processing systems"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/5.726791"},{"key":"ref3","first-page":"2280","article-title":"Adversarial examples are a natural consequence of test error in noise","author":"ford","year":"2019","journal-title":"Proceedings of the ICML International Conference on Machine Learning"},{"key":"ref6","article-title":"Towards deep learning models resistant to adversarial attacks","author":"madry","year":"2018","journal-title":"International Conference on Learning Representations"},{"key":"ref29","article-title":"Learning multiple layers of features from tiny images","author":"krizhevsky","year":"2009","journal-title":"Tech Rep"},{"key":"ref5","article-title":"Are adversarial examples inevitable?","author":"shafahi","year":"2019","journal-title":"International Conference on Learning Representations"},{"key":"ref8","first-page":"1310","article-title":"Certified adversarial robustness via randomized smoothing","author":"cohen","year":"2019","journal-title":"Proceedings of the 36th International Conference on Machine Learning ICML 2019"},{"key":"ref7","article-title":"Theoretically principled trade-off between robustness and accuracy","author":"zhang","year":"2019","journal-title":"Proceedings of the 36th International Conference on Machine Learning ICML 2019"},{"key":"ref2","first-page":"5014","article-title":"Adver-sarially robust generalization requires more data","author":"schmidt","year":"2018","journal-title":"Advances in neural information processing systems"},{"key":"ref9","article-title":"Using pre-training can improve model robustness and uncertainty","author":"hendrycks","year":"2019","journal-title":"Proceedings of the 36th International Conference on Machine Learning ICML 2019"},{"key":"ref1","article-title":"Intriguing properties of neural networks","author":"szegedy","year":"2014","journal-title":"International Conference on Learning Representations"},{"key":"ref20","article-title":"Simple black-box adversarial attacks","author":"guo","year":"2019","journal-title":"Proceedings of the 36th International Conference on Machine Learning ICML 2019"},{"key":"ref22","article-title":"Decision-based adversarial attacks: Reliable attacks against black-box machine learning models","author":"brendel","year":"2018","journal-title":"International Conference on Learning Representations"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/TEVC.2019.2890858"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/SP40000.2020.00045"},{"key":"ref23","article-title":"Black-box adversarial attacks with limited queries and information","author":"ilyas","year":"2018","journal-title":"Proceedings of the 35th International Conference on Machine Learning ICML 2018"},{"journal-title":"Blacklight Defending black-box adversarial attacks on deep neural networks","year":"2020","author":"li","key":"ref26"},{"journal-title":"Stateful detection of black-box adversarial attacks","year":"2019","author":"chen","key":"ref25"}],"event":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","start":{"date-parts":[[2021,7,18]]},"location":"Shenzhen, China","end":{"date-parts":[[2021,7,22]]}},"container-title":["2021 International Joint Conference on Neural Networks (IJCNN)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9533266\/9533267\/09534397.pdf?arnumber=9534397","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,5,10]],"date-time":"2022-05-10T15:46:03Z","timestamp":1652197563000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9534397\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,7,18]]},"references-count":36,"URL":"https:\/\/doi.org\/10.1109\/ijcnn52387.2021.9534397","relation":{},"subject":[],"published":{"date-parts":[[2021,7,18]]}}}