{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,9]],"date-time":"2025-11-09T03:45:50Z","timestamp":1762659950702},"reference-count":31,"publisher":"IEEE","license":[{"start":{"date-parts":[[2020,7,1]],"date-time":"2020-07-01T00:00:00Z","timestamp":1593561600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2020,7,1]],"date-time":"2020-07-01T00:00:00Z","timestamp":1593561600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2020,7,1]],"date-time":"2020-07-01T00:00:00Z","timestamp":1593561600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020,7]]},"DOI":"10.1109\/ijcnn48605.2020.9207686","type":"proceedings-article","created":{"date-parts":[[2020,9,30]],"date-time":"2020-09-30T00:40:33Z","timestamp":1601426433000},"page":"1-7","source":"Crossref","is-referenced-by-count":13,"title":["Adversarial Vulnerability in Doppler-based Human Activity Recognition"],"prefix":"10.1109","author":[{"given":"Zhaoyuan","family":"Yang","sequence":"first","affiliation":[]},{"given":"Yang","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Weizhong","family":"Yan","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"year":"2019","author":"virani","article-title":"Justification-based reliability in machine learning","key":"ref31"},{"year":"2018","author":"papernot","article-title":"Deep k-nearest neighbors: Towards confident, interpretable and robust deep learning","key":"ref30"},{"doi-asserted-by":"publisher","key":"ref10","DOI":"10.1109\/TEVC.2019.2890858"},{"doi-asserted-by":"publisher","key":"ref11","DOI":"10.1109\/CVPR.2018.00175"},{"doi-asserted-by":"publisher","key":"ref12","DOI":"10.1109\/TNNLS.2018.2886017"},{"year":"2017","author":"huang","article-title":"Adversarial attacks on neural network policies","key":"ref13"},{"key":"ref14","article-title":"Provable defenses against adversarial examples via the convex outer adversarial polytope","author":"kolter","year":"2017","journal-title":"CoRR"},{"doi-asserted-by":"publisher","key":"ref15","DOI":"10.1145\/2994551.2996543"},{"doi-asserted-by":"publisher","key":"ref16","DOI":"10.1109\/IJCNN.2017.7966039"},{"doi-asserted-by":"publisher","key":"ref17","DOI":"10.1109\/SP.2017.49"},{"year":"2016","author":"kurakin","article-title":"Adversarial examples in the physical world","key":"ref18"},{"year":"2017","author":"liu","article-title":"Trojaning attack on neural networks","key":"ref19"},{"year":"2017","author":"wong","article-title":"Provable defenses against adversarial examples via the convex outer adversarial polytope","key":"ref28"},{"doi-asserted-by":"publisher","key":"ref4","DOI":"10.1145\/2737095.2737117"},{"doi-asserted-by":"publisher","key":"ref27","DOI":"10.1109\/ACCESS.2019.2941376"},{"doi-asserted-by":"publisher","key":"ref3","DOI":"10.1109\/CVPR.2018.00768"},{"doi-asserted-by":"publisher","key":"ref6","DOI":"10.1145\/3241539.3241548"},{"year":"2017","author":"madry","article-title":"Towards deep learning models resistant to adversarial attacks","key":"ref29"},{"doi-asserted-by":"publisher","key":"ref5","DOI":"10.1145\/3310194"},{"year":"2015","author":"goodfellow","article-title":"Explaining and harnessing adversarial examples","key":"ref8"},{"key":"ref7","article-title":"Intriguing properties of neural networks","author":"szegedy","year":"2013","journal-title":"CoRR"},{"doi-asserted-by":"publisher","key":"ref2","DOI":"10.1145\/2809695.2817882"},{"doi-asserted-by":"publisher","key":"ref9","DOI":"10.1109\/CVPR.2017.17"},{"doi-asserted-by":"publisher","key":"ref1","DOI":"10.1109\/TII.2017.2782213"},{"key":"ref20","article-title":"Badnets: Identifying vulnerabilities in the machine learning model supply chain","author":"gu","year":"2017","journal-title":"CoRR"},{"year":"2019","author":"fawaz","article-title":"Adversarial attacks on deep neural networks for time series classification","key":"ref22"},{"doi-asserted-by":"publisher","key":"ref21","DOI":"10.1109\/MILCOM.2016.7795300"},{"year":"2019","author":"team","journal-title":"Keras Deep Learning for humans","key":"ref24"},{"key":"ref23","article-title":"Adversarial examples in the physical world","author":"kurakin","year":"2016","journal-title":"CoRR"},{"year":"2018","author":"bagdasaryan","article-title":"How to backdoor federated learning","key":"ref26"},{"year":"2018","author":"nicolae","article-title":"Adversarial robustness toolbox v0. 4.0","key":"ref25"}],"event":{"name":"2020 International Joint Conference on Neural Networks (IJCNN)","start":{"date-parts":[[2020,7,19]]},"location":"Glasgow, United Kingdom","end":{"date-parts":[[2020,7,24]]}},"container-title":["2020 International Joint Conference on Neural Networks (IJCNN)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9200848\/9206590\/09207686.pdf?arnumber=9207686","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,6,28]],"date-time":"2022-06-28T21:52:26Z","timestamp":1656453146000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9207686\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,7]]},"references-count":31,"URL":"https:\/\/doi.org\/10.1109\/ijcnn48605.2020.9207686","relation":{},"subject":[],"published":{"date-parts":[[2020,7]]}}}