{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,24]],"date-time":"2025-11-24T07:14:44Z","timestamp":1763968484189,"version":"3.37.3"},"reference-count":25,"publisher":"IEEE","license":[{"start":{"date-parts":[[2021,12,1]],"date-time":"2021-12-01T00:00:00Z","timestamp":1638316800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2021,12,1]],"date-time":"2021-12-01T00:00:00Z","timestamp":1638316800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/100006602","name":"AFRL","doi-asserted-by":"publisher","award":["AFRL-2020-0311"],"award-info":[{"award-number":["AFRL-2020-0311"]}],"id":[{"id":"10.13039\/100006602","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021,12]]},"DOI":"10.1109\/globecom46510.2021.9685101","type":"proceedings-article","created":{"date-parts":[[2022,2,2]],"date-time":"2022-02-02T21:59:04Z","timestamp":1643839144000},"page":"1-6","source":"Crossref","is-referenced-by-count":4,"title":["Active Learning Under Malicious Mislabeling and Poisoning Attacks"],"prefix":"10.1109","author":[{"given":"Jing","family":"Lin","sequence":"first","affiliation":[{"name":"ICNS Lab and Cyber Florida University of South Florida,Tampa,FL,USA"}]},{"given":"Ryan","family":"Luley","sequence":"additional","affiliation":[{"name":"High Performance Systems Branch U.S. Air Force Research Laboratory,Rome,New York,USA"}]},{"given":"Kaiqi","family":"Xiong","sequence":"additional","affiliation":[{"name":"ICNS Lab and Cyber Florida University of South Florida,Tampa,FL,USA"}]}],"member":"263","reference":[{"key":"ref10","article-title":"Adversarial active learning for deep networks: a margin based approach","author":"ducoffe","year":"2018","journal-title":"ArXiv Preprint"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW50498.2020.00402"},{"key":"ref12","article-title":"Transferable clean-label poisoning attacks on deep neural nets","author":"zhu","year":"2019","journal-title":"ArXiv Preprint"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-014-0781-x"},{"key":"ref14","article-title":"Active learning for convolutional neural networks: A core-set approach","author":"sener","year":"2017","journal-title":"ArXiv Preprint"},{"key":"ref15","article-title":"Deep batch active learning by diverse, uncertain gradient lower bounds","author":"ash","year":"0","journal-title":"International Conference on Learning Representations"},{"key":"ref16","article-title":"Poison Frogs! Targeted Clean-Label Poisoning Attacks on Neural Networks","author":"shafahi","year":"2018","journal-title":"ArXiv e-prints"},{"key":"ref17","article-title":"Cifar10 to compare visual recognition performance between deep neural networks and humans","author":"ho-phuoc","year":"2018","journal-title":"ArXiv Preprint"},{"journal-title":"Learning multiple layers of features from tiny images","year":"2009","author":"krizhevsky","key":"ref18"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2018.2807385"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2018.2886017"},{"key":"ref3","article-title":"Guarantees on learning depth-2 neural networks under a data-poisoning attack","author":"mukherjee","year":"2020","journal-title":"ArXiv Preprint"},{"journal-title":"Active Learning Literature Survey","year":"2009","author":"settles","key":"ref6"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1145\/1015330.1015349"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206627"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.565"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN.2017.7966180"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN.2014.6889457"},{"journal-title":"How Much Data Do We Create Every Day? The Mind-Blowing Stats Everyone Should Read","year":"0","author":"lackey","key":"ref1"},{"key":"ref20","article-title":"A tale of evil twins: Adversarial inputs versus poisoned mod-els","author":"pang","year":"0","journal-title":"ACM Conference on Computer and Communications Security"},{"key":"ref22","article-title":"Adversarial robustness toolbox v0.10.0","volume":"1807 1069","author":"nicolae","year":"2018","journal-title":"CoRR"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref24","first-page":"1","article-title":"On defending against label flipping attacks on malware detection systems","author":"taheri","year":"2020","journal-title":"Neural Computing and Applications"},{"key":"ref23","article-title":"Towards deep learning models resistant to adversarial attacks","author":"madry","year":"2017","journal-title":"ArXiv Preprint"},{"key":"ref25","first-page":"5","article-title":"Label sanitization against label flipping poisoning attacks","author":"paudice","year":"0","journal-title":"European Conference on Machine Learning and Knowledge Discovery in Databases"}],"event":{"name":"GLOBECOM 2021 - 2021 IEEE Global Communications Conference","start":{"date-parts":[[2021,12,7]]},"location":"Madrid, Spain","end":{"date-parts":[[2021,12,11]]}},"container-title":["2021 IEEE Global Communications Conference (GLOBECOM)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9685019\/9685006\/09685101.pdf?arnumber=9685101","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,12,7]],"date-time":"2022-12-07T00:54:07Z","timestamp":1670374447000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9685101\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,12]]},"references-count":25,"URL":"https:\/\/doi.org\/10.1109\/globecom46510.2021.9685101","relation":{},"subject":[],"published":{"date-parts":[[2021,12]]}}}