{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,18]],"date-time":"2025-12-18T14:10:56Z","timestamp":1766067056311,"version":"3.28.0"},"reference-count":39,"publisher":"IEEE","license":[{"start":{"date-parts":[[2019,7,1]],"date-time":"2019-07-01T00:00:00Z","timestamp":1561939200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2019,7,1]],"date-time":"2019-07-01T00:00:00Z","timestamp":1561939200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2019,7,1]],"date-time":"2019-07-01T00:00:00Z","timestamp":1561939200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019,7]]},"DOI":"10.1109\/iolts.2019.8854377","type":"proceedings-article","created":{"date-parts":[[2019,10,3]],"date-time":"2019-10-03T20:27:48Z","timestamp":1570134468000},"page":"182-187","source":"Crossref","is-referenced-by-count":28,"title":["QuSecNets: Quantization-based Defense Mechanism for Securing Deep Neural Network against Adversarial Attacks"],"prefix":"10.1109","author":[{"given":"Faiq","family":"Khalid","sequence":"first","affiliation":[]},{"given":"Hassan","family":"Ali","sequence":"additional","affiliation":[]},{"given":"Hammad","family":"Tariq","sequence":"additional","affiliation":[]},{"given":"Muhammad Abdullah","family":"Hanif","sequence":"additional","affiliation":[]},{"given":"Semeen","family":"Rehman","sequence":"additional","affiliation":[]},{"given":"Rehan","family":"Ahmed","sequence":"additional","affiliation":[]},{"given":"Muhammad","family":"Shafique","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.23919\/DATE.2018.8342139"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1145\/3128572.3140449"},{"key":"ref33","article-title":"Thermometer encoding: One hot way to resist adversarial examples","author":"buckman","year":"2018","journal-title":"ICLRE"},{"key":"ref32","article-title":"Feature distillation: Dnn-oriented jpeg compression against adversarial examples","author":"liu","year":"2018","journal-title":"arXiv preprint arXiv 1803 05787"},{"key":"ref31","article-title":"Defense-gan: Protecting classifiers against adversarial attacks using generative models","author":"samangouei","year":"2018","journal-title":"CoRR"},{"key":"ref30","article-title":"Simple black-box adversarial perturbations for deep networks","volume":"abs 1612 6299","author":"narodytska","year":"2016","journal-title":"CoRR"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1145\/3133956.3134057"},{"key":"ref36","article-title":"Blind pre-processing: A robust defense method against adversarial examples","author":"rakin","year":"2018","journal-title":"arXiv preprint arxiv 1802 05807"},{"key":"ref35","article-title":"Mitigating adversarial effects through randomization","author":"xie","year":"2017","journal-title":"CoRR"},{"key":"ref34","article-title":"Stochastic activation pruning for robust adversarial defense","author":"dhillon","year":"2018","journal-title":"CoRR"},{"key":"ref10","article-title":"Stealing neural networks via timing side channels","author":"duddu","year":"2018","journal-title":"arXiv preprint arXiv 1812 11720"},{"key":"ref11","first-page":"601","article-title":"Stealing machine learning models via prediction apis","author":"tram\u00e8r","year":"2016","journal-title":"Usenix Security"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2018.2886017"},{"key":"ref13","article-title":"RED-Attack: Resource efficient decision based attack for machine learning","author":"khalid","year":"2019","journal-title":"arXiv preprint arXiv 1901 10055"},{"key":"ref14","article-title":"Explaining and harnessing adversarial examples","author":"goodfellow","year":"2014","journal-title":"arXiv preprint arXiv 1412 6572"},{"key":"ref15","article-title":"Intriguing properties of neural networks","volume":"abs 1312 6199","author":"szegedy","year":"2013","journal-title":"CoRR"},{"key":"ref16","article-title":"AE-GAN: adversarial eliminating with GAN","volume":"abs 1707 5474","author":"shen","year":"2017","journal-title":"CoRR"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/SP.2016.41"},{"key":"ref18","article-title":"Defend deep neural networks against adversarial examples via fixed anddynamic quantized activation functions","author":"rakin","year":"2018","journal-title":"CoRR"},{"key":"ref19","article-title":"Towards evaluating the robustness of neural networks","volume":"abs 1608 4644","author":"carlini","year":"2016","journal-title":"CoRR"},{"key":"ref28","article-title":"Efficient decision-based black-box adversarial attacks on face recognition","author":"dong","year":"2019","journal-title":"arXiv preprint arXiv 1904 01870"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.23919\/DATE.2018.8342120"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.282"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/ISCAS.2018.8351283"},{"journal-title":"DATE","article-title":"FAdeML: understanding the impact of pre-processing noise filtering on adversarial machine learning","year":"2019","key":"ref6"},{"key":"ref29","article-title":"A geometry-inspired decision-based attack","author":"liu","year":"2019","journal-title":"arXiv preprint arXiv 1903 11593"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/IOLTS.2019.8854425"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/IOLTS.2018.8474192"},{"key":"ref7","first-page":"327","article-title":"Security for machine learning-based systems: Attacks and challenges during training and inference","year":"2018","journal-title":"FIT"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2019.2891969"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/ISVLSI.2018.00111"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1177\/0306312717741687"},{"journal-title":"CoRR","article-title":"Defensive distillation is not robust to adversarial examples","year":"2016","key":"ref20"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.14722\/ndss.2018.23198"},{"journal-title":"CoRR","article-title":"Magnet and &#x201D;efficient defenses against adversarial attacks&#x201D; are not robust to adversarial examples","year":"2017","key":"ref21"},{"key":"ref24","article-title":"Foolbox v0.8.0: A python toolbox to benchmark the robustness of machine learning models","author":"rauber","year":"2017","journal-title":"CoRR"},{"key":"ref23","article-title":"cleverhans v0.1: an adversarial machine learning library","author":"goodfellow","year":"2016","journal-title":"CoRR"},{"key":"ref26","article-title":"Adversarial examples in the physical world","volume":"abs 1607 2533","author":"kurakin","year":"2016","journal-title":"CoRR"},{"key":"ref25","article-title":"The limitations of deep learning in adversarial settings","volume":"abs 1511 7528","author":"papernot","year":"2015","journal-title":"CoRR"}],"event":{"name":"2019 IEEE 25th International Symposium on On-Line Testing And Robust System Design (IOLTS)","start":{"date-parts":[[2019,7,1]]},"location":"Rhodes, Greece","end":{"date-parts":[[2019,7,3]]}},"container-title":["2019 IEEE 25th International Symposium on On-Line Testing and Robust System Design (IOLTS)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/8846168\/8854369\/08854377.pdf?arnumber=8854377","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,7,17]],"date-time":"2022-07-17T17:51:34Z","timestamp":1658080294000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8854377\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,7]]},"references-count":39,"URL":"https:\/\/doi.org\/10.1109\/iolts.2019.8854377","relation":{},"subject":[],"published":{"date-parts":[[2019,7]]}}}