{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,31]],"date-time":"2025-10-31T08:00:07Z","timestamp":1761897607114},"reference-count":30,"publisher":"IEEE","license":[{"start":{"date-parts":[[2021,7,18]],"date-time":"2021-07-18T00:00:00Z","timestamp":1626566400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2021,7,18]],"date-time":"2021-07-18T00:00:00Z","timestamp":1626566400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2021,7,18]],"date-time":"2021-07-18T00:00:00Z","timestamp":1626566400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021,7,18]]},"DOI":"10.1109\/ijcnn52387.2021.9534432","type":"proceedings-article","created":{"date-parts":[[2021,9,21]],"date-time":"2021-09-21T20:40:52Z","timestamp":1632256852000},"source":"Crossref","is-referenced-by-count":9,"title":["Sequence Squeezing: A Defense Method Against Adversarial Examples for API Call-Based RNN Variants"],"prefix":"10.1109","author":[{"given":"Ishai","family":"Rosenberg","sequence":"first","affiliation":[]},{"given":"Asaf","family":"Shabtai","sequence":"additional","affiliation":[]},{"given":"Yuval","family":"Elovici","sequence":"additional","affiliation":[]},{"given":"Lior","family":"Rokach","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.14722\/ndss.2018.23198"},{"key":"ref10","first-page":"163","article-title":"Code vectors: understanding programs through embedded abstracted symbolic traces","author":"jordan","year":"2018","journal-title":"FSE"},{"key":"ref11","article-title":"Black-box attacks against RNN based malware detection algorithms","author":"hu","year":"2017","journal-title":"ArXiv"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/SP.2019.00044"},{"key":"ref13","article-title":"Textbugger: Generating adversarial text against real-world applications","author":"li","year":"2018","journal-title":"CoRR abs\/1812 05271"},{"key":"ref14","article-title":"Towards deep learning models resistant to adversarial attacks","author":"madry","year":"2018","journal-title":"ICLRE"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/ISI.2007.379542"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/MILCOM.2016.7795300"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/D14-1162"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN48605.2020.9207168"},{"key":"ref19","author":"rosenberg","year":"2020","journal-title":"Adver-sariallearning in the cyber security domain"},{"key":"ref28","author":"tramer","year":"2020","journal-title":"On adaptive attacks to adversarial example defenses"},{"key":"ref4","article-title":"On evaluating adversarial robustness","author":"nicholas","year":"2019","journal-title":"CoRR abs\/1902 06705"},{"key":"ref27","article-title":"Intriguing properties of neural networks","author":"christian","year":"2014","journal-title":"abs\/1312 6199"},{"key":"ref3","first-page":"274","article-title":"Obfuscated gradients give a false sense of security: Circumventing defenses to adversarial examples","author":"athalye","year":"2018","journal-title":"ICML"},{"key":"ref6","first-page":"1310","article-title":"Certified adversarial robustness via randomized smoothing","author":"cohen","year":"2019","journal-title":"ICML"},{"key":"ref29","first-page":"5283","article-title":"Provable defenses against adversarial examples via the convex outer adversarial polytope","author":"wong","year":"2018","journal-title":"ICML"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1145\/3128572.3140444"},{"key":"ref8","article-title":"On the (statistical) detection of adversarial examples","author":"grosse","year":"2017","journal-title":"ArXiv e-prints abs\/1702 06280"},{"key":"ref7","article-title":"Detecting Adversarial Samples from Artifacts","author":"feinman","year":"2017","journal-title":"ArXiv e-prints"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D18-1316"},{"key":"ref9","article-title":"Countering adversarial images using input transformations","author":"guo","year":"2018","journal-title":"ICLRE"},{"key":"ref1","article-title":"Robust neural malware detection models for emulation sequence learning","author":"rakshit","year":"2018","journal-title":"CoRR abs\/1806 10741"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1145\/3427228.3427230"},{"key":"ref22","article-title":"Defense-gan: Protecting classifiers against adversarial attacks using generative models","author":"samangouei","year":"2018","journal-title":"CoRR abs\/1805 06605"},{"key":"ref21","first-page":"490","article-title":"Generic black-box end-to-end attack against state of the art API call based malware classifiers","author":"rosenberg","year":"2018","journal-title":"RAID"},{"key":"ref24","article-title":"Attack and defense of dynamic analysis-based, adversarial neural malware classification models","author":"stokes","year":"2017","journal-title":"CoRR abs\/1712 05919"},{"key":"ref23","article-title":"Bypassing feature squeezing by increasing adversary strength","author":"sharma","year":"2018","journal-title":"ArXiv Preprint"},{"key":"ref26","first-page":"1299","article-title":"When does machine learning FAIL? generalized transferability for evasion and poisoning attacks","author":"suciu","year":"2018","journal-title":"(USENIX Security)"},{"key":"ref25","article-title":"Ensemble Methods as a Defense to Adversarial Perturbations Against Deep Neural Networks","author":"strauss","year":"2017","journal-title":"ArXiv e-prints"}],"event":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","location":"Shenzhen, China","start":{"date-parts":[[2021,7,18]]},"end":{"date-parts":[[2021,7,22]]}},"container-title":["2021 International Joint Conference on Neural Networks (IJCNN)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9533266\/9533267\/09534432.pdf?arnumber=9534432","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,5,10]],"date-time":"2022-05-10T15:46:06Z","timestamp":1652197566000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9534432\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,7,18]]},"references-count":30,"URL":"https:\/\/doi.org\/10.1109\/ijcnn52387.2021.9534432","relation":{},"subject":[],"published":{"date-parts":[[2021,7,18]]}}}