{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T06:09:27Z","timestamp":1769062167706,"version":"3.49.0"},"reference-count":35,"publisher":"IEEE","license":[{"start":{"date-parts":[[2020,12,10]],"date-time":"2020-12-10T00:00:00Z","timestamp":1607558400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2020,12,10]],"date-time":"2020-12-10T00:00:00Z","timestamp":1607558400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2020,12,10]],"date-time":"2020-12-10T00:00:00Z","timestamp":1607558400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000002","name":"National Institutes of Health","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100000002","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000181","name":"Air Force Office of Scientific Research","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100000181","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020,12,10]]},"DOI":"10.1109\/bigdata50022.2020.9378274","type":"proceedings-article","created":{"date-parts":[[2021,3,19]],"date-time":"2021-03-19T21:10:21Z","timestamp":1616188221000},"page":"1061-1070","source":"Crossref","is-referenced-by-count":21,"title":["Broadening Differential Privacy for Deep Learning Against Model Inversion Attacks"],"prefix":"10.1109","author":[{"given":"Qiuchen","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Jing","family":"Ma","sequence":"additional","affiliation":[]},{"given":"Yonghui","family":"Xiao","sequence":"additional","affiliation":[]},{"given":"Jian","family":"Lou","sequence":"additional","affiliation":[]},{"given":"Li","family":"Xiong","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref33","article-title":"Improving robustness to model inversion attacks via mutual information regularization","author":"wang","year":"2020"},{"key":"ref32","first-page":"14774","article-title":"Deep leakage from gradients","author":"zhu","year":"2019","journal-title":"NIPS"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/5.726791"},{"key":"ref30","article-title":"A motion and similarity-based fake detection method for biometric face recognition systems","author":"kim","year":"2011","journal-title":"IEEE Transactions on Consumer Electronics"},{"key":"ref35","article-title":"Differentially private federated learning: A client level perspective","author":"geyer","year":"2017","journal-title":"NIPS Workshop"},{"key":"ref34","article-title":"Learning differentially private recurrent language models","volume":"30","author":"brendan","year":"2018","journal-title":"International Conference on Learning Representations"},{"key":"ref10","article-title":"Differentially private empirical risk minimization","author":"chaudhuri","year":"2011","journal-title":"JMLR"},{"key":"ref11","article-title":"Uplink communication efficient differentially private sparse optimization with feature-wise distributed data","author":"lou","year":"2018","journal-title":"Thirty-Second AAAI Conference on Artificial Intelligence"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2020.3020955"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1145\/2976749.2978318"},{"key":"ref14","doi-asserted-by":"crossref","DOI":"10.1609\/aaai.v30i1.10165","article-title":"Differential privacy preservation for deep auto-encoders: an application of human behavior prediction","author":"phan","year":"2016","journal-title":"AAAI"},{"key":"ref15","article-title":"Semi-supervised knowledge transfer for deep learning from private training data","author":"papernot","year":"2017","journal-title":"ICLRE"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1145\/3219819.3220106"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/SP.2019.00019"},{"key":"ref18","first-page":"1895","article-title":"Evaluating differentially private machine learning in practice","author":"jayaraman","year":"2019","journal-title":"28th USENIX Security Symposium ( USENIX Security 19)"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1145\/3219819.3220076"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/SP.2017.49"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/SP.2017.41"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1186\/s13640-015-0075-4"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1145\/2810103.2813677"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1145\/3133956.3134012"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2003.819861"},{"key":"ref5","article-title":"Adversarial neural network inversion via auxiliary knowledge alignment","author":"yang","year":"2019"},{"key":"ref8","article-title":"Updates-leak: Data set inference and reconstruction attacks in online learning","author":"salem","year":"2020","journal-title":"USENIX Security Symposium"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM.2019.8737416"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2020.2979670"},{"key":"ref9","article-title":"The algorithmic foundations of differential privacy","author":"dwork","year":"2014","journal-title":"Foundations and Trends\ufffd in Theoretical Computer Science"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-019-01247-4"},{"key":"ref20","first-page":"61","article-title":"Membership inference attack against differentially private deep learning model","volume":"11","author":"rahman","year":"2018","journal-title":"Trans Data Privacy"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1145\/2508859.2516735"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1145\/3133956.3134077"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/FOCS.2010.12"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-39077-7_5"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2004.834669"},{"key":"ref25","article-title":"Robust or private? adversarial training makes models more vulnerable to privacy attacks","author":"mejia","year":"2019"}],"event":{"name":"2020 IEEE International Conference on Big Data (Big Data)","location":"Atlanta, GA, USA","start":{"date-parts":[[2020,12,10]]},"end":{"date-parts":[[2020,12,13]]}},"container-title":["2020 IEEE International Conference on Big Data (Big Data)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9377717\/9377728\/09378274.pdf?arnumber=9378274","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,12,22]],"date-time":"2022-12-22T00:55:45Z","timestamp":1671670545000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9378274\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,12,10]]},"references-count":35,"URL":"https:\/\/doi.org\/10.1109\/bigdata50022.2020.9378274","relation":{},"subject":[],"published":{"date-parts":[[2020,12,10]]}}}