{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,28]],"date-time":"2026-03-28T20:59:06Z","timestamp":1774731546500,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":65,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,11,13]],"date-time":"2022-11-13T00:00:00Z","timestamp":1668297600000},"content-version":"vor","delay-in-days":366,"URL":"http:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/100000001","name":"NSF (National Science Foundation)","doi-asserted-by":"publisher","award":["1910100"],"award-info":[{"award-number":["1910100"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"name":"NSF CNS","award":["20-46726 CAR"],"award-info":[{"award-number":["20-46726 CAR"]}]},{"name":"Amazon Research Award"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2021,11,12]]},"DOI":"10.1145\/3460120.3484579","type":"proceedings-article","created":{"date-parts":[[2021,11,13]],"date-time":"2021-11-13T12:05:27Z","timestamp":1636805127000},"page":"2146-2168","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":41,"title":["DataLens: Scalable Privacy Preserving Training via Gradient Compression and Aggregation"],"prefix":"10.1145","author":[{"given":"Boxin","family":"Wang","sequence":"first","affiliation":[{"name":"University of Illinois at Urbana-Champaign, Urbana, IL, USA"}]},{"given":"Fan","family":"Wu","sequence":"additional","affiliation":[{"name":"University of Illinois at Urbana-Champaign, Urbana, IL, USA"}]},{"given":"Yunhui","family":"Long","sequence":"additional","affiliation":[{"name":"University of Illinois at Urbana-Champaign, Urbana, IL, USA"}]},{"given":"Luka","family":"Rimanic","sequence":"additional","affiliation":[{"name":"ETH Z\u00fcrich, Zurich, Switzerland"}]},{"given":"Ce","family":"Zhang","sequence":"additional","affiliation":[{"name":"ETH Z\u00fcrich, Zurich, Switzerland"}]},{"given":"Bo","family":"Li","sequence":"additional","affiliation":[{"name":"University of Illinois at Urbana-Champaign, Urbana, IL, USA"}]}],"member":"320","published-online":{"date-parts":[[2021,11,13]]},"reference":[{"key":"e_1_3_2_2_1_1","unstructured":"2020. Opacus - Train PyTorch models with Differential Privacy. https:\/\/opacus.ai\/"},{"key":"e_1_3_2_2_2_1","doi-asserted-by":"publisher","DOI":"10.1145\/2976749.2978318"},{"key":"e_1_3_2_2_3_1","volume-title":"Felix Xinnan X Yu, Sanjiv Kumar, and Brendan McMahan.","author":"Agarwal Naman","year":"2018","unstructured":"Naman Agarwal, Ananda Theertha Suresh, Felix Xinnan X Yu, Sanjiv Kumar, and Brendan McMahan. 2018. cpSGD: Communication-efficient and differentially-private distributed SGD. In Advances in Neural Information Processing Systems."},{"key":"e_1_3_2_2_4_1","volume-title":"QSGD: Communication-Efficient SGD via Gradient Quantization and Encoding. In Advances in Neural Information Processing Systems.","author":"Alistarh Dan","year":"2017","unstructured":"Dan Alistarh, Demjan Grubic, Jerry Li, Ryota Tomioka, and Milan Vojnovic. 2017. QSGD: Communication-Efficient SGD via Gradient Quantization and Encoding. In Advances in Neural Information Processing Systems."},{"key":"e_1_3_2_2_5_1","unstructured":"Dan Alistarh Torsten Hoefler Mikael Johansson Nikola Konstantinov Sarit Khirirat and Cedric Renggli. 2018. The Convergence of Sparsified Gradient Methods. In Advances in Neural Information Processing Systems."},{"key":"e_1_3_2_2_6_1","doi-asserted-by":"crossref","unstructured":"Tal Ben-Nun and Torsten Hoefler. 2019. Demystifying Parallel and Distributed Deep Learning: An In-Depth Concurrency Analysis. ACM Comput. Surv. (2019).","DOI":"10.1145\/3320060"},{"key":"e_1_3_2_2_7_1","volume-title":"Proceedings of the 35th International Conference on Machine Learning.","author":"Bernstein Jeremy","year":"2018","unstructured":"Jeremy Bernstein, Yu-Xiang Wang, Kamyar Azizzadenesheli, and Animashree Anandkumar. 2018a. signSGD: Compressed Optimisation for Non-Convex Problems. In Proceedings of the 35th International Conference on Machine Learning."},{"key":"e_1_3_2_2_8_1","volume-title":"Proceedings of the 35th International Conference on Machine Learning","volume":"80","author":"Bernstein Jeremy","year":"2018","unstructured":"Jeremy Bernstein, Yu-Xiang Wang, Kamyar Azizzadenesheli, and Animashree Anandkumar. 2018b. signSGD : Compressed Optimisation for Non-Convex Problems. Proceedings of the 35th International Conference on Machine Learning, Vol. 80 (2018), 560--569."},{"key":"e_1_3_2_2_9_1","volume-title":"Proceedings of the 35th International Conference on Machine Learning (Proceedings of Machine Learning Research","volume":"569","author":"Bernstein Jeremy","year":"2018","unstructured":"Jeremy Bernstein, Yu-Xiang Wang, Kamyar Azizzadenesheli, and Animashree Anandkumar. 2018c. signSGD: Compressed Optimisation for Non-Convex Problems. In Proceedings of the 35th International Conference on Machine Learning (Proceedings of Machine Learning Research, Vol. 80). PMLR, 560--569."},{"key":"e_1_3_2_2_10_1","volume-title":"An Attack on InstaHide: Is Private Learning Possible with Instance Encoding? arXiv preprint arXiv:2011.05315","author":"Carlini Nicholas","year":"2020","unstructured":"Nicholas Carlini, Samuel Deng, Sanjam Garg, Somesh Jha, Saeed Mahloujifar, Mohammad Mahmoody, Shuang Song, Abhradeep Thakurta, and Florian Tramer. 2020. An Attack on InstaHide: Is Private Learning Possible with Instance Encoding? arXiv preprint arXiv:2011.05315 (2020)."},{"key":"e_1_3_2_2_11_1","volume-title":"28th $$USENIX$$ Security Symposium ($$USENIX$$ Security 19). 267--284.","author":"Carlini Nicholas","unstructured":"Nicholas Carlini, Chang Liu, \u00dalfar Erlingsson, Jernej Kos, and Dawn Song. 2019. The secret sharer: Evaluating and testing unintended memorization in neural networks. In 28th $$USENIX$$ Security Symposium ($$USENIX$$ Security 19). 267--284."},{"key":"e_1_3_2_2_12_1","first-page":"13551","article-title":"a. ScaleCom: Scalable Sparsified Gradient Compression for Communication-Efficient Distributed Training","volume":"33","author":"Chen Chia-Yu","year":"2020","unstructured":"Chia-Yu Chen, Jiamin Ni, Songtao Lu, Xiaodong Cui, Pin-Yu Chen, Xiao Sun, Naigang Wang, Swagath Venkataramani, Vijayalakshmi (viji) Srinivasan, Wei Zhang, and Kailash Gopalakrishnan. 2020 a. ScaleCom: Scalable Sparsified Gradient Compression for Communication-Efficient Distributed Training. Adv. Neural Inf. Process. Syst., Vol. 33 (2020), 13551--13563.","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"e_1_3_2_2_13_1","volume-title":"2020 b. GS-WGAN: A Gradient-Sanitized Approach for Learning Differentially Private Generators. Neural Information Processing Systems (NeurIPS)","author":"Chen Dingfan","year":"2020","unstructured":"Dingfan Chen, Tribhuvanesh Orekondy, and Mario Fritz. 2020 b. GS-WGAN: A Gradient-Sanitized Approach for Learning Differentially Private Generators. Neural Information Processing Systems (NeurIPS) (2020)."},{"key":"e_1_3_2_2_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/3372297.3417238"},{"key":"e_1_3_2_2_15_1","volume-title":"Lin (Eds.)","volume":"33","author":"Chen Xiangyi","year":"2020","unstructured":"Xiangyi Chen, Steven Z. Wu, and Mingyi Hong. 2020 c. Understanding Gradient Clipping in Private SGD: A Geometric Perspective. In Advances in Neural Information Processing Systems, H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin (Eds.), Vol. 33. Curran Associates, Inc., 13773--13782."},{"key":"e_1_3_2_2_16_1","doi-asserted-by":"publisher","DOI":"10.5555\/1791834.1791836"},{"key":"e_1_3_2_2_17_1","volume-title":"Foundations and Trends\u00ae in Theoretical Computer Science","volume":"9","author":"Dwork Cynthia","year":"2014","unstructured":"Cynthia Dwork, Aaron Roth, et al. 2014. The algorithmic foundations of differential privacy. Foundations and Trends\u00ae in Theoretical Computer Science, Vol. 9, 3--4 (2014), 211--407."},{"key":"e_1_3_2_2_18_1","volume-title":"International Conference on Machine Learning.","author":"Fu Fangcheng","year":"2020","unstructured":"Fangcheng Fu, Yuzheng Hu, Yihan He, Jiawei Jiang, Yingxia Shao, Ce Zhang, and Bin Cui. 2020. Don't Waste Your Bits! Squeeze Activations and Gradients for Deep Neural Networks via TinyScript. In International Conference on Machine Learning."},{"key":"e_1_3_2_2_19_1","unstructured":"Ian Goodfellow Jean Pouget-Abadie Mehdi Mirza Bing Xu David Warde-Farley Sherjil Ozair Aaron Courville and Yoshua Bengio. 2014. Generative adversarial nets. In Advances in neural information processing systems. 2672--2680."},{"key":"e_1_3_2_2_20_1","unstructured":"Moritz Hardt Katrina Ligett and Frank McSherry. 2012. A simple and practical algorithm for differentially private data release. In Advances in Neural Information Processing Systems. 2339--2347."},{"key":"e_1_3_2_2_21_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_3_2_2_22_1","doi-asserted-by":"publisher","DOI":"10.14778\/3342263.3342637"},{"key":"e_1_3_2_2_23_1","unstructured":"Ruoxi Jia David Dao Boxin Wang Frances Ann Hubis Nick Hynes Nezihe Merve G\u00fcrel Bo Li Ce Zhang Dawn Xiaodong Song and Costas J. Spanos. 2019 b. Towards Efficient Data Valuation Based on the Shapley Value. In AISTATS."},{"key":"e_1_3_2_2_24_1","volume-title":"Stochastic-Sign SGD for Federated Learning with Theoretical Guarantees. arXiv preprint arXiv:2002.10940","author":"Jin Richeng","year":"2020","unstructured":"Richeng Jin, Yufan Huang, Xiaofan He, Huaiyu Dai, and Tianfu Wu. 2020. Stochastic-Sign SGD for Federated Learning with Theoretical Guarantees. arXiv preprint arXiv:2002.10940 (2020)."},{"key":"e_1_3_2_2_25_1","volume-title":"The Distributed Discrete Gaussian Mechanism for Federated Learning with Secure Aggregation. arXiv preprint arXiv:2102.06387","author":"Kairouz Peter","year":"2021","unstructured":"Peter Kairouz, Ziyu Liu, and Thomas Steinke. 2021. The Distributed Discrete Gaussian Mechanism for Federated Learning with Secure Aggregation. arXiv preprint arXiv:2102.06387 (2021)."},{"key":"e_1_3_2_2_26_1","doi-asserted-by":"publisher","DOI":"10.1214\/aoms\/1177729392"},{"key":"e_1_3_2_2_27_1","volume-title":"Proceedings of the 36th International Conference on Machine Learning (Proceedings of Machine Learning Research","author":"Koloskova Anastasia","unstructured":"Anastasia Koloskova, Sebastian Stich, and Martin Jaggi. 2019 a. Decentralized Stochastic Optimization and Gossip Algorithms with Compressed Communication. In Proceedings of the 36th International Conference on Machine Learning (Proceedings of Machine Learning Research, Vol. 97), Kamalika Chaudhuri and Ruslan Salakhutdinov (Eds.). PMLR, Long Beach, California, USA, 3478--3487."},{"key":"e_1_3_2_2_28_1","volume-title":"Proceedings of the 36th International Conference on Machine Learning.","author":"Koloskova Anastasia","year":"2019","unstructured":"Anastasia Koloskova, Sebastian Stich, and Martin Jaggi. 2019 b. Decentralized Stochastic Optimization and Gossip Algorithms with Compressed Communication. In Proceedings of the 36th International Conference on Machine Learning."},{"key":"e_1_3_2_2_29_1","unstructured":"Alex Krizhevsky Vinod Nair and Geoffrey Hinton. [n. d.]. CIFAR-10 (Canadian Institute for Advanced Research). ( [n. d.]). http:\/\/www.cs.toronto.edu\/ kriz\/cifar.html"},{"key":"e_1_3_2_2_30_1","unstructured":"Yann LeCun. 1998. The MNIST database of handwritten digits. http:\/\/yann.lecun.com\/exdb\/mnist\/ (1998)."},{"key":"e_1_3_2_2_31_1","first-page":"5495","article-title":"ALICE: Towards Understanding Adversarial Learning for Joint Distribution Matching","volume":"30","author":"Li Chunyuan","year":"2017","unstructured":"Chunyuan Li, Hao Liu, Changyou Chen, Yuchen Pu, Liqun Chen, Ricardo Henao, and Lawrence Carin. 2017. ALICE: Towards Understanding Adversarial Learning for Joint Distribution Matching. Advances in Neural Information Processing Systems, Vol. 30 (2017), 5495--5503.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_2_32_1","unstructured":"Xiangru Lian Yijun Huang Yuncheng Li and Ji Liu. 2015. Asynchronous Parallel Stochastic Gradient for Nonconvex Optimization. In Advances in Neural Information Processing Systems."},{"key":"e_1_3_2_2_33_1","unstructured":"Xiangru Lian Ce Zhang Huan Zhang Cho-Jui Hsieh Wei Zhang and Ji Liu. 2017. Can Decentralized Algorithms Outperform Centralized Algorithms? A Case Study for Decentralized Parallel Stochastic Gradient Descent. In Advances in Neural Information Processing Systems."},{"key":"e_1_3_2_2_34_1","volume-title":"Proceedings of the 2nd SysML Conference.","author":"Lim Hyeontaek","year":"2019","unstructured":"Hyeontaek Lim, David G Andersen, and Michael Kaminsky. 2019. 3LC: LIGHTWEIGHT AND EFFECTIVE TRAFFIC COMPRESSION FOR DISTRIBUTED MACHINE LEARNING. In Proceedings of the 2nd SysML Conference."},{"key":"e_1_3_2_2_35_1","doi-asserted-by":"publisher","DOI":"10.1561\/9781680837018"},{"key":"e_1_3_2_2_36_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.425"},{"key":"e_1_3_2_2_37_1","volume-title":"Scalable differentially private generative student model via pate. arXiv preprint arXiv:1906.09338","author":"Long Yunhui","year":"2019","unstructured":"Yunhui Long, Suxin Lin, Zhuolin Yang, Carl A Gunter, and Bo Li. 2019. Scalable differentially private generative student model via pate. arXiv preprint arXiv:1906.09338 (2019)."},{"key":"e_1_3_2_2_38_1","unstructured":"Brendan McMahan Eider Moore Daniel Ramage Seth Hampson and Blaise Aguera y Arcas. 2017. Communication-efficient learning of deep networks from decentralized data. In Artificial Intelligence and Statistics. PMLR 1273--1282."},{"key":"e_1_3_2_2_39_1","volume-title":"A General Approach to Adding Differential Privacy to Iterative Training Procedures. arxiv","author":"McMahan H. Brendan","year":"1812","unstructured":"H. Brendan McMahan, Galen Andrew, Ulfar Erlingsson, Steve Chien, Ilya Mironov, Nicolas Papernot, and Peter Kairouz. 2019. A General Approach to Adding Differential Privacy to Iterative Training Procedures. arxiv: 1812.06210 [cs.LG]"},{"key":"e_1_3_2_2_40_1","volume-title":"Proceedings of the 35th International Conference on Machine Learning.","author":"Mescheder Lars","year":"2018","unstructured":"Lars Mescheder, Andreas Geiger, and Sebastian Nowozin. 2018. Which Training Methods for GANs do actually Converge?. In Proceedings of the 35th International Conference on Machine Learning."},{"key":"e_1_3_2_2_41_1","doi-asserted-by":"publisher","DOI":"10.1109\/CSF.2017.11"},{"key":"e_1_3_2_2_42_1","volume-title":"R'enyi Differential Privacy of the Sampled Gaussian Mechanism. arXiv preprint arXiv:1908.10530","author":"Mironov Ilya","year":"2019","unstructured":"Ilya Mironov, Kunal Talwar, and Li Zhang. 2019. R'enyi Differential Privacy of the Sampled Gaussian Mechanism. arXiv preprint arXiv:1908.10530 (2019)."},{"key":"e_1_3_2_2_43_1","volume-title":"International Conference on Learning Representations.","author":"Papernot Nicolas","year":"2017","unstructured":"Nicolas Papernot, Mart'in Abadi, Ulfar Erlingsson, Ian Goodfellow, and Kunal Talwar. 2017. Semi-supervised knowledge transfer for deep learning from private training data. In International Conference on Learning Representations."},{"key":"e_1_3_2_2_44_1","volume-title":"Scalable Private Learning with PATE. In International Conference on Learning Representations.","author":"Papernot Nicolas","year":"2018","unstructured":"Nicolas Papernot, Shuang Song, Ilya Mironov, Ananth Raghunathan, Kunal Talwar, and Ulfar Erlingsson. 2018. Scalable Private Learning with PATE. In International Conference on Learning Representations."},{"key":"e_1_3_2_2_45_1","volume-title":"Felix X. Yu, Sashank J. Reddi, and Sanjiv Kumar.","author":"Pichapati Venkatadheeraj","year":"2019","unstructured":"Venkatadheeraj Pichapati, Ananda Theertha Suresh, Felix X. Yu, Sashank J. Reddi, and Sanjiv Kumar. 2019. AdaCliP: Adaptive Clipping for Private SGD. CoRR, Vol. abs\/1908.07643 (2019). arxiv: 1908.07643 http:\/\/arxiv.org\/abs\/1908.07643"},{"key":"e_1_3_2_2_46_1","doi-asserted-by":"publisher","DOI":"10.1145\/2588555.2588575"},{"key":"e_1_3_2_2_47_1","volume-title":"Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434","author":"Radford Alec","year":"2015","unstructured":"Alec Radford, Luke Metz, and Soumith Chintala. 2015. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434 (2015)."},{"key":"e_1_3_2_2_48_1","volume-title":"International Conference on Machine Learning. PMLR, 8253--8265","author":"Rothchild Daniel","year":"2020","unstructured":"Daniel Rothchild, Ashwinee Panda, Enayat Ullah, Nikita Ivkin, Ion Stoica, Vladimir Braverman, Joseph Gonzalez, and Raman Arora. 2020. Fetchsgd: Communication-efficient federated learning with sketching. In International Conference on Machine Learning. PMLR, 8253--8265."},{"key":"e_1_3_2_2_49_1","volume-title":"Proceedings of the 28th International Conference on Neural Information Processing Systems.","author":"Sa Christopher De","year":"2015","unstructured":"Christopher De Sa, Ce Zhang, Kunle Olukotun, and Christopher R\u00e9. 2015. Taming the Wild: A Unified Analysis of HOG WILD! -Style Algorithms. In Proceedings of the 28th International Conference on Neural Information Processing Systems."},{"key":"e_1_3_2_2_50_1","volume-title":"Proceedings of the 30th International Conference on Neural Information Processing Systems. 2234--2242","author":"Salimans Tim","year":"2016","unstructured":"Tim Salimans, Ian Goodfellow, Wojciech Zaremba, Vicki Cheung, Alec Radford, and Xi Chen. 2016. Improved techniques for training GANs. In Proceedings of the 30th International Conference on Neural Information Processing Systems. 2234--2242."},{"key":"e_1_3_2_2_51_1","volume-title":"Membership Inference Attacks Against Machine Learning Models. In 2017 IEEE Symposium on Security and Privacy (SP). 3--18","author":"Shokri R.","unstructured":"R. Shokri, M. Stronati, C. Song, and V. Shmatikov. 2017. Membership Inference Attacks Against Machine Learning Models. In 2017 IEEE Symposium on Security and Privacy (SP). 3--18."},{"key":"e_1_3_2_2_52_1","volume-title":"Advances in Neural Information Processing Systems 31, S Bengio, H Wallach, H Larochelle, K Grauman, N Cesa-Bianchi, and R Garnett (Eds.). Curran Associates","author":"Tang Hanlin","unstructured":"Hanlin Tang, Shaoduo Gan, Ce Zhang, Tong Zhang, and Ji Liu. 2018. Communication Compression for Decentralized Training. In Advances in Neural Information Processing Systems 31, S Bengio, H Wallach, H Larochelle, K Grauman, N Cesa-Bianchi, and R Garnett (Eds.). Curran Associates, Inc., 7652--7662."},{"key":"e_1_3_2_2_53_1","volume-title":"Differentially Private Learning with Adaptive Clipping. CoRR","author":"Thakkar Om","year":"2019","unstructured":"Om Thakkar, Galen Andrew, and H. Brendan McMahan. 2019. Differentially Private Learning with Adaptive Clipping. CoRR, Vol. abs\/1905.03871 (2019). arxiv: 1905.03871 http:\/\/arxiv.org\/abs\/1905.03871"},{"key":"e_1_3_2_2_54_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2019.00018"},{"key":"e_1_3_2_2_55_1","volume-title":"Sai Praneeth Karimireddy, and Martin Jaggi","author":"Vogels Thijs","year":"2020","unstructured":"Thijs Vogels, Sai Praneeth Karimireddy, and Martin Jaggi. 2020. Practical Low-Rank Communication Compression in Decentralized Deep Learning. Adv. Neural Inf. Process. Syst., Vol. 33 (2020)."},{"key":"e_1_3_2_2_56_1","volume-title":"D2P-Fed: Differentially private federated learning with efficient communication. arxiv. org\/pdf\/2006.13039","author":"Wang L","year":"2020","unstructured":"L Wang, R Jia, and D Song. 2020. D2P-Fed: Differentially private federated learning with efficient communication. arxiv. org\/pdf\/2006.13039 (2020)."},{"key":"e_1_3_2_2_57_1","doi-asserted-by":"publisher","DOI":"10.5555\/3326943.3327063"},{"key":"e_1_3_2_2_58_1","volume-title":"Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. arXiv preprint arXiv:1708.07747","author":"Xiao Han","year":"2017","unstructured":"Han Xiao, Kashif Rasul, and Roland Vollgraf. 2017. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. arXiv preprint arXiv:1708.07747 (2017)."},{"key":"e_1_3_2_2_59_1","volume-title":"Differentially Private Generative Adversarial Network. arXiv preprint arXiv:1802.06739","author":"Xie Liyang","year":"2018","unstructured":"Liyang Xie, Kaixiang Lin, Shu Wang, Fei Wang, and Jiayu Zhou. 2018. Differentially Private Generative Adversarial Network. arXiv preprint arXiv:1802.06739 (2018)."},{"key":"e_1_3_2_2_60_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i01.5462"},{"key":"e_1_3_2_2_61_1","doi-asserted-by":"publisher","DOI":"10.1109\/CSF.2018.00027"},{"key":"e_1_3_2_2_62_1","volume-title":"PATE-GAN: Generating Synthetic Data with Differential Privacy Guarantees. In International Conference on Learning Representations.","author":"Yoon Jinsung","unstructured":"Jinsung Yoon, James Jordon, and Mihaela van der Schaar. 2019. PATE-GAN: Generating Synthetic Data with Differential Privacy Guarantees. In International Conference on Learning Representations."},{"key":"e_1_3_2_2_63_1","doi-asserted-by":"publisher","DOI":"10.1145\/3134428"},{"key":"e_1_3_2_2_64_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00033"},{"key":"e_1_3_2_2_65_1","volume-title":"Places: A 10 million Image Database for Scene Recognition","author":"Zhou Bolei","year":"2017","unstructured":"Bolei Zhou, Agata Lapedriza, Aditya Khosla, Aude Oliva, and Antonio Torralba. 2017. Places: A 10 million Image Database for Scene Recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence (2017)."}],"event":{"name":"CCS '21: 2021 ACM SIGSAC Conference on Computer and Communications Security","location":"Virtual Event Republic of Korea","acronym":"CCS '21","sponsor":["SIGSAC ACM Special Interest Group on Security, Audit, and Control"]},"container-title":["Proceedings of the 2021 ACM SIGSAC Conference on Computer and Communications Security"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3460120.3484579","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3460120.3484579","content-type":"application\/pdf","content-version":"vor","intended-application":"syndication"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3460120.3484579","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,18]],"date-time":"2025-11-18T20:45:23Z","timestamp":1763498723000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3460120.3484579"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,11,12]]},"references-count":65,"alternative-id":["10.1145\/3460120.3484579","10.1145\/3460120"],"URL":"https:\/\/doi.org\/10.1145\/3460120.3484579","relation":{},"subject":[],"published":{"date-parts":[[2021,11,12]]},"assertion":[{"value":"2021-11-13","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}