{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,19]],"date-time":"2026-02-19T14:48:44Z","timestamp":1771512524613,"version":"3.50.1"},"reference-count":37,"publisher":"IEEE","license":[{"start":{"date-parts":[[2023,10,28]],"date-time":"2023-10-28T00:00:00Z","timestamp":1698451200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,10,28]],"date-time":"2023-10-28T00:00:00Z","timestamp":1698451200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["CCF-1763747"],"award-info":[{"award-number":["CCF-1763747"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023,10,28]]},"DOI":"10.1109\/iccad57390.2023.10323702","type":"proceedings-article","created":{"date-parts":[[2023,11,30]],"date-time":"2023-11-30T18:58:45Z","timestamp":1701370725000},"page":"1-9","source":"Crossref","is-referenced-by-count":8,"title":["RNA-ViT: Reduced-Dimension Approximate Normalized Attention Vision Transformers for Latency Efficient Private Inference"],"prefix":"10.1109","author":[{"given":"Dake","family":"Chen","sequence":"first","affiliation":[{"name":"University of Southern California,Los Angeles,CA,USA"}]},{"given":"Yuke","family":"Zhang","sequence":"additional","affiliation":[{"name":"University of Southern California,Los Angeles,CA,USA"}]},{"given":"Souvik","family":"Kundu","sequence":"additional","affiliation":[{"name":"Intel Labs,San Diego,USA"}]},{"given":"Chenghao","family":"Li","sequence":"additional","affiliation":[{"name":"University of Southern California,Los Angeles,CA,USA"}]},{"given":"Peter A.","family":"Beerel","sequence":"additional","affiliation":[{"name":"University of Southern California,Los Angeles,CA,USA"}]}],"member":"263","reference":[{"key":"ref1","article-title":"MPCViT: Searching for MPC-friendly vision transformer with heterogeneous attention","author":"Zeng","year":"2022","journal-title":"arXiv preprint"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2010.11929"},{"key":"ref3","volume-title":"Escaping the big data paradigm with compact transformers","author":"Hassani","year":"2021"},{"key":"ref4","first-page":"344","article-title":"DNR: a tunable robust pruning framework through dynamic network rewiring of dnns","volume-title":"Proceedings of the 26th Asia and South Pacific Design Automation Conference","author":"Kundu"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00681"},{"key":"ref8","article-title":"SegFormer: Simple and efficient design for semantic segmentation with transformers","author":"Xie","year":"2021","journal-title":"arXiv preprint"},{"key":"ref9","first-page":"9181","article-title":"Analyzing the confidentiality of undistillable teachers in knowledge distillation","volume":"34","author":"Kundu","year":"2021","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref10","article-title":"Undistillable: Making a nasty teacher that cannot teach students","author":"Ma","year":"2021","journal-title":"arXiv preprint"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/SP.2017.12"},{"key":"ref12","first-page":"1651","article-title":"GAZELLE: A low latency framework for secure neural network inference","volume-title":"27th USENIX Security Symposium (USENIX Security 18)","author":"Juvekar"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1145\/3411501.3419418"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/SP40001.2021.00098"},{"key":"ref15","article-title":"Cheetah: Lean and fast secure two-party deep neural network inference","author":"Huang","year":"2022","journal-title":"Cryptology ePrint Archive, Paper 2022\/207"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1145\/3489517.3530680"},{"key":"ref17","article-title":"Learning to linearize deep neural networks for secure and efficient private inference","volume-title":"International Conference on Learning Representation","author":"Kundu"},{"key":"ref18","first-page":"3947","article-title":"Selective network linearization for efficient private inference","volume-title":"International Conference on Machine Learning","author":"Cho"},{"key":"ref19","author":"Li","year":"2022","journal-title":"MPCFormer: fast, performant and private Transformer inference with MPC"},{"key":"ref20","article-title":"Faster CryptoNets: Leveraging sparsity for real-world encrypted inference","author":"Chou","year":"2018","journal-title":"arXiv preprint"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00813"},{"key":"ref22","first-page":"1501","article-title":"XONN: XNOR-based oblivious deep neural network inference","volume-title":"28th USENIX Security Symposium (USENIX Security 19)","author":"Riazi"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/DAC56929.2023.10247682"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.findings-acl.277"},{"key":"ref25","article-title":"Iron: Private inference on transformers","author":"Hao","year":"2022","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref26","first-page":"4839","article-title":"DeepReDuce: ReLU reduction for fast private inference","volume-title":"International Conference on Machine Learning","author":"Jha"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW59228.2023.00494"},{"key":"ref28","article-title":"Linformer: Self-attention with linear complexity","author":"Wang","year":"2020","journal-title":"arXiv preprint"},{"key":"ref29","article-title":"cosFormer: Rethinking softmax in attention","volume-title":"International Conference on Learning Representations","author":"Qin"},{"key":"ref30","article-title":"Is attention better than matrix decomposition?","volume-title":"International Conference on Learning Representations","author":"Geng"},{"key":"ref31","first-page":"21297","article-title":"Soft: Softmax-free transformer with linear complexity","volume":"34","author":"Lu","year":"2021","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2022.3211006"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1016\/0045-7825(79)90002-1"},{"key":"ref34","first-page":"4961","article-title":"CRYPTEN: Secure multi-party computation meets machine learning","volume":"34","author":"Knott","year":"2021","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref35","first-page":"5156","article-title":"Transformers are RNNs: Fast autoregressive transformers with linear attention","volume-title":"International Conference on Machine Learning","author":"Katharopoulos"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.4140\/TCP.n.2015.249"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.74"}],"event":{"name":"2023 IEEE\/ACM International Conference on Computer Aided Design (ICCAD)","location":"San Francisco, CA, USA","start":{"date-parts":[[2023,10,28]]},"end":{"date-parts":[[2023,11,2]]}},"container-title":["2023 IEEE\/ACM International Conference on Computer Aided Design (ICCAD)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/10323590\/10323543\/10323702.pdf?arnumber=10323702","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,2]],"date-time":"2024-03-02T22:04:49Z","timestamp":1709417089000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10323702\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,10,28]]},"references-count":37,"URL":"https:\/\/doi.org\/10.1109\/iccad57390.2023.10323702","relation":{},"subject":[],"published":{"date-parts":[[2023,10,28]]}}}