{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,10]],"date-time":"2026-01-10T19:40:33Z","timestamp":1768074033865,"version":"3.49.0"},"publisher-location":"Cham","reference-count":52,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031728549","type":"print"},{"value":"9783031728556","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,11,9]],"date-time":"2024-11-09T00:00:00Z","timestamp":1731110400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,9]],"date-time":"2024-11-09T00:00:00Z","timestamp":1731110400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-72855-6_18","type":"book-chapter","created":{"date-parts":[[2024,11,8]],"date-time":"2024-11-08T18:48:44Z","timestamp":1731091724000},"page":"307-325","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["CLAMP-ViT: Contrastive Data-Free Learning for\u00a0Adaptive Post-training Quantization of\u00a0ViTs"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0000-4763-3321","authenticated-orcid":false,"given":"Akshat","family":"Ramachandran","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3533-9405","authenticated-orcid":false,"given":"Souvik","family":"Kundu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5738-6942","authenticated-orcid":false,"given":"Tushar","family":"Krishna","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,9]]},"reference":[{"issue":"1\u20134","key":"18_CR1","first-page":"1","volume":"37","author":"C Baskin","year":"2021","unstructured":"Baskin, C., et al.: Uniq: uniform noise injection for non-uniform quantization of neural networks. ACM Trans. Comput. Syst. (TOCS) 37(1\u20134), 1\u201315 (2021)","journal-title":"ACM Trans. Comput. Syst. (TOCS)"},{"key":"18_CR2","doi-asserted-by":"crossref","unstructured":"Cai, Y., Yao, Z., Dong, Z., Gholami, A., Mahoney, M.W., Keutzer, K.: Zeroq: a novel zero shot quantization framework. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13169\u201313178 (2020)","DOI":"10.1109\/CVPR42600.2020.01318"},{"key":"18_CR3","doi-asserted-by":"crossref","unstructured":"Cai, Z., Vasconcelos, N.: Cascade r-cnn: delving into high quality object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6154\u20136162 (2018)","DOI":"10.1109\/CVPR.2018.00644"},{"key":"18_CR4","doi-asserted-by":"publisher","unstructured":"Cao, Y.H., Sun, P., Huang, Y., Wu, J., Zhou, S.: Synergistic self-supervised and quantization learning. In: European Conference on Computer Vision, pp. 587\u2013604. Springer, Heidelberg (2022). https:\/\/doi.org\/10.1007\/978-3-031-20056-4_34","DOI":"10.1007\/978-3-031-20056-4_34"},{"key":"18_CR5","doi-asserted-by":"publisher","unstructured":"Chen, H., et al.: Bootstrap generalization ability from loss landscape perspective. In: European Conference on Computer Vision, pp. 500\u2013517. Springer, Heidelberg (2022). https:\/\/doi.org\/10.1007\/978-3-031-25075-0_34","DOI":"10.1007\/978-3-031-25075-0_34"},{"key":"18_CR6","unstructured":"Chen, K., et\u00a0al.: Mmdetection: open mmlab detection toolbox and benchmark. arXiv preprint arXiv:1906.07155 (2019)"},{"key":"18_CR7","first-page":"14835","volume":"34","author":"K Choi","year":"2021","unstructured":"Choi, K., Hong, D., Park, N., Kim, Y., Lee, J.: Qimera: data-free quantization with synthetic boundary supporting samples. Adv. Neural. Inf. Process. Syst. 34, 14835\u201314847 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"18_CR8","doi-asserted-by":"crossref","unstructured":"Chuang, C.Y., et al.: Robust contrastive learning against noisy views. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 16670\u201316681 (2022)","DOI":"10.1109\/CVPR52688.2022.01617"},{"key":"18_CR9","unstructured":"Contributors, M.: MMSegmentation: openmmlab semantic segmentation toolbox and benchmark (2020). https:\/\/github.com\/open-mmlab\/mmsegmentation"},{"key":"18_CR10","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: Imagenet: a large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248\u2013255. IEEE (2009)","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"18_CR11","doi-asserted-by":"crossref","unstructured":"Dong, P., Li, L., Wei, Z., Niu, X., Tian, Z., Pan, H.: Emq: evolving training-free proxies for automated mixed precision quantization. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 17076\u201317086 (2023)","DOI":"10.1109\/ICCV51070.2023.01566"},{"key":"18_CR12","unstructured":"Dosovitskiy, A., et\u00a0al.: An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"issue":"10","key":"18_CR13","doi-asserted-by":"publisher","first-page":"5253","DOI":"10.1109\/TNNLS.2021.3069886","volume":"33","author":"W Fei","year":"2021","unstructured":"Fei, W., Dai, W., Li, C., Zou, J., Xiong, H.: General bitwidth assignment for efficient deep convolutional neural network quantization. IEEE Trans. Neural Netw. Learn. Syst. 33(10), 5253\u20135267 (2021)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"18_CR14","doi-asserted-by":"crossref","unstructured":"Frumkin, N., Gope, D., Marculescu, D.: Jumping through local minima: quantization in the loss landscape of vision transformers. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 16978\u201316988 (2023)","DOI":"10.1109\/ICCV51070.2023.01557"},{"key":"18_CR15","doi-asserted-by":"crossref","unstructured":"Fu, Y., Yu, Q., Li, M., Ouyang, X., Chandra, V., Lin, Y.: Contrastive quant: quantization makes stronger contrastive learning. In: Proceedings of the 59th ACM\/IEEE Design Automation Conference, pp. 205\u2013210 (2022)","DOI":"10.1145\/3489517.3530419"},{"key":"18_CR16","unstructured":"Huang, H., Yu, P.S., Wang, C.: An introduction to image synthesis with generative adversarial nets. arXiv preprint arXiv:1803.04469 (2018)"},{"key":"18_CR17","unstructured":"Hubara, I., Courbariaux, M., Soudry, D., El-Yaniv, R., Bengio, Y.: Binarized neural networks. Adv. Neural Inf. Process. Syst. 29 (2016)"},{"key":"18_CR18","unstructured":"Hubara, I., Nahshan, Y., Hanani, Y., Banner, R., Soudry, D.: Accurate post training quantization with small calibration sets. In: International Conference on Machine Learning, pp. 4466\u20134475. PMLR (2021)"},{"issue":"7","key":"18_CR19","doi-asserted-by":"publisher","first-page":"2925","DOI":"10.1109\/TNNLS.2020.3008996","volume":"32","author":"N Kim","year":"2020","unstructured":"Kim, N., Shin, D., Choi, W., Kim, G., Park, J.: Exploiting retraining-based mixed-precision quantization for low-cost dnn accelerator design. IEEE Trans. Neural Netw. Learn. Syst. 32(7), 2925\u20132938 (2020)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"18_CR20","first-page":"9181","volume":"34","author":"S Kundu","year":"2021","unstructured":"Kundu, S., Sun, Q., Fu, Y., Pedram, M., Beerel, P.: Analyzing the confidentiality of undistillable teachers in knowledge distillation. Adv. Neural. Inf. Process. Syst. 34, 9181\u20139192 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"18_CR21","doi-asserted-by":"crossref","unstructured":"Kundu, S., Wang, S., Sun, Q., Beerel, P.A., Pedram, M.: Bmpq: bit-gradient sensitivity-driven mixed-precision quantization of dnns from scratch. In: 2022 Design, Automation & Test in Europe Conference & Exhibition (DATE), pp. 588\u2013591. IEEE (2022)","DOI":"10.23919\/DATE54114.2022.9774740"},{"key":"18_CR22","unstructured":"Li, H., Xu, Z., Taylor, G., Studer, C., Goldstein, T.: Visualizing the loss landscape of neural nets. Adv. Neural Inf. Process. Syst. 31 (2018)"},{"key":"18_CR23","first-page":"34451","volume":"35","author":"Y Li","year":"2022","unstructured":"Li, Y., Xu, S., Zhang, B., Cao, X., Gao, P., Guo, G.: Q-vit: accurate and fully quantized low-bit vision transformer. Adv. Neural. Inf. Process. Syst. 35, 34451\u201334463 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"18_CR24","first-page":"12934","volume":"35","author":"Y Li","year":"2022","unstructured":"Li, Y., et al.: Efficientformer: vision transformers at mobilenet speed. Adv. Neural. Inf. Process. Syst. 35, 12934\u201312949 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"18_CR25","doi-asserted-by":"crossref","unstructured":"Li, Z., Chen, M., Xiao, J., Gu, Q.: Psaq-vit v2: toward accurate and general data-free quantization for vision transformers. IEEE Trans. Neural Netw. Learn. Syst. (2023)","DOI":"10.1109\/TNNLS.2023.3301007"},{"key":"18_CR26","doi-asserted-by":"crossref","unstructured":"Li, Z., Gu, Q.: I-vit: integer-only quantization for efficient vision transformer inference. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 17065\u201317075 (2023)","DOI":"10.1109\/ICCV51070.2023.01565"},{"key":"18_CR27","doi-asserted-by":"publisher","unstructured":"Li, Z., Ma, L., Chen, M., Xiao, J., Gu, Q.: Patch similarity aware data-free quantization for vision transformers. In: European Conference on Computer Vision, pp. 154\u2013170. Springer, Heidelberg (2022). https:\/\/doi.org\/10.1007\/978-3-031-20083-0_10","DOI":"10.1007\/978-3-031-20083-0_10"},{"key":"18_CR28","doi-asserted-by":"crossref","unstructured":"Li, Z., Xiao, J., Yang, L., Gu, Q.: Repq-vit: scale reparameterization for post-training quantization of vision transformers. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 17227\u201317236 (2023)","DOI":"10.1109\/ICCV51070.2023.01580"},{"key":"18_CR29","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1007\/978-3-319-10602-1_48","volume-title":"Computer Vision \u2013 ECCV 2014","author":"T-Y Lin","year":"2014","unstructured":"Lin, T.-Y., et al.: Microsoft COCO: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8693, pp. 740\u2013755. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10602-1_48"},{"key":"18_CR30","doi-asserted-by":"crossref","unstructured":"Lin, Y., Zhang, T., Sun, P., Li, Z., Zhou, S.: Fq-vit: post-training quantization for fully quantized vision transformer. arXiv preprint arXiv:2111.13824 (2021)","DOI":"10.24963\/ijcai.2022\/164"},{"key":"18_CR31","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"18_CR32","first-page":"28092","volume":"34","author":"Z Liu","year":"2021","unstructured":"Liu, Z., Wang, Y., Han, K., Zhang, W., Ma, S., Gao, W.: Post-training quantization for vision transformer. Adv. Neural. Inf. Process. Syst. 34, 28092\u201328103 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"18_CR33","unstructured":"Peters, J.W., Welling, M.: Probabilistic binary neural networks. arXiv preprint arXiv:1809.03368 (2018)"},{"key":"18_CR34","doi-asserted-by":"crossref","unstructured":"Ramachandran, A., Dhiman, A., Vandrotti, B.S., Kim, J.: Ntrans-net: a multi-scale neutrosophic-uncertainty guided transformer network for indoor depth completion. In: 2023 IEEE International Conference on Image Processing (ICIP), pp. 905\u2013909. IEEE (2023)","DOI":"10.1109\/ICIP49359.2023.10222228"},{"key":"18_CR35","doi-asserted-by":"crossref","unstructured":"Ramachandran, A., Wan, Z., Jeong, G., Gustafson, J., Krishna, T.: Algorithm-hardware co-design of distribution-aware logarithmic-posit encodings for efficient dnn inference. arXiv preprint arXiv:2403.05465 (2024)","DOI":"10.1145\/3649329.3656544"},{"key":"18_CR36","unstructured":"Ranjan, N., Savakis, A.: Lrp-qvit: mixed-precision vision transformer quantization via layer-wise relevance propagation. arXiv preprint arXiv:2401.11243 (2024)"},{"key":"18_CR37","doi-asserted-by":"crossref","unstructured":"Strudel, R., Garcia, R., Laptev, I., Schmid, C.: Segmenter: transformer for semantic segmentation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 7262\u20137272 (2021)","DOI":"10.1109\/ICCV48922.2021.00717"},{"key":"18_CR38","unstructured":"Touvron, H., Cord, M., Douze, M., Massa, F., Sablayrolles, A., J\u00e9gou, H.: Training data-efficient image transformers & distillation through attention. In: International conference on machine learning, pp. 10347\u201310357. PMLR (2021)"},{"key":"18_CR39","doi-asserted-by":"publisher","first-page":"13","DOI":"10.1016\/j.neucom.2023.02.028","volume":"535","author":"J Wang","year":"2023","unstructured":"Wang, J., Li, J., Li, W., Xuan, L., Zhang, T., Wang, W.: Positive-negative equal contrastive loss for semantic segmentation. Neurocomputing 535, 13\u201324 (2023)","journal-title":"Neurocomputing"},{"key":"18_CR40","doi-asserted-by":"publisher","unstructured":"Wightman, R.: Pytorch image models (2019). https:\/\/github.com\/rwightman\/pytorch-image-models. https:\/\/doi.org\/10.5281\/zenodo.4414861","DOI":"10.5281\/zenodo.4414861"},{"key":"18_CR41","doi-asserted-by":"crossref","unstructured":"Xiao, J., Li, Z., Yang, L., Gu, Q.: Patch-wise mixed-precision quantization of vision transformer. arXiv preprint arXiv:2305.06559 (2023)","DOI":"10.1109\/IJCNN54540.2023.10191205"},{"key":"18_CR42","doi-asserted-by":"crossref","unstructured":"Xiao, T., Liu, Y., Zhou, B., Jiang, Y., Sun, J.: Unified perceptual parsing for scene understanding. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 418\u2013434 (2018)","DOI":"10.1007\/978-3-030-01228-1_26"},{"key":"18_CR43","doi-asserted-by":"publisher","unstructured":"Yeh, C.H., Hong, C.Y., Hsu, Y.C., Liu, T.L., Chen, Y., LeCun, Y.: Decoupled contrastive learning. In: European Conference on Computer Vision. pp. 668\u2013684. Springer, Heidelberg (2022). https:\/\/doi.org\/10.1007\/978-3-031-19809-0_38","DOI":"10.1007\/978-3-031-19809-0_38"},{"key":"18_CR44","doi-asserted-by":"crossref","unstructured":"Yin, H., et al.: Dreaming to distill: data-free knowledge transfer via deepinversion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8715\u20138724 (2020)","DOI":"10.1109\/CVPR42600.2020.00874"},{"key":"18_CR45","doi-asserted-by":"publisher","unstructured":"Yuan, Z., Xue, C., Chen, Y., Wu, Q., Sun, G.: Ptq4vit: post-training quantization for vision transformers with twin uniform quantization. In: European Conference on Computer Vision, pp. 191\u2013207. Springer, Heidelberg (2022). https:\/\/doi.org\/10.1007\/978-3-031-19775-8_12","DOI":"10.1007\/978-3-031-19775-8_12"},{"key":"18_CR46","doi-asserted-by":"crossref","unstructured":"Zhang, D., Yang, J., Ye, D., Hua, G.: Lq-nets: learned quantization for highly accurate and compact deep neural networks. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 365\u2013382 (2018)","DOI":"10.1007\/978-3-030-01237-3_23"},{"key":"18_CR47","unstructured":"Zhang, S., Zhou, Q., Wang, Z., Wang, F., Yan, J.: Patch-level contrastive learning via positional query for visual pre-training (2023)"},{"key":"18_CR48","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Guo, X., Poggi, M., Zhu, Z., Huang, G., Mattoccia, S.: Completionformer: depth completion with convolutions and vision transformers. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 18527\u201318536 (2023)","DOI":"10.1109\/CVPR52729.2023.01777"},{"key":"18_CR49","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Chen, D., Kundu, S., Li, C., Beerel, P.A.: Sal-vit: towards latency efficient private inference on vit using selective attention search with a learnable softmax approximation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5116\u20135125 (2023)","DOI":"10.1109\/ICCV51070.2023.00472"},{"key":"18_CR50","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Lu, X., Cao, G., Yang, Y., Jiao, L., Liu, F.: Vit-yolo: transformer-based yolo for object detection. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 2799\u20132808 (2021)","DOI":"10.1109\/ICCVW54120.2021.00314"},{"key":"18_CR51","doi-asserted-by":"crossref","unstructured":"Zhong, Y., et al.: Intraq: learning synthetic images with intra-class heterogeneity for zero-shot network quantization. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12339\u201312348 (2022)","DOI":"10.1109\/CVPR52688.2022.01202"},{"key":"18_CR52","doi-asserted-by":"publisher","first-page":"302","DOI":"10.1007\/s11263-018-1140-0","volume":"127","author":"B Zhou","year":"2019","unstructured":"Zhou, B., et al.: Semantic understanding of scenes through the ade20k dataset. Int. J. Comput. Vision 127, 302\u2013321 (2019)","journal-title":"Int. J. Comput. Vision"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-72855-6_18","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,8]],"date-time":"2024-11-08T19:06:13Z","timestamp":1731092773000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-72855-6_18"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,9]]},"ISBN":["9783031728549","9783031728556"],"references-count":52,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-72855-6_18","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,9]]},"assertion":[{"value":"9 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}