{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T17:19:31Z","timestamp":1777655971642,"version":"3.51.4"},"publisher-location":"Cham","reference-count":55,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031726835","type":"print"},{"value":"9783031726842","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,11,3]],"date-time":"2024-11-03T00:00:00Z","timestamp":1730592000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,3]],"date-time":"2024-11-03T00:00:00Z","timestamp":1730592000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-72684-2_24","type":"book-chapter","created":{"date-parts":[[2024,11,2]],"date-time":"2024-11-02T19:06:34Z","timestamp":1730574394000},"page":"420-437","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":7,"title":["PQ-SAM: Post-training Quantization for\u00a0Segment Anything Model"],"prefix":"10.1007","author":[{"given":"Xiaoyu","family":"Liu","sequence":"first","affiliation":[]},{"given":"Xin","family":"Ding","sequence":"additional","affiliation":[]},{"given":"Lei","family":"Yu","sequence":"additional","affiliation":[]},{"given":"Yuanyuan","family":"Xi","sequence":"additional","affiliation":[]},{"given":"Wei","family":"Li","sequence":"additional","affiliation":[]},{"given":"Zhijun","family":"Tu","sequence":"additional","affiliation":[]},{"given":"Jie","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Hanting","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Baoqun","family":"Yin","sequence":"additional","affiliation":[]},{"given":"Zhiwei","family":"Xiong","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,3]]},"reference":[{"issue":"5","key":"24_CR1","doi-asserted-by":"publisher","first-page":"898","DOI":"10.1109\/TPAMI.2010.161","volume":"33","author":"P Arbelaez","year":"2010","unstructured":"Arbelaez, P., Maire, M., Fowlkes, C., Malik, J.: Contour detection and hierarchical image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 33(5), 898\u2013916 (2010)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"24_CR2","unstructured":"Brown, T., et\u00a0al.: Language models are few-shot learners. NeurIPS (2020)"},{"issue":"12","key":"24_CR3","doi-asserted-by":"publisher","first-page":"1247","DOI":"10.1038\/s41592-019-0612-7","volume":"16","author":"JC Caicedo","year":"2019","unstructured":"Caicedo, J.C., et al.: Nucleus segmentation across imaging experiments: the 2018 data science bowl. Nat. Methods 16(12), 1247\u20131253 (2019)","journal-title":"Nat. Methods"},{"key":"24_CR4","doi-asserted-by":"crossref","unstructured":"Cheng, H.K., Oh, S.W., Price, B., Schwing, A., Lee, J.Y.: Tracking anything with decoupled video segmentation. In: ICCV (2023)","DOI":"10.1109\/ICCV51070.2023.00127"},{"key":"24_CR5","unstructured":"Choi, J., Chuang, P.I.J., Wang, Z., Venkataramani, S., Srinivasan, V., Gopalakrishnan, K.: Bridging the accuracy gap for 2-bit quantized neural networks (QNN). arXiv preprint arXiv:1807.06964 (2018)"},{"key":"24_CR6","unstructured":"Choi, J., Wang, Z., Venkataramani, S., Chuang, P.I.J., Srinivasan, V., Gopalakrishnan, K.: PACT: parameterized clipping activation for quantized neural networks. arXiv preprint arXiv:1805.06085 (2018)"},{"key":"24_CR7","unstructured":"Chowdhery, A., et\u00a0al.: Palm: scaling language modeling with pathways. arXiv preprint arXiv:2204.02311 (2022)"},{"key":"24_CR8","doi-asserted-by":"crossref","unstructured":"Ciampi, L., Santiago, C., Costeira, J.P., Gennaro, C., Amato, G.: Domain adaptation for traffic density estimation. In: VISAPP (2021)","DOI":"10.5220\/0010303401850195"},{"key":"24_CR9","doi-asserted-by":"crossref","unstructured":"Cordts, M., et al.: The cityscapes dataset for semantic urban scene understanding. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.350"},{"key":"24_CR10","unstructured":"Ding, X., et\u00a0al.: CBQ: cross-block quantization for large language models. arXiv preprint arXiv:2312.07950 (2023)"},{"key":"24_CR11","doi-asserted-by":"crossref","unstructured":"Gong, S., et al.: 3DSAM-adapter: holistic adaptation of SAM from 2D to 3D for promptable medical image segmentation. arXiv preprint arXiv:2306.13465 (2023)","DOI":"10.1016\/j.media.2024.103324"},{"key":"24_CR12","doi-asserted-by":"crossref","unstructured":"Gupta, A., Dollar, P., Girshick, R.: LVIS: a dataset for large vocabulary instance segmentation. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00550"},{"key":"24_CR13","unstructured":"Han, S., Mao, H., Dally, W.J.: Deep compression: compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv preprint arXiv:1510.00149 (2015)"},{"key":"24_CR14","doi-asserted-by":"crossref","unstructured":"He, K., Chen, X., Xie, S., Li, Y., Doll\u00e1r, P., Girshick, R.: Masked autoencoders are scalable vision learners. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01553"},{"key":"24_CR15","unstructured":"Hinton, G., Vinyals, O., Dean, J.: Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 (2015)"},{"key":"24_CR16","doi-asserted-by":"crossref","unstructured":"Jacob, B., et al.: Quantization and training of neural networks for efficient integer-arithmetic-only inference. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00286"},{"key":"24_CR17","unstructured":"Jia, C., et al.: Scaling up visual and vision-language representation learning with noisy text supervision. In: ICML (2021)"},{"key":"24_CR18","unstructured":"Kingma, D.P., Ba, J.L.: Adam: a method for stochastic optimization. In: ICLR (2015)"},{"key":"24_CR19","doi-asserted-by":"crossref","unstructured":"Kirillov, A., et\u00a0al.: Segment anything. In: ICCV (2023)","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"24_CR20","doi-asserted-by":"crossref","unstructured":"Li, R., Wang, Y., Liang, F., Qin, H., Yan, J., Fan, R.: Fully quantized network for object detection. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00292"},{"key":"24_CR21","doi-asserted-by":"crossref","unstructured":"Li, Z., Gu, Q.: I-ViT: integer-only quantization for efficient vision transformer inference. In: ICCV (2023)","DOI":"10.1109\/ICCV51070.2023.01565"},{"key":"24_CR22","doi-asserted-by":"publisher","unstructured":"Li, Z., Ma, L., Chen, M., Xiao, J., Gu, Q.: Patch similarity aware data-free quantization for vision transformers. In: Avidan, S., Brostow, G., Cisse, M., Farinella, G.M., Hassner, T. (eds.) Computer Vision \u2013 ECCV 2022. ECCV 2022. LNCS, vol. 13671, pp. 154\u2013170. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-20083-0_10","DOI":"10.1007\/978-3-031-20083-0_10"},{"key":"24_CR23","doi-asserted-by":"crossref","unstructured":"Li, Z., Xiao, J., Yang, L., Gu, Q.: Repq-vit: scale reparameterization for post-training quantization of vision transformers. In: ICCV (2023)","DOI":"10.1109\/ICCV51070.2023.01580"},{"key":"24_CR24","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1007\/978-3-319-10602-1_48","volume-title":"Computer Vision \u2013 ECCV 2014","author":"T-Y Lin","year":"2014","unstructured":"Lin, T.-Y., et al.: Microsoft COCO: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8693, pp. 740\u2013755. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10602-1_48"},{"key":"24_CR25","doi-asserted-by":"crossref","unstructured":"Lin, Y., Zhang, T., Sun, P., Li, Z., Zhou, S.: Fq-vit: post-training quantization for fully quantized vision transformer. arXiv preprint arXiv:2111.13824 (2021)","DOI":"10.24963\/ijcai.2022\/164"},{"key":"24_CR26","doi-asserted-by":"crossref","unstructured":"Liu, X., et al.: Cross-dimension affinity distillation for 3D EM neuron segmentation. In: CVPR (2024)","DOI":"10.1109\/CVPR52733.2024.01056"},{"key":"24_CR27","doi-asserted-by":"publisher","unstructured":"Liu, X., Hu, B., Huang, W., Zhang, Y., Xiong, Z.: Efficient biomedical instance segmentation via knowledge distillation. In: Wang, L., Dou, Q., Fletcher, P.T., Speidel, S., Li, S. (eds.) Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2022. MICCAI 2022. LNCS, vol. 13434, pp. 14\u201324. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-16440-8_2","DOI":"10.1007\/978-3-031-16440-8_2"},{"key":"24_CR28","unstructured":"Liu, Z., Wang, Y., Han, K., Zhang, W., Ma, S., Gao, W.: Post-training quantization for vision transformer. In: NeurIPS (2021)"},{"key":"24_CR29","unstructured":"Loshchilov, I., Hutter, F.: SGDR: stochastic gradient descent with warm restarts. arXiv preprint arXiv:1608.03983 (2016)"},{"key":"24_CR30","unstructured":"Martin, D., Fowlkes, C., Tal, D., Malik, J.: A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In: ICCV (2001)"},{"key":"24_CR31","unstructured":"Paszke, A., et\u00a0al.: Pytorch: an imperative style, high-performance deep learning library. In: NeurIPS (2019)"},{"key":"24_CR32","first-page":"20","volume":"9","author":"M Pugliatti","year":"2022","unstructured":"Pugliatti, M., Topputo, F.: Doors: dataset for boulders segmentation. Zenodo 9, 20 (2022)","journal-title":"Zenodo"},{"key":"24_CR33","unstructured":"Radford, A., et\u00a0al.: Learning transferable visual models from natural language supervision. In: ICML (2021)"},{"key":"24_CR34","doi-asserted-by":"crossref","unstructured":"Ross, S.M.: Introductory Statistics. Academic Press, Cambridge (2017)","DOI":"10.1016\/B978-0-12-804317-2.00031-X"},{"key":"24_CR35","doi-asserted-by":"crossref","unstructured":"Shamir, R.R., Duchin, Y., Kim, J., Sapiro, G., Harel, N.: Continuous dice coefficient: a method for evaluating probabilistic segmentations. arXiv preprint arXiv:1906.11031 (2019)","DOI":"10.1101\/306977"},{"key":"24_CR36","unstructured":"Shao, W., et al.: Omniquant: omnidirectionally calibrated quantization for large language models. In: ICLR (2023)"},{"key":"24_CR37","unstructured":"Trotter, C., et al.: NDD20: a large-scale few-shot dolphin dataset for coarse and fine-grained categorisation. arXiv preprint arXiv:2005.13359 (2020)"},{"key":"24_CR38","doi-asserted-by":"crossref","unstructured":"Tu, Z., Hu, J., Chen, H., Wang, Y.: Toward accurate post-training quantization for image super resolution. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.00567"},{"key":"24_CR39","unstructured":"Upton, G., Cook, I.: Understanding Statistics. Oxford University Press, Oxford (1996)"},{"key":"24_CR40","unstructured":"Vasu, P.K.A., Gabriel, J., Zhu, J., Tuzel, O., Ranjan, A.: Fastvit: a fast hybrid vision transformer using structural reparameterization. In: ICCV (2023)"},{"key":"24_CR41","doi-asserted-by":"crossref","unstructured":"Wei, X., et al.: Outlier suppression+: accurate quantization of large language models by equivalent and effective shifting and scaling. In: EMNLP (2023)","DOI":"10.18653\/v1\/2023.emnlp-main.102"},{"key":"24_CR42","unstructured":"Xie, D., et al.: Edit everything: a text-guided generative system for images editing. arXiv preprint arXiv:2304.14006 (2023)"},{"key":"24_CR43","doi-asserted-by":"crossref","unstructured":"Xie, Y., Liao, Y.: Efficient-ViT: a light-weight classification model based on CNN and ViT. In: ICIGP (2023)","DOI":"10.1145\/3582649.3582676"},{"key":"24_CR44","doi-asserted-by":"crossref","unstructured":"Xiong, Y., et\u00a0al.: Efficientsam: leveraged masked image pretraining for efficient segment anything. In: CVPR (2024)","DOI":"10.1109\/CVPR52733.2024.01525"},{"key":"24_CR45","unstructured":"Yang, J., Gao, M., Li, Z., Gao, S., Wang, F., Zheng, F.: Track anything: segment anything meets videos. arXiv preprint arXiv:2304.11968 (2023)"},{"key":"24_CR46","unstructured":"Yang, L., et al.: iShape: a first step towards irregular shape instance segmentation. arXiv preprint arXiv:2109.15068 (2021)"},{"key":"24_CR47","doi-asserted-by":"publisher","unstructured":"Yang, X., Ye, J., Wang, X.: Factorizing knowledge in neural networks. In: Avidan, S., Brostow, G., Cisse, M., Farinella, G.M., Hassner, T. (eds.) Computer Vision \u2013 ECCV 2022. ECCV 2022. LNCS, vol. 13694, pp 73\u201391. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19830-4_5","DOI":"10.1007\/978-3-031-19830-4_5"},{"key":"24_CR48","first-page":"25739","volume":"35","author":"X Yang","year":"2022","unstructured":"Yang, X., Zhou, D., Liu, S., Ye, J., Wang, X.: Deep model reassembly. NeurIPS 35, 25739\u201325753 (2022)","journal-title":"NeurIPS"},{"key":"24_CR49","doi-asserted-by":"crossref","unstructured":"Yu, F., Huang, K., Wang, M., Cheng, Y., Chu, W., Cui, L.: Width & depth pruning for vision transformers. In: AAAI (2022)","DOI":"10.1609\/aaai.v36i3.20222"},{"key":"24_CR50","doi-asserted-by":"publisher","unstructured":"Yuan, Z., Xue, C., Chen, Y., Wu, Q., Sun, G.: PTQ4ViT: post-training quantization for vision transformers with twin uniform quantization. In: Avidan, S., Brostow, G., Cisse, M., Farinella, G.M., Hassner, T. (eds.) Computer Vision \u2013 ECCV 2022. ECCV 2022. LNCS, vol. 13672, pp. 191\u2013207. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19775-8_12","DOI":"10.1007\/978-3-031-19775-8_12"},{"key":"24_CR51","unstructured":"Zhang, C., et al.: Faster segment anything: towards lightweight sam for mobile applications. arXiv preprint arXiv:2306.14289 (2023)"},{"key":"24_CR52","unstructured":"Zhang, C., et\u00a0al.: One small step for generative ai, one giant leap for agi: a complete survey on chatgpt in aigc era. arXiv preprint arXiv:2304.06488 (2023)"},{"key":"24_CR53","doi-asserted-by":"crossref","unstructured":"Zhang, Q., et al.: Distilling semantic priors from SAM to efficient image restoration models. In: CVPR (2024)","DOI":"10.1109\/CVPR52733.2024.02401"},{"key":"24_CR54","unstructured":"Zhao, X., et al.: Fast segment anything. arXiv preprint arXiv:2306.12156 (2023)"},{"key":"24_CR55","doi-asserted-by":"crossref","unstructured":"Zwillinger, D., Kokoska, S.: CRC Standard Probability and Statistics Tables and Formulae. CRC Press, Boca Raton (1999)","DOI":"10.1201\/9780367802417"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-72684-2_24","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,2]],"date-time":"2024-11-02T19:11:42Z","timestamp":1730574702000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-72684-2_24"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,3]]},"ISBN":["9783031726835","9783031726842"],"references-count":55,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-72684-2_24","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,3]]},"assertion":[{"value":"3 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}