{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,22]],"date-time":"2025-10-22T23:35:26Z","timestamp":1761176126003,"version":"build-2065373602"},"reference-count":0,"publisher":"IOS Press","isbn-type":[{"value":"9781643686318","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,10,21]],"date-time":"2025-10-21T00:00:00Z","timestamp":1761004800000},"content-version":"unspecified","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,10,21]]},"abstract":"<jats:p>Post-training quantization (PTQ) is widely utilized in Vision Transformers (ViTs) for its computational efficiency and retraining elimination. However, the unique architecture of ViTs introduces significant quantization challenges. Dynamic fluctuations in channel activations, particularly post-LayerNorm, result in distributional mismatches. Additionally, the heavy-tailed nature of post-Softmax activations compromises the accurate representation of critical attention regions, vital for ViT performance. Moreover, weight quantization at low bit-widths leads to a loss of structural information, degrading global feature representation. To address these challenges, we introduce the Group-aware Collaborative Quantization framework (GCQ-ViT), which significantly improves both the accuracy and efficiency of ViT quantization. The GCQ-ViT framework integrates a novel dynamic perception grouping quantization mechanism to ensure distributional consistency within groups, thus reducing hardware expense. It also utilizes a self-adaptive displaced uniform log2 quantizer, optimizing shift factors and nonlinear intervals to enhance representation in high-density regions of post-Softmax activations. Additionally, we propose a dynamic dimension-aware error compensation method to correct quantization errors across channel dimensions using a residual mean compensation skill, ensuring robust feature preservation. Extensive experiments on image classification, object detection, and instance segmentation tasks demonstrate that GCQ-ViT outperforms the current leading PTQ methods, setting a new benchmark for ViT quantization.<\/jats:p>","DOI":"10.3233\/faia250834","type":"book-chapter","created":{"date-parts":[[2025,10,22]],"date-time":"2025-10-22T09:43:34Z","timestamp":1761126214000},"source":"Crossref","is-referenced-by-count":0,"title":["GCQ-ViT: Group-Aware Collaborative Post-Training Quantization for Vision Transformers"],"prefix":"10.3233","author":[{"given":"Pan","family":"Peng","sequence":"first","affiliation":[{"name":"School of Software and AI, Yunnan University, Kunming, China"}]},{"given":"Wenbin","family":"Guo","sequence":"additional","affiliation":[{"name":"School of Software and AI, Yunnan University, Kunming, China"}]},{"given":"Ping","family":"Wei","sequence":"additional","affiliation":[{"name":"School of Software and AI, Yunnan University, Kunming, China"}]},{"given":"Wei","family":"Zhou","sequence":"additional","affiliation":[{"name":"School of Software and AI, Yunnan University, Kunming, China"}]}],"member":"7437","container-title":["Frontiers in Artificial Intelligence and Applications","ECAI 2025"],"original-title":[],"link":[{"URL":"https:\/\/ebooks.iospress.nl\/pdf\/doi\/10.3233\/FAIA250834","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,22]],"date-time":"2025-10-22T09:43:34Z","timestamp":1761126214000},"score":1,"resource":{"primary":{"URL":"https:\/\/ebooks.iospress.nl\/doi\/10.3233\/FAIA250834"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,21]]},"ISBN":["9781643686318"],"references-count":0,"URL":"https:\/\/doi.org\/10.3233\/faia250834","relation":{},"ISSN":["0922-6389","1879-8314"],"issn-type":[{"value":"0922-6389","type":"print"},{"value":"1879-8314","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,10,21]]}}}