{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,7,9]],"date-time":"2025-07-09T15:10:06Z","timestamp":1752073806754,"version":"3.41.2"},"publisher-location":"New York, NY, USA","reference-count":26,"publisher":"ACM","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,12,14]]},"DOI":"10.1145\/3719384.3719426","type":"proceedings-article","created":{"date-parts":[[2025,7,9]],"date-time":"2025-07-09T14:44:03Z","timestamp":1752072243000},"page":"292-298","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["ESPCA: An Efficient Spatial-Sensitive Partial Channel Attention Mechanism in CNNs"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0004-5531-8275","authenticated-orcid":false,"given":"Tianze","family":"Huang","sequence":"first","affiliation":[{"name":"Changshu Institute of Technology, Suzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-4132-3116","authenticated-orcid":false,"given":"Yan","family":"Qian","sequence":"additional","affiliation":[{"name":"Changshu Institute of Technology, Suzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7422-9877","authenticated-orcid":false,"given":"Gaofei","family":"Sun","sequence":"additional","affiliation":[{"name":"Changshu Institute of Technology, Suzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-2963-832X","authenticated-orcid":false,"given":"Jiahao","family":"Yu","sequence":"additional","affiliation":[{"name":"Changshu Institute of Technology, Suzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7807-2600","authenticated-orcid":false,"given":"Zhenjiang","family":"Qian","sequence":"additional","affiliation":[{"name":"Changshu Institute of Technology, Suzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4061-6067","authenticated-orcid":false,"given":"Feng","family":"Li","sequence":"additional","affiliation":[{"name":"Changshu Institute of Technology, Suzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-4946-6425","authenticated-orcid":false,"given":"Xiaobing","family":"Xian","sequence":"additional","affiliation":[{"name":"Changshu Institute of Technology, Suzhou, China"}]}],"member":"320","published-online":{"date-parts":[[2025,7,9]]},"reference":[{"key":"e_1_3_3_1_1_2","volume-title":"Very deep convolutional networks for large-scale image recognition.\u00a0arXiv preprint arXiv:1409.1556","author":"Simonyan K.","year":"2014","unstructured":"Simonyan, K., & Zisserman, A. 2014. Very deep convolutional networks for large-scale image recognition.\u00a0arXiv preprint arXiv:1409.1556 (2014)."},{"key":"e_1_3_3_1_2_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_3_3_1_3_2","volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition.11976-11986","author":"Liu Z.","year":"2022","unstructured":"Liu, Z., Mao, H., Wu, C. Y., et al. 2022. A ConvNet for the 2020s. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition.11976-11986."},{"key":"e_1_3_3_1_4_2","volume-title":"In\u00a0Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.4510-4520","author":"Sandler M.","year":"2018","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., & Chen, L. C. 2018. Mobilenetv2: Inverted residuals and linear bottlenecks. In\u00a0Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.4510-4520."},{"key":"e_1_3_3_1_5_2","volume-title":"In\u00a0Proceedings of the IEEE\/CVF International Conference on Computer Vision.1314-1324","author":"Howard A.","year":"2019","unstructured":"Howard, A., Sandler, M., Chu, G., Chen, L. C., Chen, B., Tan, M., ... & Adam, H. 2019. Searching for mobilenetv3. In\u00a0Proceedings of the IEEE\/CVF International Conference on Computer Vision.1314-1324."},{"key":"e_1_3_3_1_6_2","volume-title":"Yolov10: Real-time end-to-end object detection.\u00a0arXiv preprint arXiv:2405.14458","author":"Wang A.","year":"2024","unstructured":"Wang, A., Chen, H., Liu, L., Chen, K., Lin, Z., Han, J., & Ding, G. 2024. Yolov10: Real-time end-to-end object detection.\u00a0arXiv preprint arXiv:2405.14458 (2024)."},{"key":"e_1_3_3_1_7_2","volume-title":"Proceedings of the IEEE International Conference on Computer Vision.2980-2988","author":"Lin T. Y.","year":"2017","unstructured":"Lin, T. Y., Goyal, P., Girshick, R., He, K., & Doll\u00e1r, P. 2017. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision.2980-2988."},{"key":"e_1_3_3_1_8_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"e_1_3_3_1_9_2","volume-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2881-2890","author":"Zhao H.","year":"2017","unstructured":"Zhao, H., Shi, J., Qi, X., Wang, X., & Jia, J. 2017. Pyramid scene parsing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2881-2890."},{"key":"e_1_3_3_1_10_2","volume-title":"In\u00a0Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.7132-7141","author":"Hu J.","year":"2018","unstructured":"Hu, J., Shen, L., & Sun, G. 2018. Squeeze-and-excitation networks. In\u00a0Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.7132-7141."},{"key":"e_1_3_3_1_11_2","volume-title":"In\u00a0Proceedings of the European Conference on Computer Vision (ECCV).3-19","author":"Woo S.","year":"2018","unstructured":"Woo, S., Park, J., Lee, J. Y., & Kweon, I. S. 2018. Cbam: Convolutional block attention module. In\u00a0Proceedings of the European Conference on Computer Vision (ECCV).3-19."},{"key":"e_1_3_3_1_12_2","volume-title":"In\u00a0Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition.11534-11542","author":"Wang Q.","year":"2020","unstructured":"Wang, Q., Wu, B., Zhu, P., Li, P., Zuo, W., & Hu, Q. 2020. ECA-Net: Efficient channel attention for deep convolutional neural networks. In\u00a0Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition.11534-11542."},{"key":"e_1_3_3_1_13_2","volume-title":"In\u00a0Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition.13713-13722","author":"Hou Q.","year":"2021","unstructured":"Hou, Q., Zhou, D., & Feng, J. 2021. Coordinate attention for efficient mobile network design. In\u00a0Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition.13713-13722."},{"key":"e_1_3_3_1_14_2","volume-title":"Gather-excite: Exploiting feature context in convolutional neural networks.\u00a0Advances in neural information processing systems,31.","author":"Hu J.","year":"2018","unstructured":"Hu, J., Shen, L., Albanie, S., Sun, G., & Vedaldi, A. 2018. Gather-excite: Exploiting feature context in convolutional neural networks.\u00a0Advances in neural information processing systems,31."},{"key":"e_1_3_3_1_15_2","volume-title":"In\u00a0Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition.3024-3033","author":"Gao Z.","year":"2019","unstructured":"Gao, Z., Xie, J., Wang, Q., & Li, P. 2019. Global second-order pooling convolutional networks. In\u00a0Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition.3024-3033."},{"key":"e_1_3_3_1_16_2","doi-asserted-by":"crossref","unstructured":"Russakovsky O. Deng J. Su H. Krause J. Satheesh S. Ma S. ... & Fei-Fei L. 2015. Imagenet large scale visual recognition challenge.\u00a0International journal of computer vision \u00a0115 211-252.","DOI":"10.1007\/s11263-015-0816-y"},{"key":"e_1_3_3_1_17_2","unstructured":"Krizhevsky A. & Hinton G. 2009. Learning multiple layers of features from tiny images."},{"key":"e_1_3_3_1_18_2","doi-asserted-by":"crossref","unstructured":"Everingham M. Eslami S. A. Van Gool L. Williams C. K. Winn J. & Zisserman A. 2015. The Pascal visual object classes challenge: A retrospective.\u00a0International journal of computer vision \u00a0111 98-136.","DOI":"10.1007\/s11263-014-0733-5"},{"key":"e_1_3_3_1_19_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2011.6126343"},{"key":"e_1_3_3_1_20_2","volume-title":"In\u00a0Proceedings of the IEEE International Conference on Computer Vision.618-626","author":"Selvaraju R. R.","year":"2017","unstructured":"Selvaraju, R. R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., & Batra, D. 2017. Grad-cam: Visual explanations from deep networks via gradient-based localization. In\u00a0Proceedings of the IEEE International Conference on Computer Vision.618-626."},{"key":"e_1_3_3_1_21_2","first-page":"401","volume-title":"In\u00a0Fifteenth International Conference on Graphics and Image Processing (ICGIP","author":"Yu J.","year":"2024","unstructured":"Yu, J., Hu, Z., Sun, G., Xing, X., & Li, F. 2024. A novel channel attention module for integrity and importance of feature map. In\u00a0Fifteenth International Conference on Graphics and Image Processing (ICGIP 2023),\u00a0Vol. 13089, pp. 401-409. SPIE."},{"key":"e_1_3_3_1_22_2","volume-title":"SHViT: Single-Head Vision Transformer with Memory Efficient Macro Design.\u00a0arXiv preprint arXiv:2401.16456","author":"Yun S.","year":"2024","unstructured":"Yun, S., & Ro, Y. 2024. SHViT: Single-Head Vision Transformer with Memory Efficient Macro Design.\u00a0arXiv preprint arXiv:2401.16456 (2024)."},{"key":"e_1_3_3_1_23_2","volume-title":"In\u00a0Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition.12021-12031","author":"Chen J.","year":"2023","unstructured":"Chen, J., Kao, S. H., He, H., Zhuo, W., Wen, S., Lee, C. H., & Chan, S. H. G. 2023. Run, Don't walk: Chasing higher FLOPS for faster neural networks. In\u00a0Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition.12021-12031."},{"key":"e_1_3_3_1_24_2","volume-title":"In\u00a0Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition.1580-1589","author":"Han K.","year":"2020","unstructured":"Han, K., Wang, Y., Tian, Q., Guo, J., Xu, C., & Xu, C. 2020. Ghostnet: More features from cheap operations. In\u00a0Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition.1580-1589."},{"key":"e_1_3_3_1_25_2","volume-title":"Mobilenets: Efficient convolutional neural networks for mobile vision applications.\u00a0arXiv preprint arXiv:1704.04861.","author":"Howard A. G.","year":"2017","unstructured":"Howard, A. G., Zhu, M., Chen, B., Kalenichenko, D., Wang, W., Weyand, T., ... & Adam, H. 2017. Mobilenets: Efficient convolutional neural networks for mobile vision applications.\u00a0arXiv preprint arXiv:1704.04861. (2017)."},{"key":"e_1_3_3_1_26_2","unstructured":"Liu Z. Hao Z. Han K. Tang Y. & Wang Y. 2024. GhostNetV3: Exploring the Training Strategies for Compact Models.\u00a0arXiv preprint arXiv:2404.11202. (2024)."}],"event":{"name":"AICCC 2024: 2024 the 7th Artificial Intelligence and Cloud Computing Conference","location":"Tokyo Japan","acronym":"AICCC 2024"},"container-title":["Proceedings of the 2024 7th Artificial Intelligence and Cloud Computing Conference"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3719384.3719426","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,7,9]],"date-time":"2025-07-09T14:46:50Z","timestamp":1752072410000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3719384.3719426"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,14]]},"references-count":26,"alternative-id":["10.1145\/3719384.3719426","10.1145\/3719384"],"URL":"https:\/\/doi.org\/10.1145\/3719384.3719426","relation":{},"subject":[],"published":{"date-parts":[[2024,12,14]]},"assertion":[{"value":"2025-07-09","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}