{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,4]],"date-time":"2026-04-04T17:40:49Z","timestamp":1775324449907,"version":"3.50.1"},"publisher-location":"Singapore","reference-count":44,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819533923","type":"print"},{"value":"9789819533930","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,11,1]],"date-time":"2025-11-01T00:00:00Z","timestamp":1761955200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,11,1]],"date-time":"2025-11-01T00:00:00Z","timestamp":1761955200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-3393-0_38","type":"book-chapter","created":{"date-parts":[[2025,10,31]],"date-time":"2025-10-31T09:51:26Z","timestamp":1761904286000},"page":"467-478","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["SNN-PAR: Energy Efficient Pedestrian Attribute Recognition via\u00a0Spiking Neural Networks"],"prefix":"10.1007","author":[{"given":"Haiyang","family":"Wang","sequence":"first","affiliation":[]},{"given":"Qian","family":"Zhu","sequence":"additional","affiliation":[]},{"given":"Mowen","family":"She","sequence":"additional","affiliation":[]},{"given":"Yabo","family":"Li","sequence":"additional","affiliation":[]},{"given":"Haoyu","family":"Song","sequence":"additional","affiliation":[]},{"given":"Minghe","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Jin","family":"Tang","sequence":"additional","affiliation":[]},{"given":"Xiao","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,11,1]]},"reference":[{"issue":"11","key":"38_CR1","doi-asserted-by":"publisher","first-page":"1949","DOI":"10.1109\/TMM.2015.2477680","volume":"17","author":"AH Abdulnabi","year":"2015","unstructured":"Abdulnabi, A.H., Wang, G., Lu, J., Jia, K.: Multi-task CNN model for attribute prediction. IEEE Trans. Multimedia 17(11), 1949\u20131959 (2015)","journal-title":"IEEE Trans. Multimedia"},{"key":"38_CR2","doi-asserted-by":"crossref","unstructured":"Burelo, K., Sharifshazileh, M., Krayenb\u00fchl, N., Ramantani, G., Indiveri, G., Sarnthein, J.: A spiking neural network (SNN) for detecting high frequency oscillations (HFOS) in the intraoperative ECOG (2020). https:\/\/arxiv.org\/abs\/2011.08783","DOI":"10.1038\/s41598-021-85827-w"},{"key":"38_CR3","doi-asserted-by":"publisher","first-page":"54","DOI":"10.1007\/s11263-014-0788-3","volume":"113","author":"Y Cao","year":"2015","unstructured":"Cao, Y., Chen, Y., Khosla, D.: Spiking deep convolutional neural networks for energy-efficient object recognition. Int. J. Comput. Vision 113, 54\u201366 (2015)","journal-title":"Int. J. Comput. Vision"},{"issue":"10","key":"38_CR4","doi-asserted-by":"publisher","first-page":"6994","DOI":"10.1109\/TCSVT.2022.3178144","volume":"32","author":"X Cheng","year":"2022","unstructured":"Cheng, X., Jia, M., Wang, Q., Zhang, J.: A simple visual-textual baseline for pedestrian attribute recognition. IEEE Trans. Circuits Syst. Video Technol. 32(10), 6994\u20137004 (2022)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"38_CR5","unstructured":"Chung, J., Gulcehre, C., Cho, K., Bengio, Y.: Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:1412.3555 (2014)"},{"issue":"1","key":"38_CR6","doi-asserted-by":"publisher","first-page":"6","DOI":"10.1007\/s44267-025-00078-x","volume":"3","author":"Y Deng","year":"2025","unstructured":"Deng, Y., Chen, Z., Li, C., Tang, J.: Uncertainty-aware coarse-to-fine alignment for text-image person retrieval. Vis. Intell. 3(1), 6 (2025)","journal-title":"Vis. Intell."},{"key":"38_CR7","doi-asserted-by":"crossref","unstructured":"Deng, Y., Luo, P., Loy, C.C., Tang, X.: Pedestrian attribute recognition at far distance. In: Proceedings of the 22nd ACM International Conference on Multimedia, pp. 789\u2013792 (2014)","DOI":"10.1145\/2647868.2654966"},{"key":"38_CR8","doi-asserted-by":"crossref","unstructured":"Duwek, H.C., Shalumov, A., Tsur, E.E.: Image reconstruction from neuromorphic event cameras using Laplacian-prediction and poisson integration with spiking and artificial neural networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1333\u20131341 (2021)","DOI":"10.1109\/CVPRW53098.2021.00147"},{"key":"38_CR9","unstructured":"Eshraghian, J.K., et al.: Training spiking neural networks using lessons from deep learning (2023). https:\/\/arxiv.org\/abs\/2109.12894"},{"key":"38_CR10","doi-asserted-by":"publisher","first-page":"49","DOI":"10.1109\/TMM.2020.3045286","volume":"24","author":"H Fan","year":"2022","unstructured":"Fan, H., Hu, H.M., Liu, S., Lu, W., Pu, S.: Correlation graph convolutional network for pedestrian attribute recognition. IEEE Trans. Multimedia 24, 49\u201360 (2022). https:\/\/doi.org\/10.1109\/TMM.2020.3045286","journal-title":"IEEE Trans. Multimedia"},{"key":"38_CR11","doi-asserted-by":"crossref","unstructured":"Fan, X., Zhang, Y., Lu, Y., Wang, H.: ParFormer: transformer-based multi-task network for pedestrian attribute recognition. IEEE Trans. Circ. Syst. Video Technol. 34, 411\u2013423 (2023)","DOI":"10.1109\/TCSVT.2023.3285411"},{"key":"38_CR12","doi-asserted-by":"crossref","unstructured":"Fang, W., Yu, Z., Chen, Y., Masquelier, T., Huang, T., Tian, Y.: Incorporating learnable membrane time constant to enhance learning of spiking neural networks. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 2661\u20132671 (2021)","DOI":"10.1109\/ICCV48922.2021.00266"},{"issue":"4","key":"38_CR13","doi-asserted-by":"publisher","first-page":"1088","DOI":"10.1007\/s11263-022-01591-y","volume":"130","author":"H Guo","year":"2022","unstructured":"Guo, H., Fan, X., Wang, S.: Visual attention consistency for human attribute recognition. Int. J. Comput. Vision 130(4), 1088\u20131106 (2022)","journal-title":"Int. J. Comput. Vision"},{"key":"38_CR14","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"38_CR15","doi-asserted-by":"crossref","unstructured":"Heo, B., Lee, M., Yun, S., Choi, J.Y.: Knowledge transfer via distillation of activation boundaries formed by hidden neurons. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a033, pp. 3779\u20133787 (2019)","DOI":"10.1609\/aaai.v33i01.33013779"},{"key":"38_CR16","unstructured":"Hinton, G.: Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 (2015)"},{"key":"38_CR17","doi-asserted-by":"crossref","unstructured":"Huang, Y., Zhang, Z., Wu, Q., Zhong, Y., Wang, L.: Attribute-guided pedestrian retrieval: bridging person re-id with internal attribute variability. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 17689\u201317699 (2024)","DOI":"10.1109\/CVPR52733.2024.01675"},{"key":"38_CR18","doi-asserted-by":"crossref","unstructured":"Jia, J., Chen, X., Huang, K.: Spatial and semantic consistency regularizations for pedestrian attribute recognition. arXiv e-prints arXiv:2109.05686 (2021)","DOI":"10.1109\/ICCV48922.2021.00100"},{"key":"38_CR19","doi-asserted-by":"publisher","unstructured":"Jia, J., Gao, N., He, F., Chen, X., Huang, K.: Learning disentangled attribute representations for robust pedestrian attribute recognition. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 36, no. 1, pp. 1069\u20131077 (2022). https:\/\/doi.org\/10.1609\/aaai.v36i1.19991, https:\/\/ojs.aaai.org\/index.php\/AAAI\/article\/view\/19991","DOI":"10.1609\/aaai.v36i1.19991"},{"key":"38_CR20","doi-asserted-by":"crossref","unstructured":"Kabilan, R., Muthukumaran, N.: A neuromorphic model for image recognition using SNN. In: 2021 6th International Conference on Inventive Computation Technologies (ICICT), pp. 720\u2013725. IEEE (2021)","DOI":"10.1109\/ICICT50816.2021.9358663"},{"key":"38_CR21","unstructured":"Li, D., Zhang, Z., Chen, X., Ling, H., Huang, K.: A richly annotated dataset for pedestrian attribute recognition. arXiv e-prints arXiv:1603.07054 (2016)"},{"key":"38_CR22","doi-asserted-by":"crossref","unstructured":"Li, Q., Zhao, X., He, R., Huang, K.: Visual-semantic graph reasoning for pedestrian attribute recognition. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a033, pp. 8634\u20138641 (2019)","DOI":"10.1609\/aaai.v33i01.33018634"},{"key":"38_CR23","doi-asserted-by":"crossref","unstructured":"Li, Y., et al.: ATTMOT: improving multiple-object tracking by introducing auxiliary pedestrian attributes. IEEE Trans. Neural Netw. Learn. Syst. 36, 5454\u20135468 (2024)","DOI":"10.1109\/TNNLS.2024.3384446"},{"key":"38_CR24","doi-asserted-by":"publisher","first-page":"151","DOI":"10.1016\/j.patcog.2019.06.006","volume":"95","author":"Y Lin","year":"2019","unstructured":"Lin, Y., et al.: Improving person re-identification by attribute and identity learning. Pattern Recogn. 95, 151\u2013161 (2019)","journal-title":"Pattern Recogn."},{"key":"38_CR25","doi-asserted-by":"crossref","unstructured":"Liu, X., et al.: HydraPlus-Net: attentive deep features for pedestrian analysis. arXiv e-prints arXiv:1709.09930 (2017)","DOI":"10.1109\/ICCV.2017.46"},{"key":"38_CR26","unstructured":"Romero, A., Ballas, N., Kahou, S.E., Chassang, A., Gatta, C., Bengio, Y.: FitNets: hints for thin deep nets. arXiv preprint arXiv:1412.6550 (2014)"},{"key":"38_CR27","doi-asserted-by":"crossref","unstructured":"Chen, L., Song, J., Zhang, X., Shang, M.: MCFL: multi-label contrastive focal loss for deep imbalanced pedestrian attribute recognition. Neural Comput. Appl. 34, 16701\u201316715 (2022)","DOI":"10.1007\/s00521-022-07300-7"},{"key":"38_CR28","doi-asserted-by":"publisher","first-page":"159","DOI":"10.1016\/j.neucom.2022.05.028","volume":"497","author":"Z Tang","year":"2022","unstructured":"Tang, Z., Huang, J.: DRFormer: learning dual relations using transformer for pedestrian attribute recognition. Neurocomputing 497, 159\u2013169 (2022)","journal-title":"Neurocomputing"},{"key":"38_CR29","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, pp. 6000\u20136010 (2017)"},{"key":"38_CR30","doi-asserted-by":"crossref","unstructured":"Wang, J., Yang, Y., Mao, J., Huang, Z., Huang, C., Xu, W.: CNN-RNN: a unified framework for multi-label image classification. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2285\u20132294 (2016)","DOI":"10.1109\/CVPR.2016.251"},{"key":"38_CR31","doi-asserted-by":"crossref","unstructured":"Wang, J., Zhu, X., Gong, S., Li, W.: Attribute recognition by joint recurrent learning of context and correlation. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 531\u2013540 (2017)","DOI":"10.1109\/ICCV.2017.65"},{"key":"38_CR32","doi-asserted-by":"crossref","unstructured":"Wang, X., Jin, J., Li, C., Tang, J., Zhang, C., Wang, W.: Pedestrian attribute recognition via clip based prompt vision-language fusion. IEEE Trans. Circ. Syst. Video Technol. 35, 148\u2013161 (2024)","DOI":"10.1109\/TCSVT.2024.3454366"},{"issue":"1","key":"38_CR33","doi-asserted-by":"publisher","first-page":"8","DOI":"10.1007\/s44267-025-00080-3","volume":"3","author":"X Wang","year":"2025","unstructured":"Wang, X., et al.: Pre-training on high-resolution x-ray images: an experimental study. Vis. Intell. 3(1), 8 (2025)","journal-title":"Vis. Intell."},{"key":"38_CR34","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2021.108220","volume":"121","author":"X Wang","year":"2022","unstructured":"Wang, X., et al.: Pedestrian attribute recognition: a survey. Pattern Recogn. 121, 108220 (2022)","journal-title":"Pattern Recogn."},{"key":"38_CR35","doi-asserted-by":"publisher","unstructured":"Wu, J., Huang, Y., Gao, Z., Hong, Y., Zhao, J., Du, X.: Inter-attribute awareness for pedestrian attribute recognition. Pattern Recognit. 131, 108865 (2022). https:\/\/doi.org\/10.1016\/j.patcog.2022.108865, https:\/\/www.sciencedirect.com\/science\/article\/pii\/S0031320322003466","DOI":"10.1016\/j.patcog.2022.108865"},{"issue":"10","key":"38_CR36","doi-asserted-by":"publisher","first-page":"2731","DOI":"10.1007\/s11263-021-01499-z","volume":"129","author":"Y Yang","year":"2021","unstructured":"Yang, Y., et al.: Cascaded split-and-aggregate learning with feature recombination for pedestrian attribute recognition. Int. J. Comput. Vision 129(10), 2731\u20132744 (2021)","journal-title":"Int. J. Comput. Vision"},{"key":"38_CR37","doi-asserted-by":"publisher","unstructured":"Yang, Z., Li, Z., Shao, M., Shi, D., Yuan, Z., Yuan, C.: Masked generative distillation. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) European Conference on Computer Vision, pp. 53\u201369. Springer (2022). https:\/\/doi.org\/10.1007\/978-3-031-20083-0_4","DOI":"10.1007\/978-3-031-20083-0_4"},{"key":"38_CR38","unstructured":"Zagoruyko, S., Komodakis, N.: Paying more attention to attention: improving the performance of convolutional neural networks via attention transfer. arXiv preprint arXiv:1612.03928 (2016)"},{"key":"38_CR39","doi-asserted-by":"crossref","unstructured":"Zeng, Y., et\u00a0al.: BrainCog: a spiking neural network based, brain-inspired cognitive intelligence engine for brain-inspired AI and brain simulation. Patterns 4(8), 100789 (2023)","DOI":"10.1016\/j.patter.2023.100789"},{"key":"38_CR40","doi-asserted-by":"publisher","unstructured":"Zheng, A., Wang, H., Wang, J., Huang, H., He, R., Hussain, A.: Diverse features discovery transformer for pedestrian attribute recognition. Eng. Appl. Artif. Intell. 119, 105708 (2023). https:\/\/doi.org\/10.1016\/j.engappai.2022.105708, https:\/\/www.sciencedirect.com\/science\/article\/pii\/S0952197622006984","DOI":"10.1016\/j.engappai.2022.105708"},{"key":"38_CR41","unstructured":"Zhou, C., et al.: SpikingFormer: spike-driven residual learning for transformer-based spiking neural network. arXiv preprint arXiv:2304.11954 (2023)"},{"key":"38_CR42","unstructured":"Zhou, H., et al.: Rethinking soft labels for knowledge distillation: a bias-variance tradeoff perspective. arXiv preprint arXiv:2102.00650 (2021)"},{"key":"38_CR43","doi-asserted-by":"crossref","unstructured":"Zhou, S., Li, X., Chen, Y., Chandrasekaran, S.T., Sanyal, A.: Temporal-coded deep spiking neural network with easy training and robust performance. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a035, pp. 11143\u201311151 (2021)","DOI":"10.1609\/aaai.v35i12.17329"},{"key":"38_CR44","unstructured":"Zhou, Z., et al.: SpikFormer: when spiking neural network meets transformer. arXiv preprint arXiv:2209.15425 (2022)"}],"container-title":["Lecture Notes in Computer Science","Image and Graphics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-3393-0_38","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,4,4]],"date-time":"2026-04-04T16:30:32Z","timestamp":1775320232000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-3393-0_38"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,1]]},"ISBN":["9789819533923","9789819533930"],"references-count":44,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-3393-0_38","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,11,1]]},"assertion":[{"value":"1 November 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICIG","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Image and Graphics","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Xuzhou","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"31 October 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2 November 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"13","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icig2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/icig.csig.org.cn\/2025\/index.html","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}