{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T18:24:04Z","timestamp":1772907844003,"version":"3.50.1"},"reference-count":56,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2024,12,1]],"date-time":"2024-12-01T00:00:00Z","timestamp":1733011200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2024,12,1]],"date-time":"2024-12-01T00:00:00Z","timestamp":1733011200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2024,12,1]],"date-time":"2024-12-01T00:00:00Z","timestamp":1733011200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2024,12,1]],"date-time":"2024-12-01T00:00:00Z","timestamp":1733011200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2024,12,1]],"date-time":"2024-12-01T00:00:00Z","timestamp":1733011200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2024,12,1]],"date-time":"2024-12-01T00:00:00Z","timestamp":1733011200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,12,1]],"date-time":"2024-12-01T00:00:00Z","timestamp":1733011200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62088102"],"award-info":[{"award-number":["62088102"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Pattern Recognition"],"published-print":{"date-parts":[[2024,12]]},"DOI":"10.1016\/j.patcog.2024.110698","type":"journal-article","created":{"date-parts":[[2024,6,21]],"date-time":"2024-06-21T16:18:52Z","timestamp":1718986732000},"page":"110698","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":9,"special_numbering":"C","title":["FMGNet: An efficient feature-multiplex group network for real-time vision task"],"prefix":"10.1016","volume":"156","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-3572-7053","authenticated-orcid":false,"given":"Hao","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Yongqiang","family":"Ma","sequence":"additional","affiliation":[]},{"given":"Kaipeng","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Nanning","family":"Zheng","sequence":"additional","affiliation":[]},{"given":"Shenqi","family":"Lai","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.patcog.2024.110698_b1","doi-asserted-by":"crossref","unstructured":"Mark Sandler, Andrew Howard, Menglong Zhu, Andrey Zhmoginov, Liang-Chieh Chen, MobileNetV2: Inverted residuals and linear bottlenecks, in: CVPR, 2018, pp. 4510\u20134520.","DOI":"10.1109\/CVPR.2018.00474"},{"key":"10.1016\/j.patcog.2024.110698_b2","doi-asserted-by":"crossref","first-page":"272","DOI":"10.1016\/j.patcog.2018.10.029","article-title":"LightweightNet: Toward fast and lightweight convolutional neural networks via architecture distillation","volume":"88","author":"Xu","year":"2019","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.patcog.2024.110698_b3","series-title":"ICML","first-page":"6105","article-title":"EfficientNet: Rethinking model scaling for convolutional neural networks","author":"Tan","year":"2019"},{"key":"10.1016\/j.patcog.2024.110698_b4","doi-asserted-by":"crossref","unstructured":"Ilija Radosavovic, Raj Prateek Kosaraju, Ross Girshick, Kaiming He, Piotr Doll\u00e1r, Designing network design spaces, in: CVPR, 2020, pp. 10428\u201310436.","DOI":"10.1109\/CVPR42600.2020.01044"},{"key":"10.1016\/j.patcog.2024.110698_b5","doi-asserted-by":"crossref","unstructured":"Pavan Kumar Anasosalu Vasu, James Gabriel, Jeff Zhu, Oncel Tuzel, Anurag Ranjan, MobileOne: An Improved One Millisecond Mobile Backbone, in: CVPR, 2023, pp. 7907\u20137917.","DOI":"10.1109\/CVPR52729.2023.00764"},{"key":"10.1016\/j.patcog.2024.110698_b6","article-title":"Scgnet: Shifting and cascaded group network","author":"Zhang","year":"2023","journal-title":"IEEE Trans. Circuit Syst. Video Technol."},{"key":"10.1016\/j.patcog.2024.110698_b7","series-title":"ICLR","article-title":"Do wide and deep networks learn the same things? Uncovering how neural network representations vary with width and depth","author":"Nguyen","year":"2021"},{"key":"10.1016\/j.patcog.2024.110698_b8","unstructured":"Vinod Nair, Geoffrey E. Hinton, Rectified linear units improve restricted boltzmann machines, in: ICML, 2010."},{"key":"10.1016\/j.patcog.2024.110698_b9","doi-asserted-by":"crossref","unstructured":"Jie Hu, Li Shen, Gang Sun, Squeeze-and-Excitation Networks, in: CVPR, 2018, pp. 7132\u20137141.","DOI":"10.1109\/CVPR.2018.00745"},{"key":"10.1016\/j.patcog.2024.110698_b10","doi-asserted-by":"crossref","unstructured":"Andrew Howard, Mark Sandler, Grace Chu, Liang-Chieh Chen, Bo Chen, Mingxing Tan, Weijun Wang, Yukun Zhu, Ruoming Pang, Vijay Vasudevan, et al., Searching for MobileNetV3, in: ICCV, 2019, pp. 1314\u20131324.","DOI":"10.1109\/ICCV.2019.00140"},{"key":"10.1016\/j.patcog.2024.110698_b11","series-title":"Very deep convolutional networks for large-scale image recognition","author":"Simonyan","year":"2014"},{"key":"10.1016\/j.patcog.2024.110698_b12","doi-asserted-by":"crossref","unstructured":"Christian Szegedy, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed, Dragomir Anguelov, Dumitru Erhan, Vincent Vanhoucke, Andrew Rabinovich, Going deeper with convolutions, in: CVPR, 2015, pp. 1\u20139.","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"10.1016\/j.patcog.2024.110698_b13","unstructured":"Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun, Deep residual learning for image recognition, in: CVPR, 2016, pp. 770\u2013778."},{"key":"10.1016\/j.patcog.2024.110698_b14","doi-asserted-by":"crossref","unstructured":"Ke Sun, Bin Xiao, Dong Liu, Jingdong Wang, Deep high-resolution representation learning for human pose estimation, in: CVPR, 2019, pp. 5693\u20135703.","DOI":"10.1109\/CVPR.2019.00584"},{"key":"10.1016\/j.patcog.2024.110698_b15","unstructured":"Zhuang Liu, Hanzi Mao, Chao-Yuan Wu, Christoph Feichtenhofer, Trevor Darrell, Saining Xie, A convnet for the 2020s, in: CVPR, 2022, pp. 11976\u201311986."},{"key":"10.1016\/j.patcog.2024.110698_b16","doi-asserted-by":"crossref","unstructured":"Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin, Baining Guo, Swin transformer: Hierarchical vision transformer using shifted windows, in: ICCV, 2021, pp. 10012\u201310022.","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"10.1016\/j.patcog.2024.110698_b17","series-title":"ICML","first-page":"10347","article-title":"Training data-efficient image transformers & distillation through attention","author":"Touvron","year":"2021"},{"key":"10.1016\/j.patcog.2024.110698_b18","unstructured":"Sanghyun Woo, Shoubhik Debnath, Ronghang Hu, Xinlei Chen, Zhuang Liu, In So Kweon, Saining Xie, Convnext v2: Co-designing and scaling convnets with masked autoencoders, in: CVPR, 2023, pp. 16133\u201316142."},{"key":"10.1016\/j.patcog.2024.110698_b19","doi-asserted-by":"crossref","unstructured":"Olaf Ronneberger, Philipp Fischer, Thomas Brox, U-Net: Convolutional networks for biomedical image segmentation, in: International Conference on Medical Image Computing and Computer-Assisted Intervention, 2015, pp. 234\u2013241.","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"10.1016\/j.patcog.2024.110698_b20","series-title":"SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and\u00a1 0.5 MB model size","author":"Iandola","year":"2016"},{"key":"10.1016\/j.patcog.2024.110698_b21","doi-asserted-by":"crossref","unstructured":"Amir Gholami, Kiseok Kwon, Bichen Wu, Zizheng Tai, Xiangyu Yue, Peter Jin, Sicheng Zhao, Kurt Keutzer, SqueezeNext: Hardware-aware neural network design, in: CVPR Workshops, 2018, pp. 1638\u20131647.","DOI":"10.1109\/CVPRW.2018.00215"},{"key":"10.1016\/j.patcog.2024.110698_b22","series-title":"MobileNets: Efficient convolutional neural networks for mobile vision applications","author":"Howard","year":"2017"},{"key":"10.1016\/j.patcog.2024.110698_b23","doi-asserted-by":"crossref","unstructured":"Xiangyu Zhang, Xinyu Zhou, Mengxiao Lin, Jian Sun, ShuffleNet: An extremely efficient convolutional neural network for mobile devices, in: CVPR, 2018, pp. 6848\u20136856.","DOI":"10.1109\/CVPR.2018.00716"},{"key":"10.1016\/j.patcog.2024.110698_b24","unstructured":"Ningning Ma, Xiangyu Zhang, Hai-Tao Zheng, Jian Sun, ShuffleNet V2: Practical guidelines for efficient cnn architecture design, in: ECCV, 2018, pp. 116\u2013131."},{"key":"10.1016\/j.patcog.2024.110698_b25","doi-asserted-by":"crossref","unstructured":"Esteban Real, Alok Aggarwal, Yanping Huang, Quoc V. Le, Regularized evolution for image classifier architecture search, in: AAAI, Vol. 33, 2019, pp. 4780\u20134789.","DOI":"10.1609\/aaai.v33i01.33014780"},{"key":"10.1016\/j.patcog.2024.110698_b26","unstructured":"Yanyu Li, Ju Hu, Yang Wen, Georgios Evangelidis, Kamyar Salahi, Yanzhi Wang, Sergey Tulyakov, Jian Ren, Rethinking vision transformers for mobilenet size and speed, in: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2023, pp. 16889\u201316900."},{"key":"10.1016\/j.patcog.2024.110698_b27","doi-asserted-by":"crossref","unstructured":"Junting Pan, Adrian Bulat, Fuwen Tan, Xiatian Zhu, Lukasz Dudziak, Hongsheng Li, Georgios Tzimiropoulos, Brais Martinez, Edgevits: Competing light-weight cnns on mobile devices with vision transformers, in: ECCV, 2022, pp. 294\u2013311.","DOI":"10.1007\/978-3-031-20083-0_18"},{"key":"10.1016\/j.patcog.2024.110698_b28","series-title":"FastViT: A fast hybrid vision transformer using structural reparameterization","author":"Vasu","year":"2023"},{"key":"10.1016\/j.patcog.2024.110698_b29","doi-asserted-by":"crossref","unstructured":"Mustafa Munir, William Avery, Radu Marculescu, MobileViG: Graph-Based Sparse Attention for Mobile Vision Applications, in: CVPR, 2023, pp. 2210\u20132218.","DOI":"10.1109\/CVPRW59228.2023.00215"},{"key":"10.1016\/j.patcog.2024.110698_b30","series-title":"SwiftFormer: Efficient additive attention for transformer-based real-time mobile vision applications","author":"Shaker","year":"2023"},{"key":"10.1016\/j.patcog.2024.110698_b31","series-title":"2021 International Conference on Advanced Mechatronic Systems","first-page":"75","article-title":"A high-efficiency dirty-egg detection system based on YOLOv4 and TensorRT","author":"Wang","year":"2021"},{"key":"10.1016\/j.patcog.2024.110698_b32","doi-asserted-by":"crossref","unstructured":"Wei Liu, Dragomir Anguelov, Dumitru Erhan, Christian Szegedy, Scott E. Reed, Cheng-Yang Fu, Alexander C. Berg, SSD: Single Shot MultiBox Detector, in: Bastian Leibe, Jiri Matas, Nicu Sebe, Max Welling (Eds.), ECCV, Vol. 9905, 2016, pp. 21\u201337.","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"10.1016\/j.patcog.2024.110698_b33","doi-asserted-by":"crossref","unstructured":"Bin Xiao, Haiping Wu, Yichen Wei, Simple baselines for human pose estimation and tracking, in: ECCV, 2018, pp. 466\u2013481.","DOI":"10.1007\/978-3-030-01231-1_29"},{"key":"10.1016\/j.patcog.2024.110698_b34","series-title":"ACM MM","first-page":"274","article-title":"Learning discriminative features with multiple granularities for person re-identification","author":"Wang","year":"2018"},{"key":"10.1016\/j.patcog.2024.110698_b35","doi-asserted-by":"crossref","unstructured":"Liang-Chieh Chen, Yukun Zhu, George Papandreou, Florian Schroff, Hartwig Adam, Encoder-decoder with atrous separable convolution for semantic image segmentation, in: ECCV, 2018, pp. 801\u2013818.","DOI":"10.1007\/978-3-030-01234-2_49"},{"key":"10.1016\/j.patcog.2024.110698_b36","series-title":"CVPR","first-page":"248","article-title":"ImageNet: A large-scale hierarchical image database","author":"Deng","year":"2009"},{"key":"10.1016\/j.patcog.2024.110698_b37","doi-asserted-by":"crossref","unstructured":"Barret Zoph, Vijay Vasudevan, Jonathon Shlens, Quoc V. Le, Learning transferable architectures for scalable image recognition, in: CVPR, 2018, pp. 8697\u20138710.","DOI":"10.1109\/CVPR.2018.00907"},{"key":"10.1016\/j.patcog.2024.110698_b38","series-title":"DARTS: Differentiable architecture search","author":"Liu","year":"2018"},{"key":"10.1016\/j.patcog.2024.110698_b39","doi-asserted-by":"crossref","unstructured":"Chenxi Liu, Barret Zoph, Maxim Neumann, Jonathon Shlens, Wei Hua, Li-Jia Li, Li Fei-Fei, Alan L. Yuille, Jonathan Huang, Kevin Murphy, Progressive Neural Architecture Search, in: Vittorio Ferrari, Martial Hebert, Cristian Sminchisescu, Yair Weiss (Eds.), ECCV, Vol. 11205, 2018, pp. 19\u201335.","DOI":"10.1007\/978-3-030-01246-5_2"},{"key":"10.1016\/j.patcog.2024.110698_b40","series-title":"IGCV3: Interleaved low-rank group convolutions for efficient deep neural networks","author":"Sun","year":"2018"},{"key":"10.1016\/j.patcog.2024.110698_b41","doi-asserted-by":"crossref","unstructured":"Daquan Zhou, Qibin Hou, Yunpeng Chen, Jiashi Feng, Shuicheng Yan, Rethinking bottleneck structure for efficient mobile network design, in: ECCV, 2020, pp. 680\u2013697.","DOI":"10.1007\/978-3-030-58580-8_40"},{"key":"10.1016\/j.patcog.2024.110698_b42","article-title":"HF-HRNet: a simple hardware friendly high-resolution network","author":"Zhang","year":"2024","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10.1016\/j.patcog.2024.110698_b43","first-page":"1","article-title":"Open-vocabulary animal keypoint detection with semantic-feature matching","author":"Zhang","year":"2024","journal-title":"International Journal of Computer Vision"},{"key":"10.1016\/j.patcog.2024.110698_b44","doi-asserted-by":"crossref","unstructured":"Yipeng Chen, Cairong Zhao, Tianli Sun, Single Image Based Metric Learning via Overlapping Blocks Model for Person Re-Identification, in: CVPR, 2019, pp. 647\u2013656.","DOI":"10.1109\/CVPRW.2019.00091"},{"key":"10.1016\/j.patcog.2024.110698_b45","doi-asserted-by":"crossref","unstructured":"Yunhang Shen, Rongrong Ji, Xiaopeng Hong, Feng Zheng, Xiaowei Guo, Yongjian Wu, Feiyue Huang, A Part Power Set Model for Scale-Free Person Retrieval, in: IJCAI, 2019, pp. 3397\u20133403.","DOI":"10.24963\/ijcai.2019\/471"},{"key":"10.1016\/j.patcog.2024.110698_b46","doi-asserted-by":"crossref","unstructured":"Binghui Chen, Weihong Deng, Jiani Hu, Mixed high-order attention network for person re-identification, in: ICCV, 2019, pp. 371\u2013381.","DOI":"10.1109\/ICCV.2019.00046"},{"issue":"10","key":"10.1016\/j.patcog.2024.110698_b47","doi-asserted-by":"crossref","first-page":"3433","DOI":"10.1109\/TCSVT.2019.2957467","article-title":"Complementation-reinforced attention network for person re-identification","volume":"30","author":"Han","year":"2019","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10.1016\/j.patcog.2024.110698_b48","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2021.108432","article-title":"Loss function search for person re-identification","volume":"124","author":"Gu","year":"2022","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.patcog.2024.110698_b49","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2022.108899","article-title":"Dimension-aware attention for efficient mobile networks","volume":"131","author":"Mo","year":"2022","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.patcog.2024.110698_b50","doi-asserted-by":"crossref","unstructured":"Hao Luo, Youzhi Gu, Xingyu Liao, Shenqi Lai, Wei Jiang, Bag of tricks and a strong baseline for deep person re-identification, in: CVPR Workshops, 2019.","DOI":"10.1109\/CVPRW.2019.00190"},{"key":"10.1016\/j.patcog.2024.110698_b51","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2022.109151","article-title":"Similarity learning with deep CRF for person re-identification","volume":"135","author":"Xiang","year":"2023","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.patcog.2024.110698_b52","doi-asserted-by":"crossref","unstructured":"Wei Li, Rui Zhao, Tong Xiao, Xiaogang Wang, Deepreid: Deep filter pairing neural network for person re-identification, in: CVPR, 2014, pp. 152\u2013159.","DOI":"10.1109\/CVPR.2014.27"},{"key":"10.1016\/j.patcog.2024.110698_b53","doi-asserted-by":"crossref","unstructured":"Zhun Zhong, Liang Zheng, Donglin Cao, Shaozi Li, Re-ranking person re-identification with k-reciprocal encoding, in: CVPR, 2017, pp. 1318\u20131327.","DOI":"10.1109\/CVPR.2017.389"},{"key":"10.1016\/j.patcog.2024.110698_b54","doi-asserted-by":"crossref","unstructured":"Ergys Ristani, Francesco Solera, Roger Zou, Rita Cucchiara, Carlo Tomasi, Performance measures and a data set for multi-target, multi-camera tracking, in: ECCV, 2016, pp. 17\u201335.","DOI":"10.1007\/978-3-319-48881-3_2"},{"key":"10.1016\/j.patcog.2024.110698_b55","doi-asserted-by":"crossref","unstructured":"Liang Zheng, Liyue Shen, Lu Tian, Shengjin Wang, Jingdong Wang, Qi Tian, Scalable person re-identification: A benchmark, in: ICCV, 2015, pp. 1116\u20131124.","DOI":"10.1109\/ICCV.2015.133"},{"key":"10.1016\/j.patcog.2024.110698_b56","doi-asserted-by":"crossref","unstructured":"Marius Cordts, Mohamed Omran, Sebastian Ramos, Timo Rehfeld, Markus Enzweiler, Rodrigo Benenson, Uwe Franke, Stefan Roth, Bernt Schiele, The cityscapes dataset for semantic urban scene understanding, in: CVPR, 2016, pp. 3213\u20133223.","DOI":"10.1109\/CVPR.2016.350"}],"container-title":["Pattern Recognition"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0031320324004497?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0031320324004497?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2024,11,22]],"date-time":"2024-11-22T11:20:55Z","timestamp":1732274455000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0031320324004497"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12]]},"references-count":56,"alternative-id":["S0031320324004497"],"URL":"https:\/\/doi.org\/10.1016\/j.patcog.2024.110698","relation":{},"ISSN":["0031-3203"],"issn-type":[{"value":"0031-3203","type":"print"}],"subject":[],"published":{"date-parts":[[2024,12]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"FMGNet: An efficient feature-multiplex group network for real-time vision task","name":"articletitle","label":"Article Title"},{"value":"Pattern Recognition","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.patcog.2024.110698","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2024 Elsevier Ltd. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"110698"}}