{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,4]],"date-time":"2025-11-04T11:06:40Z","timestamp":1762254400787,"version":"3.37.3"},"reference-count":59,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2023,1,30]],"date-time":"2023-01-30T00:00:00Z","timestamp":1675036800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,30]],"date-time":"2023-01-30T00:00:00Z","timestamp":1675036800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U21A20487"],"award-info":[{"award-number":["U21A20487"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012156","name":"Shenzhen Technical Project","doi-asserted-by":"publisher","award":["JCYJ20180507182610734","KCXFZ20201221173411032"],"award-info":[{"award-number":["JCYJ20180507182610734","KCXFZ20201221173411032"]}],"id":[{"id":"10.13039\/501100012156","id-type":"DOI","asserted-by":"publisher"}]},{"name":"CAS Key Technology Talent Program"},{"DOI":"10.13039\/501100013289","name":"SIAT Innovation Program for Excellent Young Researchers","doi-asserted-by":"crossref","award":["E1G032"],"award-info":[{"award-number":["E1G032"]}],"id":[{"id":"10.13039\/501100013289","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Neural Process Lett"],"published-print":{"date-parts":[[2023,10]]},"DOI":"10.1007\/s11063-023-11150-2","type":"journal-article","created":{"date-parts":[[2023,1,30]],"date-time":"2023-01-30T13:05:23Z","timestamp":1675083923000},"page":"6595-6609","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Semantic-Aware Feature Aggregation for Few-Shot Image Classification"],"prefix":"10.1007","volume":"55","author":[{"given":"Fusheng","family":"Hao","sequence":"first","affiliation":[]},{"given":"Fuxiang","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Fengxiang","family":"He","sequence":"additional","affiliation":[]},{"given":"Qieshi","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Chengqun","family":"Song","sequence":"additional","affiliation":[]},{"given":"Jun","family":"Cheng","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,1,30]]},"reference":[{"key":"11150_CR1","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: IEEE conference on computer vision and pattern recognition (CVPR), pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"11150_CR2","doi-asserted-by":"crossref","unstructured":"Hu J, Shen L, Sun G (2018) Squeeze-and-excitation networks. In: IEEE conference on computer vision and pattern recognition (CVPR), pp 7132\u20137141","DOI":"10.1109\/CVPR.2018.00745"},{"issue":"2","key":"11150_CR3","doi-asserted-by":"publisher","first-page":"386","DOI":"10.1109\/TPAMI.2018.2844175","volume":"42","author":"K He","year":"2020","unstructured":"He K, Gkioxari G, Doll\u00e1r P, Girshick R (2020) Mask R-CNN. IEEE Trans Pattern Anal Mach Intell (IEEE TPAMI) 42(2):386\u2013397","journal-title":"IEEE Trans Pattern Anal Mach Intell (IEEE TPAMI)"},{"key":"11150_CR4","doi-asserted-by":"publisher","first-page":"2789","DOI":"10.1007\/s11063-020-10228-5","volume":"51","author":"H Zhao","year":"2020","unstructured":"Zhao H, Li Z, Fang L, Zhang T (2020) A balanced feature fusion ssd for object detection. Neural Process Lett (NPL) 51:2789\u20132806","journal-title":"Neural Process Lett (NPL)"},{"key":"11150_CR5","doi-asserted-by":"publisher","DOI":"10.1007\/s11063-022-10894-7","author":"Y He","year":"2022","unstructured":"He Y, Zang C, Zeng P, Dong Q, Liu D, Liu Y (2022) Convolutional shrinkage neural networks based model-agnostic meta-learning for few-shot learning. Neural Process Lett (NPL). https:\/\/doi.org\/10.1007\/s11063-022-10894-7","journal-title":"Neural Process Lett (NPL)"},{"key":"11150_CR6","doi-asserted-by":"publisher","DOI":"10.1007\/s11063-022-10918-2","author":"J Zhao","year":"2022","unstructured":"Zhao J, Tang T, Yu Y, Wang J, Yang T, Chen M, Wu J (2022) Adaptive meta transfer learning with efficient self-attention for few-shot bearing fault diagnosis. Neural Process Lett (NPL). https:\/\/doi.org\/10.1007\/s11063-022-10918-2","journal-title":"Neural Process Lett (NPL)"},{"key":"11150_CR7","doi-asserted-by":"crossref","unstructured":"Nie L, Li X, Gong T, Zhan D (2022) Few shot learning-based fast adaptation for human activity recognition. Pattern Recogn Lett (PRL) 159:100\u2013107","DOI":"10.1016\/j.patrec.2022.04.014"},{"key":"11150_CR8","unstructured":"Xu W, Xu Y, Wang H, Tu Z (2021) Constellation nets for few-shot learning. In: International conference on learning representations (ICLR)"},{"key":"11150_CR9","doi-asserted-by":"crossref","unstructured":"Nguyen VN, L\u00f8kse S, Wickstr\u00f8m K, Kampffmeyer M, Roverso D, Jenssen R (2020) SEN: a novel feature normalization dissimilarity measure for prototypical few-shot learning networks. In: European conference on computer vision (ECCV), pp 118\u2013134","DOI":"10.1007\/978-3-030-58592-1_8"},{"key":"11150_CR10","unstructured":"Yang S, Liu L, Xu M (2021) Free lunch for few-shot learning: distribution calibration. In: International conference on learning representations (ICLR)"},{"key":"11150_CR11","doi-asserted-by":"crossref","unstructured":"Lee K, Maji S, Ravichandran A, Soatto S (2019) Meta-learning with differentiable convex optimization. In: IEEE conference on computer vision and pattern recognition (CVPR), pp 10657\u201310665","DOI":"10.1109\/CVPR.2019.01091"},{"key":"11150_CR12","doi-asserted-by":"crossref","unstructured":"Li A, Huang W, Lan X, Feng J, Li Z, Wang L (2020) Boosting few-shot learning with adaptive margin loss. In: IEEE conference on computer vision and pattern recognition (CVPR), pp 12576\u201312584","DOI":"10.1109\/CVPR42600.2020.01259"},{"key":"11150_CR13","doi-asserted-by":"crossref","unstructured":"Li W, Xu J, Huo J, Wang L, Yang G, Luo J (2019) Distribution consistency based covariance metric networks for few-shot learning. In: AAAI conference on artificial intelligence (AAAI), pp 8642\u20138649","DOI":"10.1609\/aaai.v33i01.33018642"},{"key":"11150_CR14","doi-asserted-by":"crossref","unstructured":"Li W, Wang L, Huo J, Shi Y, Gao Y, Luo J (2020) Asymmetric distribution measure for few-shot learning. Proceedings of the twenty-ninth international joint conference on artificial intelligence (IJCAI), pp 2957\u20132963","DOI":"10.24963\/ijcai.2020\/409"},{"key":"11150_CR15","doi-asserted-by":"crossref","unstructured":"Sung F, Yang Y, Zhang L, Xiang T, Torr PHS, Hospedales TM (2018) Learning to compare: relation network for few-shot learning. In: IEEE conference on computer vision and pattern recognition (CVPR), pp 1199\u20131208","DOI":"10.1109\/CVPR.2018.00131"},{"key":"11150_CR16","doi-asserted-by":"crossref","unstructured":"Hao F, He F, Cheng J, Wang L, Cao J, Tao D (2019) Collect and select: semantic alignment metric learning for few-shot learning. In: IEEE international conference on computer vision (ICCV), pp 8460\u20138469","DOI":"10.1109\/ICCV.2019.00855"},{"key":"11150_CR17","doi-asserted-by":"crossref","unstructured":"Huang S, Zhang M, Kang Y, Wang D (2021) Attributes-guided and pure-visual attention alignment for few-shot recognition. In: AAAI conference on artificial intelligence (AAAI)","DOI":"10.1609\/aaai.v35i9.16957"},{"key":"11150_CR18","doi-asserted-by":"crossref","unstructured":"Zhang C, Cai Y, Lin G, Shen C (2020) DeepEMD: few-shot image classification with differentiable earth mover\u2019s distance and structured classifiers. In: IEEE conference on computer vision and pattern recognition (CVPR), pp 12203\u201312213","DOI":"10.1109\/CVPR42600.2020.01222"},{"key":"11150_CR19","unstructured":"Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser L, Polosukhin I (2017) Attention is all you need. In: Advances in neural information processing systems (NeurIPS)"},{"key":"11150_CR20","doi-asserted-by":"crossref","unstructured":"Huynh D, Elhamifar E (2020) A shared multi-attention framework for multi-label zero-shot learning. In: IEEE conference on computer vision and pattern recognition (CVPR), pp 8776\u20138786","DOI":"10.1109\/CVPR42600.2020.00880"},{"key":"11150_CR21","doi-asserted-by":"crossref","unstructured":"Ye J, He J, Peng X, Wu W, Qiao Y (2020) Attention-driven dynamic graph convolutional network for multi-label image recognition. In: European conference on computer vision (ECCV), pp 649\u2013665","DOI":"10.1007\/978-3-030-58589-1_39"},{"issue":"3","key":"11150_CR22","doi-asserted-by":"publisher","first-page":"1371","DOI":"10.1109\/TPAMI.2020.3025814","volume":"44","author":"T Chen","year":"2022","unstructured":"Chen T, Lin L, Chen R, Hui X, Wu H (2022) Knowledge-guided multi-label few-shot learning for general image recognition. IEEE Trans Pattern Anal Mach Intell (IEEE TPAMI) 44(3):1371\u20131384","journal-title":"IEEE Trans Pattern Anal Mach Intell (IEEE TPAMI)"},{"key":"11150_CR23","unstructured":"Jake S, Kevin S, Richard Z (2017) Prototypical networks for few-shot learning. In: Advances in neural information processing systems (NeurIPS), pp 4077\u20134087"},{"key":"11150_CR24","unstructured":"Laenen S, Bertinetto L (2021) On episodes, prototypical networks, and few-shot learning. In: Advances in neural information processing systems (NeurIPS)"},{"key":"11150_CR25","doi-asserted-by":"crossref","unstructured":"Li W, Wang L, Xu J, Huo J, Yang G, Luo J (2019) Revisiting local descriptor based image-to-class measure for few-shot learning. In: IEEE conference on computer vision and pattern recognition (CVPR), pp 7253\u20137260","DOI":"10.1109\/CVPR.2019.00743"},{"key":"11150_CR26","doi-asserted-by":"crossref","unstructured":"Wertheimer D, Tang L, Hariharan B (2021) Few-shot classification with feature map reconstruction networks. In: IEEE conference on computer vision and pattern recognition (CVPR), pp 8012\u20138021","DOI":"10.1109\/CVPR46437.2021.00792"},{"key":"11150_CR27","doi-asserted-by":"publisher","first-page":"1520","DOI":"10.1109\/TIP.2022.3143005","volume":"31","author":"Z Ji","year":"2022","unstructured":"Ji Z, Hou Z, Liu X, Pang Y, Han J (2022) Information symmetry matters: a modal-alternating propagation network for few-shot learning. IEEE Trans Image Process (IEEE TIP) 31:1520\u20131531","journal-title":"IEEE Trans Image Process (IEEE TIP)"},{"issue":"9","key":"11150_CR28","doi-asserted-by":"publisher","first-page":"5443","DOI":"10.1109\/TSMC.2021.3125343","volume":"52","author":"Z Ji","year":"2022","unstructured":"Ji Z, An P, Liu X, Pang Y, Shao L, Zhang Z (2022) Task-oriented high-order context graph networks for few-shot human-object interaction recognition. IEEE Trans Syst Man Cybern Syst (IEEE TSMC) 52(9):5443\u20135455","journal-title":"IEEE Trans Syst Man Cybern Syst (IEEE TSMC)"},{"issue":"8","key":"11150_CR29","doi-asserted-by":"publisher","first-page":"7852","DOI":"10.1109\/TCYB.2021.3049537","volume":"52","author":"X Liu","year":"2022","unstructured":"Liu X, Ji Z, Pang Y, Han J, Li X (2022) Dgig-net: dynamic graph-in-graph networks for few-shot human-object interaction. IEEE Trans Cybern (IEEE TCYB) 52(8):7852\u20137864","journal-title":"IEEE Trans Cybern (IEEE TCYB)"},{"key":"11150_CR30","unstructured":"Finn C, Abbeel P, Levine S (2017) Model-agnostic meta-learning for fast adaptation of deep networks. In: International conference on machine learning (ICML), pp 1126\u20131135"},{"key":"11150_CR31","unstructured":"Oriol V, Charles B, Tim L, Kavukcuoglu K, Daan W (2016) Matching networks for one shot learning. In: Advances in neural information processing systems (NeurIPS), pp 3630\u20133638"},{"key":"11150_CR32","unstructured":"Li Z, Zhou F, Chen F, Li H (2017) Meta-sgd: learning to learn quickly for few-shot learning. Preprint arXiv:1707.09835"},{"key":"11150_CR33","unstructured":"Dhillon GS, Chaudhari P, Ravichandran A, Soatto S (2020) A baseline for few-shot image classification. In: International conference on learning representations (ICLR)"},{"key":"11150_CR34","doi-asserted-by":"crossref","unstructured":"Tian Y, Wang Y, Krishnan D, Tenenbaum JB, Isola P (2020) Rethinking few-shot image classification: a good embedding is all you need? In: European conference on computer vision (ECCV), pp 266\u2013282","DOI":"10.1007\/978-3-030-58568-6_16"},{"key":"11150_CR35","unstructured":"Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Georg\u00a0Heigold SG, Uszkoreit J, Houlsby N (2021) An image is worth 16x16 words: transformers for image recognition at scale. In: International conference on learning representations (ICLR)"},{"key":"11150_CR36","unstructured":"Hou R, Chang H, Ma B, Shan S, Chen X (2019) Cross attention network for few-shot classification. In: Advances in neural information processing systems (NeurIPS)"},{"key":"11150_CR37","doi-asserted-by":"crossref","unstructured":"Ye H-J, Hu H, Zhan D-C, Sha F (2020) Few-shot learning via embedding adaptation with set-to-set functions. In: IEEE conference on computer vision and pattern recognition (CVPR), pp 8808\u20138817","DOI":"10.1109\/CVPR42600.2020.00883"},{"key":"11150_CR38","unstructured":"Doersch C, Gupta A, Zisserman A (2020) Crosstransformers: spatially-aware few-shot transfer. In: Advances in neural information processing systems (NeurIPS)"},{"key":"11150_CR39","unstructured":"Zagoruyko S, Komodakis N (2017) Paying more attention to attention: improving the performance of convolutional neural networks via attention transfer. In: International conference on learning representations (ICLR)"},{"key":"11150_CR40","unstructured":"Ren M, Triantafillou E, Ravi S, Snell J, Swersky K, Tenenbaum JB, Larochelle H, Zemel RS (2018) Meta-learning for semi-supervised few-shot classification. In: International conference on learning representations (ICLR)"},{"key":"11150_CR41","unstructured":"Ravi S, Larochelle H (2017) Optimization as a model for few-shot learning. In: International conference on learning representations (ICLR)"},{"key":"11150_CR42","unstructured":"Oreshkin B, Lopez PR, Lacoste A (2018) TADAM: task dependent adaptive metric for improved few-shot learning. In: Advances in neural information processing systems (NeurIPS), pp 719\u2013729"},{"key":"11150_CR43","unstructured":"Rusu AA, Rao D, Sygnowski J, Vinyals O, Pascanu R, Osindero S, Hadsell R (2019) Meta-learning with latent embedding optimization. In: International conference on learning representations (ICLR)"},{"key":"11150_CR44","doi-asserted-by":"crossref","unstructured":"Gidaris S, Komodakis N (2018) Dynamic few-shot visual learning without forgetting. In: IEEE Conference on computer vision and pattern recognition (CVPR), pp 4367\u20134375","DOI":"10.1109\/CVPR.2018.00459"},{"key":"11150_CR45","unstructured":"Kingma DP, Ba J (2015) Adam: a method for stochastic optimization. In: International conference on learning representations (ICLR)"},{"key":"11150_CR46","doi-asserted-by":"crossref","unstructured":"Afrasiyabi A, Jean-Fran\u00e7ois Gagn\u00e9 C (2021) Mixture-based feature space learning for few-shot image classification. In: IEEE international conference on computer vision (ICCV), pp 9041\u20139051","DOI":"10.1109\/ICCV48922.2021.00891"},{"key":"11150_CR47","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2021.3132912","author":"F Hao","year":"2021","unstructured":"Hao F, He F, Cheng J, Tao D (2021) Global-local interplay in semantic alignment for few-shot learning. IEEE Trans Circuits Syst Video Technol (IEEE TCSVT). https:\/\/doi.org\/10.1109\/TCSVT.2021.3132912","journal-title":"IEEE Trans Circuits Syst Video Technol (IEEE TCSVT)"},{"key":"11150_CR48","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2021.3123813","author":"J Cheng","year":"2021","unstructured":"Cheng J, Hao F, He F, Liu L, Zhang Q (2021) Mixer-based semantic spread for few-shot learning. IEEE Trans Multimed (IEEE TMM). https:\/\/doi.org\/10.1109\/TMM.2021.3123813","journal-title":"IEEE Trans Multimed (IEEE TMM)"},{"key":"11150_CR49","doi-asserted-by":"crossref","unstructured":"Ma R, Fang P, Drummond T, Harandi M (2022) Adaptive poincar\u00e9 point to set distance for few-shot classification. In: AAAI conference on artificial intelligence (AAAI)","DOI":"10.1609\/aaai.v36i2.20087"},{"key":"11150_CR50","doi-asserted-by":"publisher","unstructured":"Huang H, Wu Z, Li W, Huo J, Gao Y (2021) Local descriptor-based multi-prototype network for few-shot learning. Pattern Recogn (PR). https:\/\/doi.org\/10.1016\/j.patcog.2021.107935","DOI":"10.1016\/j.patcog.2021.107935"},{"key":"11150_CR51","doi-asserted-by":"crossref","unstructured":"Gidaris S, Komodakis N (2019) Generating classification weights with gnn denoising autoencoders for few-shot learning. In: IEEE conference on computer vision and pattern recognition (CVPR), pp 21\u201330","DOI":"10.1109\/CVPR.2019.00011"},{"key":"11150_CR52","doi-asserted-by":"crossref","unstructured":"Chen K, Lee C-G (2022) Meta-free few-shot learning via representation learning with weight averaging. In: International joint conference on neural networks (IJCNN), pp 1\u20138","DOI":"10.1109\/IJCNN55064.2022.9892722"},{"key":"11150_CR53","doi-asserted-by":"publisher","unstructured":"Ye H-J, Ming L, Zhan D-C, Chao W-L (2022) Few-shot learning with a strong teacher. IEEE Trans Pattern Anal Mach Intell (IEEE TPAMI). https:\/\/doi.org\/10.1109\/TPAMI.2022.3160362","DOI":"10.1109\/TPAMI.2022.3160362"},{"key":"11150_CR54","doi-asserted-by":"crossref","unstructured":"Lu Y, Wen L, Liu J, Liu Y, Tian X (2022) Self-supervision can be a good few-shot learner. In: European conference on computer vision (ECCV), pp 740\u2013758","DOI":"10.1007\/978-3-031-19800-7_43"},{"key":"11150_CR55","doi-asserted-by":"crossref","unstructured":"Shen Z, Liu Z, Qin J, Savvides M, Cheng K-T (2021) Partial is better than all: revisiting fine-tuning strategy for few-shot learning. In: AAAI conference on artificial intelligence (AAAI)","DOI":"10.1609\/aaai.v35i11.17155"},{"key":"11150_CR56","doi-asserted-by":"publisher","first-page":"81","DOI":"10.1016\/j.patrec.2020.07.015","volume":"140","author":"Z Ji","year":"2020","unstructured":"Ji Z, Chai X, Yu Y, Pang Y, Zhang Z (2020) Improved prototypical networks for few-shot learning. Pattern Recogn Lett (PRL) 140:81\u201387","journal-title":"Pattern Recogn Lett (PRL)"},{"key":"11150_CR57","doi-asserted-by":"crossref","unstructured":"Zhang H, Koniusz P, Jian S, Li H, Torr PHS (2021) Rethinking class relations: absolute-relative supervised and unsupervised few-shot learning. In: IEEE conference on computer vision and pattern recognition (CVPR), pp 9432\u20139441","DOI":"10.1109\/CVPR46437.2021.00931"},{"key":"11150_CR58","doi-asserted-by":"crossref","unstructured":"Chen Z, Ge J, Zhan H, Huang S, Wang D (2021) Pareto self-supervised training for few-shot learning. In: IEEE conference on computer vision and pattern recognition (CVPR), pp 13663\u201313672","DOI":"10.1109\/CVPR46437.2021.01345"},{"key":"11150_CR59","doi-asserted-by":"crossref","unstructured":"Qiao S, Liu C, Shen W, Yuille A (2018) Few-shot image recognition by predicting parameters from activations. In: IEEE conference on computer vision and pattern recognition (CVPR), pp 7229\u20137238","DOI":"10.1109\/CVPR.2018.00755"}],"container-title":["Neural Processing Letters"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11063-023-11150-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11063-023-11150-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11063-023-11150-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,9,29]],"date-time":"2023-09-29T16:25:08Z","timestamp":1696004708000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11063-023-11150-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,1,30]]},"references-count":59,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2023,10]]}},"alternative-id":["11150"],"URL":"https:\/\/doi.org\/10.1007\/s11063-023-11150-2","relation":{},"ISSN":["1370-4621","1573-773X"],"issn-type":[{"type":"print","value":"1370-4621"},{"type":"electronic","value":"1573-773X"}],"subject":[],"published":{"date-parts":[[2023,1,30]]},"assertion":[{"value":"10 January 2023","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 January 2023","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"This article does not contain any studies with human participants or animals performed by any of the authors.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical Approval"}}]}}