{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,13]],"date-time":"2026-03-13T16:24:43Z","timestamp":1773419083294,"version":"3.50.1"},"reference-count":60,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2025,4,2]],"date-time":"2025-04-02T00:00:00Z","timestamp":1743552000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,4,2]],"date-time":"2025-04-02T00:00:00Z","timestamp":1743552000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Pattern Anal Applic"],"published-print":{"date-parts":[[2025,6]]},"DOI":"10.1007\/s10044-025-01456-3","type":"journal-article","created":{"date-parts":[[2025,4,4]],"date-time":"2025-04-04T13:16:32Z","timestamp":1743772592000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Strengthen contrastive semantic consistency for fine-grained image classification"],"prefix":"10.1007","volume":"28","author":[{"given":"Yupeng","family":"Wang","sequence":"first","affiliation":[]},{"given":"Yongli","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Qiaolin","family":"Ye","sequence":"additional","affiliation":[]},{"given":"Wenxi","family":"Lang","sequence":"additional","affiliation":[]},{"given":"Can","family":"Xu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,4,2]]},"reference":[{"key":"1456_CR1","doi-asserted-by":"publisher","first-page":"2","DOI":"10.1007\/s10044-023-01207-2","volume":"27","author":"S Qu","year":"2024","unstructured":"Qu S, Wang Z, Wu J, Feng Y (2024) FBRNet: a feature fusion and border refinement network for real-time semantic segmentation. Pattern Anal Appl 27:2. https:\/\/doi.org\/10.1007\/s10044-023-01207-2","journal-title":"Pattern Anal Appl"},{"key":"1456_CR2","doi-asserted-by":"publisher","first-page":"655","DOI":"10.1007\/s10044-022-01113-z","volume":"26","author":"G Chen","year":"2023","unstructured":"Chen G, Chen Q, Long S et al (2023) Quantum convolutional neural network for image classification. Pattern Anal Appl 26:655\u2013667. https:\/\/doi.org\/10.1007\/s10044-022-01113-z","journal-title":"Pattern Anal Appl"},{"key":"1456_CR3","doi-asserted-by":"publisher","first-page":"1481","DOI":"10.1007\/s10044-023-01176-6","volume":"26","author":"Z Hu","year":"2023","unstructured":"Hu Z, Huang Y, Zheng H et al (2023) Graph-based fine-grained model selection for multi-source domain. Pattern Anal Appl 26:1481\u20131492. https:\/\/doi.org\/10.1007\/s10044-023-01176-6","journal-title":"Pattern Anal Appl"},{"key":"1456_CR4","unstructured":"Wah C, Branson S, Welinder P et al (2011) The caltech-ucsd birds-200-2011 dataset"},{"key":"1456_CR5","unstructured":"Khosla A, Jayadevaprakash N, Yao B, Li F-F (2011) Novel dataset for fine-grained image categorization: Stanford dogs. In: Proc. CVPR workshop on fine-grained visual categorization (FGVC). Citeseer"},{"key":"1456_CR6","doi-asserted-by":"crossref","unstructured":"Krause J, Stark M, Deng J, Fei-Fei L (2013) 3d object representations for fine-grained categorization. In: Proceedings of the IEEE international conference on computer vision workshops. pp 554\u2013561","DOI":"10.1109\/ICCVW.2013.77"},{"key":"1456_CR7","unstructured":"Maji S, Rahtu E, Kannala J et al (2013) Fine-grained visual classification of aircraft. arXiv preprint arXiv:13065151"},{"key":"1456_CR8","unstructured":"Ren S, He K, Girshick R, Sun J (2015) Faster r-cnn: towards real-time object detection with region proposal networks. Adv Neural Inf Process Syst 28"},{"key":"1456_CR9","doi-asserted-by":"crossref","unstructured":"Ge W, Lin X, Yu Y (2019) Weakly supervised complementary parts models for fine-grained image classification from the bottom up. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp 3034\u20133043","DOI":"10.1109\/CVPR.2019.00315"},{"key":"1456_CR10","doi-asserted-by":"crossref","unstructured":"Rao Y, Chen G, Lu J, Zhou J (2021) Counterfactual attention learning for fine-grained visual categorization and re-identification. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. pp 1025\u20131034","DOI":"10.1109\/ICCV48922.2021.00106"},{"key":"1456_CR11","doi-asserted-by":"crossref","unstructured":"Sun M, Yuan Y, Zhou F, Ding E (2018) Multi-attention multi-class constraint for fine-grained image recognition. In: Proceedings of the european conference on computer vision (ECCV). pp 805\u2013821","DOI":"10.1007\/978-3-030-01270-0_49"},{"key":"1456_CR12","doi-asserted-by":"crossref","unstructured":"Zhuang P, Wang Y, Qiao Y (2020) Learning attentive pairwise interaction for fine-grained classification. In: Proceedings of the AAAI conference on artificial intelligence. pp 13130\u201313137","DOI":"10.1609\/aaai.v34i07.7016"},{"key":"1456_CR13","doi-asserted-by":"publisher","first-page":"103414","DOI":"10.1016\/j.jvcir.2021.103414","volume":"82","author":"C Wang","year":"2022","unstructured":"Wang C, Qian Y, Gong W et al (2022) Cross-layer progressive attention bilinear fusion method for fine-grained visual classification. J Vis Commun Image Represent 82:103414","journal-title":"J Vis Commun Image Represent"},{"key":"1456_CR14","doi-asserted-by":"publisher","first-page":"102468","DOI":"10.1016\/j.displa.2023.102468","volume":"79","author":"Q Zhu","year":"2023","unstructured":"Zhu Q, Kuang W, Li Z (2023) A collaborative gated attention network for fine-grained visual classification. Displays 79:102468. https:\/\/doi.org\/10.1016\/j.displa.2023.102468","journal-title":"Displays"},{"key":"1456_CR15","doi-asserted-by":"publisher","unstructured":"Hu T, Qi H, Huang Q, Lu Y (2019) See better before looking closer: weakly supervised data augmentation network for fine-grained visual classification. ArXiv Preprint arXiv 190109891. https:\/\/doi.org\/10.48550\/arXiv.1901.09891","DOI":"10.48550\/arXiv.1901.09891"},{"key":"1456_CR16","doi-asserted-by":"crossref","unstructured":"He J, Chen J-N, Liu S et al (2022) Transfg: A transformer architecture for fine-grained recognition. In: Proceedings of the AAAI Conference on Artificial Intelligence. pp 852\u2013860","DOI":"10.1609\/aaai.v36i1.19967"},{"key":"1456_CR17","doi-asserted-by":"crossref","unstructured":"Wang J, Yu X, Gao Y (2021) Feature fusion vision transformer for fine-grained visual categorization. arXiv preprint arXiv:210702341","DOI":"10.5244\/C.35.192"},{"key":"1456_CR18","doi-asserted-by":"crossref","unstructured":"Fu J, Zheng H, Mei T (2017) Look closer to see better: Recurrent attention convolutional neural network for fine-grained image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 4438\u20134446","DOI":"10.1109\/CVPR.2017.476"},{"key":"1456_CR19","doi-asserted-by":"crossref","unstructured":"Zhang C, Yao Y, Liu H et al (2020) Web-supervised network with softly update-drop training for fine-grained visual classification. In: Proceedings of the AAAI conference on artificial intelligence. pp 12781\u201312788","DOI":"10.1609\/aaai.v34i07.6973"},{"key":"1456_CR20","doi-asserted-by":"crossref","unstructured":"Zhang N, Donahue J, Girshick R, Darrell T (2014) Part-based R-CNNs for fine-grained category detection. In: Computer Vision\u2013ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6\u201312, 2014, Proceedings, Part I 13. Springer, pp 834\u2013849","DOI":"10.1007\/978-3-319-10590-1_54"},{"key":"1456_CR21","doi-asserted-by":"publisher","first-page":"10","DOI":"10.1109\/TIP.2017.2751960","volume":"27","author":"H Yao","year":"2017","unstructured":"Yao H, Zhang S, Yan C et al (2017) AutoBD: automated bi-level description for scalable fine-grained visual categorization. IEEE Trans Image Process 27:10\u201323","journal-title":"IEEE Trans Image Process"},{"key":"1456_CR22","doi-asserted-by":"crossref","unstructured":"Lam M, Mahasseni B, Todorovic S (2017) Fine-grained recognition as hsnet search for informative image parts. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 2520\u20132529","DOI":"10.1109\/CVPR.2017.688"},{"key":"1456_CR23","doi-asserted-by":"crossref","unstructured":"Li M, Lei L, Sun H et al (2022) Fine-grained visual classification via multilayer bilinear pooling with object localization. Vis Comput 1\u201310","DOI":"10.1007\/s00371-020-02052-8"},{"key":"1456_CR24","doi-asserted-by":"crossref","unstructured":"Wang Z, Wang S, Zhang P et al (2019) Weakly supervised fine-grained image classification via correlation-guided discriminative learning. In: Proceedings of the 27th ACM International Conference on Multimedia. pp 1851\u20131860","DOI":"10.1145\/3343031.3350976"},{"key":"1456_CR25","doi-asserted-by":"crossref","unstructured":"Li H, Zhang X, Tian Q, Xiong H Attribute mix: Semantic data augmentation for fine grained recognition. In: 2020 IEEE International Conference on Visual Communications and, Processing I (2020) (VCIP). IEEE, pp 243\u2013246","DOI":"10.1109\/VCIP49819.2020.9301763"},{"key":"1456_CR26","unstructured":"Yang S, Liu S, Yang C, Wang C (2021) Re-rank coarse classification with local region enhanced features for fine-grained image recognition. arXiv preprint arXiv:210209875"},{"key":"1456_CR27","doi-asserted-by":"crossref","unstructured":"Behera A, Wharton Z, Hewage PRPG, Bera A (2021) Context-aware attentional pooling (cap) for fine-grained visual classification. In: Proceedings of the AAAI conference on artificial intelligence. pp 929\u2013937","DOI":"10.1609\/aaai.v35i2.16176"},{"key":"1456_CR28","unstructured":"Behera A, Wharton Z, Liu Y et al (2020) Regional attention network (ran) for head pose and fine-grained gesture recognition. IEEE Trans Affect Comput"},{"key":"1456_CR29","doi-asserted-by":"publisher","first-page":"1785","DOI":"10.1109\/TMM.2019.2954747","volume":"22","author":"C Liu","year":"2019","unstructured":"Liu C, Xie H, Zha Z et al (2019) Bidirectional attention-recognition model for fine-grained object classification. IEEE Trans Multimedia 22:1785\u20131795","journal-title":"IEEE Trans Multimedia"},{"key":"1456_CR30","doi-asserted-by":"publisher","first-page":"502","DOI":"10.1109\/TMM.2019.2928494","volume":"22","author":"P Rodriguez","year":"2019","unstructured":"Rodriguez P, Velazquez D, Cucurull G et al (2019) Pay attention to the activations: A modular attention mechanism for fine-grained image recognition. IEEE Trans Multimedia 22:502\u2013514","journal-title":"IEEE Trans Multimedia"},{"key":"1456_CR31","doi-asserted-by":"publisher","first-page":"104033","DOI":"10.1016\/j.dsp.2023.104033","volume":"137","author":"X Zhang","year":"2023","unstructured":"Zhang X, Zhang K, Zhao W et al (2023) A benchmark dataset and approach for fine-grained visual categorization in complex scenes. Digit Signal Process 137:104033. https:\/\/doi.org\/10.1016\/j.dsp.2023.104033","journal-title":"Digit Signal Process"},{"key":"1456_CR32","doi-asserted-by":"crossref","unstructured":"Zheng H, Fu J, Zha Z-J, Luo J (2019) Looking for the devil in the details: Learning trilinear attention sampling network for fine-grained image recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp 5012\u20135021","DOI":"10.1109\/CVPR.2019.00515"},{"key":"1456_CR33","doi-asserted-by":"publisher","first-page":"2826","DOI":"10.1109\/TIP.2021.3055617","volume":"30","author":"Y Ding","year":"2021","unstructured":"Ding Y, Ma Z, Wen S et al (2021) AP-CNN: weakly supervised attention pyramid convolutional neural network for fine-grained visual classification. IEEE Trans Image Process 30:2826\u20132836","journal-title":"IEEE Trans Image Process"},{"key":"1456_CR34","first-page":"2","volume":"9","author":"A Jaiswal","year":"2020","unstructured":"Jaiswal A, Babu AR, Zadeh MZ et al (2020) A survey on contrastive self-supervised learning. Technol (Basel) 9:2","journal-title":"Technol (Basel)"},{"key":"1456_CR35","unstructured":"Oord A, van den, Li Y, Vinyals O (2018) Representation learning with contrastive predictive coding. arXiv preprint arXiv:180703748"},{"key":"1456_CR36","doi-asserted-by":"crossref","unstructured":"Wu Z, Xiong Y, Yu SX, Lin D (2018) Unsupervised feature learning via non-parametric instance discrimination. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 3733\u20133742","DOI":"10.1109\/CVPR.2018.00393"},{"key":"1456_CR37","unstructured":"Chen X, Fan H, Girshick R, He K (2020) Improved baselines with momentum contrastive learning. arXiv preprint arXiv:200304297"},{"key":"1456_CR38","doi-asserted-by":"crossref","unstructured":"He K, Fan H, Wu Y et al (2020) Momentum contrast for unsupervised visual representation learning. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. pp 9729\u20139738","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"1456_CR39","first-page":"18661","volume":"33","author":"P Khosla","year":"2020","unstructured":"Khosla P, Teterwak P, Wang C et al (2020) Supervised contrastive learning. Adv Neural Inf Process Syst 33:18661\u201318673","journal-title":"Adv Neural Inf Process Syst"},{"key":"1456_CR40","doi-asserted-by":"crossref","unstructured":"Wang X, Zhang R, Shen C et al (2021) Dense contrastive learning for self-supervised visual pre-training. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp 3024\u20133033","DOI":"10.1109\/CVPR46437.2021.00304"},{"key":"1456_CR41","unstructured":"Li J, Zhou P, Xiong C, Hoi SCH (2020) Prototypical contrastive learning of unsupervised representations. arXiv preprint arXiv:200504966"},{"key":"1456_CR42","doi-asserted-by":"crossref","unstructured":"Zhou T, Zhang M, Zhao F, Li J (2022) Regional semantic contrast and aggregation for weakly supervised semantic segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp 4299\u20134309","DOI":"10.1109\/CVPR52688.2022.00426"},{"key":"1456_CR43","doi-asserted-by":"publisher","first-page":"5398","DOI":"10.1109\/TPAMI.2024.3367952","volume":"46","author":"T Zhou","year":"2024","unstructured":"Zhou T, Wang W (2024) Cross-Image pixel contrasting for semantic segmentation. IEEE Trans Pattern Anal Mach Intell 46:5398\u20135412. https:\/\/doi.org\/10.1109\/TPAMI.2024.3367952","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"1456_CR44","doi-asserted-by":"crossref","unstructured":"Fang Z, Jiang X, Tang H, Li Z (2024) Learning contrastive self-distillation for ultra-fine-grained visual categorization targeting limited samples. IEEE transactions on circuits and systems for video technology","DOI":"10.1109\/TCSVT.2024.3370731"},{"key":"1456_CR45","doi-asserted-by":"crossref","unstructured":"Van Horn G, Branson S, Farrell R et al (2015) Building a bird recognition app and large scale dataset with citizen scientists: The fine print in fine-grained dataset collection. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 595\u2013604","DOI":"10.1109\/CVPR.2015.7298658"},{"key":"1456_CR46","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"1456_CR47","doi-asserted-by":"crossref","unstructured":"Szegedy C, Vanhoucke V, Ioffe S et al (2016) Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 2818\u20132826","DOI":"10.1109\/CVPR.2016.308"},{"key":"1456_CR48","unstructured":"Alexey D (2020) An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv: 201011929"},{"key":"1456_CR49","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2024.3406443","author":"H Yao","year":"2024","unstructured":"Yao H, Miao Q, Zhao P et al (2024) Exploration of class center for fine-grained visual classification. IEEE Trans Circuits Syst Video Technol. https:\/\/doi.org\/10.1109\/TCSVT.2024.3406443","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"1456_CR50","doi-asserted-by":"crossref","unstructured":"Do T, Tran H, Tjiputra E Fine-grained visual classification using self assessment classifier. In: 2024 IEEE Conference on Artificial, Intelligence et al (2024) (CAI). IEEE, pp 597\u2013602","DOI":"10.1109\/CAI59869.2024.00117"},{"key":"1456_CR51","doi-asserted-by":"crossref","unstructured":"Huang S, Wang X, Tao D (2021) Stochastic partial swap: Enhanced model generalization and interpretability for fine-grained recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision. pp 620\u2013629","DOI":"10.1109\/ICCV48922.2021.00066"},{"key":"1456_CR52","doi-asserted-by":"publisher","first-page":"479","DOI":"10.1049\/cvi2.12105","volume":"16","author":"W Li","year":"2022","unstructured":"Li W, Li S, Yin L et al (2022) A novel visual classification framework on panoramic attention mechanism network. IET Comput Vision 16:479\u2013488","journal-title":"IET Comput Vision"},{"key":"1456_CR53","doi-asserted-by":"crossref","unstructured":"Du R, Chang D, Bhunia AK et al (2020) Fine-grained visual classification via progressive multi-granularity training of jigsaw patches. In: European Conference on Computer Vision. Springer, pp 153\u2013168","DOI":"10.1007\/978-3-030-58565-5_10"},{"key":"1456_CR54","doi-asserted-by":"publisher","first-page":"306","DOI":"10.1016\/j.neunet.2023.01.050","volume":"161","author":"P Zhao","year":"2023","unstructured":"Zhao P, Li Y, Tang B et al (2023) Feature relocation network for fine-grained image classification. Neural Netw 161:306\u2013317. https:\/\/doi.org\/10.1016\/j.neunet.2023.01.050","journal-title":"Neural Netw"},{"key":"1456_CR55","doi-asserted-by":"crossref","unstructured":"Song J, Yang R (2021) Feature boosting, suppression, and diversification for fine-grained visual classification. In: 2021 International Joint Conference on Neural Networks (IJCNN). IEEE, pp 1\u20138","DOI":"10.1109\/IJCNN52387.2021.9534004"},{"key":"1456_CR56","doi-asserted-by":"publisher","first-page":"137","DOI":"10.1016\/j.neucom.2022.04.037","volume":"492","author":"X Liu","year":"2022","unstructured":"Liu X, Wang L, Han X (2022) Transformer with peak suppression and knowledge guidance for fine-grained image recognition. Neurocomputing 492:137\u2013149","journal-title":"Neurocomputing"},{"key":"1456_CR57","doi-asserted-by":"crossref","unstructured":"Zhu H, Ke W, Li D et al (2022) Dual cross-attention learning for fine-grained visual categorization and object re-identification. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. pp 4692\u20134702","DOI":"10.1109\/CVPR52688.2022.00465"},{"key":"1456_CR58","doi-asserted-by":"publisher","first-page":"103837","DOI":"10.1016\/j.cviu.2023.103837","volume":"237","author":"S Ye","year":"2023","unstructured":"Ye S, Yu S, Hou W et al (2023) Coping with change: learning invariant and minimum sufficient representations for fine-grained visual categorization. Comput Vis Image Underst 237:103837","journal-title":"Comput Vis Image Underst"},{"key":"1456_CR59","doi-asserted-by":"publisher","first-page":"9521","DOI":"10.1109\/TPAMI.2021.3126668","volume":"44","author":"R Du","year":"2022","unstructured":"Du R, Xie J, Ma Z et al (2022) Progressive learning of Category-Consistent Multi-Granularity features for Fine-Grained visual classification. IEEE Trans Pattern Anal Mach Intell 44:9521\u20139535. https:\/\/doi.org\/10.1109\/TPAMI.2021.3126668","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"1456_CR60","doi-asserted-by":"publisher","first-page":"109305","DOI":"10.1016\/j.patcog.2023.109305","volume":"137","author":"X Ke","year":"2023","unstructured":"Ke X, Cai Y, Chen B et al (2023) Granularity-aware distillation and structure modeling region proposal network for fine-grained image classification. Pattern Recognit 137:109305","journal-title":"Pattern Recognit"}],"container-title":["Pattern Analysis and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10044-025-01456-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10044-025-01456-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10044-025-01456-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,7,2]],"date-time":"2025-07-02T16:41:32Z","timestamp":1751474492000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10044-025-01456-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,4,2]]},"references-count":60,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2025,6]]}},"alternative-id":["1456"],"URL":"https:\/\/doi.org\/10.1007\/s10044-025-01456-3","relation":{},"ISSN":["1433-7541","1433-755X"],"issn-type":[{"value":"1433-7541","type":"print"},{"value":"1433-755X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,4,2]]},"assertion":[{"value":"23 February 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 March 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 April 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"75"}}