{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,28]],"date-time":"2026-03-28T12:51:09Z","timestamp":1774702269488,"version":"3.50.1"},"reference-count":65,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62376145"],"award-info":[{"award-number":["62376145"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100004480","name":"Shanxi Province Natural Science Foundation","doi-asserted-by":"publisher","award":["202303021211024"],"award-info":[{"award-number":["202303021211024"]}],"id":[{"id":"10.13039\/501100004480","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Knowledge-Based Systems"],"published-print":{"date-parts":[[2026,5]]},"DOI":"10.1016\/j.knosys.2026.115678","type":"journal-article","created":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T00:18:54Z","timestamp":1772583534000},"page":"115678","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["SPG-Mask: Fine-Grained visual classification with Self-Supervised foreground structural prior driven dynamic mask optimization"],"prefix":"10.1016","volume":"340","author":[{"ORCID":"https:\/\/orcid.org\/0009-0003-5849-5464","authenticated-orcid":false,"given":"Huiming","family":"He","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1732-8606","authenticated-orcid":false,"given":"Hongju","family":"Yang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6356-7926","authenticated-orcid":false,"given":"Yao","family":"Li","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1111-8461","authenticated-orcid":false,"given":"Fuyuan","family":"Cao","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"issue":"11","key":"10.1016\/j.knosys.2026.115678_bib0001","doi-asserted-by":"crossref","first-page":"6558","DOI":"10.1109\/TCSVT.2023.3263870","article-title":"Semantic-guided information alignment network for fine-grained image recognition","volume":"33","author":"Wang","year":"2023","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10.1016\/j.knosys.2026.115678_bib0002","doi-asserted-by":"crossref","DOI":"10.1109\/TCSVT.2024.3433415","article-title":"TMFF: Trustworthy multi-focus fusion framework for multi-label sewer defect classification in sewer inspection videos","author":"Hu","year":"2024","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"9","key":"10.1016\/j.knosys.2026.115678_bib0003","doi-asserted-by":"crossref","first-page":"4452","DOI":"10.1109\/TCSVT.2023.3277462","article-title":"Vision transformer with hybrid shifted windows for gastrointestinal endoscopy image classification","volume":"33","author":"Wang","year":"2023","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"9","key":"10.1016\/j.knosys.2026.115678_bib0004","doi-asserted-by":"crossref","first-page":"5009","DOI":"10.1109\/TCSVT.2023.3248791","article-title":"Dual transformer with multi-grained assembly for fine-grained visual classification","volume":"33","author":"Ji","year":"2023","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10.1016\/j.knosys.2026.115678_bib0005","doi-asserted-by":"crossref","first-page":"704","DOI":"10.1016\/j.patcog.2017.10.002","article-title":"Mask-CNN: localizing parts and selecting descriptors for fine-grained bird species categorization","volume":"76","author":"Wei","year":"2018","journal-title":"Pattern Recognit."},{"issue":"6","key":"10.1016\/j.knosys.2026.115678_bib0006","doi-asserted-by":"crossref","first-page":"1137","DOI":"10.1109\/TPAMI.2016.2577031","article-title":"Faster R-CNN: towards real-time object detection with region proposal networks","volume":"39","author":"Ren","year":"2016","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.knosys.2026.115678_bib0007","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"5012","article-title":"Looking for the devil in the details: learning trilinear attention sampling network for fine-grained image recognition","author":"Zheng","year":"2019"},{"key":"10.1016\/j.knosys.2026.115678_bib0008","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"10818","article-title":"Channel interaction networks for fine-grained image categorization","volume":"34","author":"Gao","year":"2020"},{"issue":"7","key":"10.1016\/j.knosys.2026.115678_bib0009","doi-asserted-by":"crossref","first-page":"3488","DOI":"10.1109\/TNNLS.2021.3112768","article-title":"Multiresolution discriminative mixup network for fine-grained visual categorization","volume":"34","author":"Xu","year":"2021","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"10.1016\/j.knosys.2026.115678_bib0010","series-title":"9Th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3\u20137, 2021","article-title":"An image is worth 16x16 words: transformers for image recognition at scale","author":"Dosovitskiy","year":"2021"},{"key":"10.1016\/j.knosys.2026.115678_bib0011","article-title":"Revisiting fine-Grained classification: a dual-Branch method for noise-Resilient and global-Local discriminative learning","author":"Yang","year":"2025","journal-title":"Expert Syst. Appl."},{"key":"10.1016\/j.knosys.2026.115678_bib0012","doi-asserted-by":"crossref","DOI":"10.1109\/TIP.2024.3390565","article-title":"Graph-represented distribution similarity index for full-reference image quality assessment","author":"Shen","year":"2024","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.knosys.2026.115678_bib0013","doi-asserted-by":"crossref","DOI":"10.1109\/TCSVT.2024.3406443","article-title":"Exploration of class center for fine-grained visual classification","author":"Yao","year":"2024","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10.1016\/j.knosys.2026.115678_bib0014","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"852","article-title":"Transfg: a transformer architecture for fine-grained recognition","volume":"36","author":"He","year":"2022"},{"key":"10.1016\/j.knosys.2026.115678_bib0015","series-title":"2022IEEE International Conference on Image Processing, ICIP 2022, Bordeaux, France, 16\u201319 October 2022","article-title":"Mask-Vit: an object mask embedding in vision transformer for fine-Grained visual classification","author":"Su","year":"2022"},{"key":"10.1016\/j.knosys.2026.115678_bib0016","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2023.109979","article-title":"A vision transformer for fine-grained classification by reducing noise and enhancing discriminative information","volume":"145","author":"Zhang","year":"2024","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.knosys.2026.115678_bib0017","article-title":"An attention-Locating algorithm for eliminating background effects in fine-grained visual classification","author":"Huang","year":"2025","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10.1016\/j.knosys.2026.115678_bib0018","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2023.123021","article-title":"SwinFG: a fine-grained recognition scheme based on swin transformer","volume":"244","author":"Ma","year":"2024","journal-title":"Expert Syst. Appl."},{"key":"10.1016\/j.knosys.2026.115678_bib0019","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2023.110042","article-title":"Two-stage fine-grained image classification model based on multi-granularity feature fusion","volume":"146","author":"Xu","year":"2024","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.knosys.2026.115678_bib0020","series-title":"IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, June 18\u201324, 2022","article-title":"Fine-Grained object classification via self-Supervised pose alignment","author":"Yang","year":"2022"},{"key":"10.1016\/j.knosys.2026.115678_bib0021","series-title":"2019 IEEE\/CVF International Conference on Computer Vision, ICCV 2019, Seoul, Korea (South), October 27 November 2, 2019","article-title":"Cross-X learning for fine-Grained visual categorization","author":"Luo","year":"2019"},{"key":"10.1016\/j.knosys.2026.115678_bib0022","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"13130","article-title":"Learning attentive pairwise interaction for fine-grained classification","volume":"34","author":"Zhuang","year":"2020"},{"key":"10.1016\/j.knosys.2026.115678_bib0023","article-title":"Mask-Aware hierarchical aggregation transformer for occluded person re-identification","author":"Zhang","year":"2025","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10.1016\/j.knosys.2026.115678_bib0024","doi-asserted-by":"crossref","DOI":"10.1016\/j.imavis.2024.104923","article-title":"R2-Trans: fine-grained visual categorization with redundancy reduction","volume":"143","author":"Ye","year":"2024","journal-title":"Image Vis. Comput."},{"key":"10.1016\/j.knosys.2026.115678_bib0025","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"4148","article-title":"Learning a discriminative filter bank within a CNN for fine-grained recognition","author":"Wang","year":"2018"},{"key":"10.1016\/j.knosys.2026.115678_bib0026","doi-asserted-by":"crossref","first-page":"748","DOI":"10.1109\/TIP.2021.3135477","article-title":"Cross-Part learning for fine-Grained image classification","volume":"31","author":"Liu","year":"2022","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.knosys.2026.115678_bib0027","series-title":"Proceedings of the IEEE International Conference on Computer Vision","first-page":"3218","article-title":"P-Cnn: pose-based cnn features for action recognition","author":"Ch\u00e9ron","year":"2015"},{"issue":"12","key":"10.1016\/j.knosys.2026.115678_bib0028","doi-asserted-by":"crossref","first-page":"9521","DOI":"10.1109\/TPAMI.2021.3126668","article-title":"Progressive learning of category-Consistent multi-Granularity features for fine-Grained visual classification","volume":"44","author":"Du","year":"2022","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.knosys.2026.115678_bib0029","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"12047","article-title":"Fine-grained recognition: accounting for subtle differences between similar classes","volume":"34","author":"Sun","year":"2020"},{"issue":"8","key":"10.1016\/j.knosys.2026.115678_bib0030","doi-asserted-by":"crossref","first-page":"5112","DOI":"10.1109\/TNNLS.2021.3126046","article-title":"On the imaginary wings: text-Assisted complex-Valued fusion network for fine-Grained visual classification","volume":"34","author":"Guan","year":"2023","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"10.1016\/j.knosys.2026.115678_bib0031","series-title":"IEEE\/CVF International Conference on Computer Vision, ICCV 2023, Paris, France, October 1\u20136, 2023","first-page":"1621","article-title":"Learning gabor texture features for fine-Grained recognition","author":"Zhu","year":"2023"},{"key":"10.1016\/j.knosys.2026.115678_bib0032","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2023.109305","article-title":"Granularity-aware distillation and structure modeling region proposal network for fine-grained image classification","volume":"137","author":"Ke","year":"2023","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.knosys.2026.115678_bib0033","first-page":"1","article-title":"Interweaving insights: high-order feature interaction for fine-grained visual recognition","author":"Sikdar","year":"2024","journal-title":"Int. J. Comput. Vis."},{"key":"10.1016\/j.knosys.2026.115678_bib0034","series-title":"2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, June 13\u201319, 2020","article-title":"Attention convolutional binary neural tree for fine-Grained visual categorization","author":"Ji","year":"2020"},{"key":"10.1016\/j.knosys.2026.115678_bib0035","series-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","first-page":"1025","article-title":"Counterfactual attention learning for fine-grained visual categorization and re-identification","author":"Rao","year":"2021"},{"key":"10.1016\/j.knosys.2026.115678_bib0036","doi-asserted-by":"crossref","first-page":"137","DOI":"10.1016\/j.neucom.2022.04.037","article-title":"Transformer with peak suppression and knowledge guidance for fine-grained image recognition","volume":"492","author":"Liu","year":"2022","journal-title":"Neurocomputing"},{"key":"10.1016\/j.knosys.2026.115678_bib0037","series-title":"MM \u201921: ACM Multimedia Conference, Virtual Event, China, October 20 \u2013 24, 2021","article-title":"RAMS-Trans: Recurrent attention multi-scale transformer for fine-grained image recognition","author":"Hu","year":"2021"},{"key":"10.1016\/j.knosys.2026.115678_bib0038","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2023.109547","article-title":"AA-Trans: core attention aggregating transformer with information entropy selector for fine-grained visual classification","volume":"140","author":"Wang","year":"2023","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.knosys.2026.115678_bib0039","series-title":"IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2022, Virtual and Singapore, 23\u201327 May 2022","article-title":"A free lunch from ViT: adaptive attention multi-scale fusion transformer for fine-grained visual recognition","author":"Zhang","year":"2022"},{"key":"10.1016\/j.knosys.2026.115678_bib0040","series-title":"International Conference on Machine Learning, ICML 2022, 17\u201323 July 2022, Baltimore, Maryland, USA","article-title":"ViT-NeT: interpretable vision transformers with neural tree decoder","author":"Kim","year":"2022"},{"key":"10.1016\/j.knosys.2026.115678_bib0041","doi-asserted-by":"crossref","first-page":"9015","DOI":"10.1109\/TMM.2023.3244340","article-title":"Fine-Grained visual classification via internal ensemble learning transformer","volume":"25","author":"Xu","year":"2023","journal-title":"IEEE Trans. Multim."},{"key":"10.1016\/j.knosys.2026.115678_bib0042","series-title":"MM \u201922: The 30Th ACM International Conference on Multimedia, Lisboa, Portugal, October 10\u2013 14, 2022","article-title":"SIM-Trans: Structure information modeling transformer for fine-grained visual categorization","author":"Sun","year":"2022"},{"key":"10.1016\/j.knosys.2026.115678_bib0043","article-title":"Multi-Granularity part sampling attention for fine-Grained visual classification","author":"Wang","year":"2024","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.knosys.2026.115678_bib0044","doi-asserted-by":"crossref","first-page":"5312","DOI":"10.1109\/TIP.2024.3459788","article-title":"SIM-OFE: Structure information mining and object-Aware feature enhancement for fine-Grained visual categorization","volume":"33","author":"Sun","year":"2024","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.knosys.2026.115678_bib0045","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"2570","article-title":"Delving into multimodal prompting for fine-grained visual classification","volume":"38","author":"Jiang","year":"2024"},{"key":"10.1016\/j.knosys.2026.115678_bib0046","doi-asserted-by":"crossref","first-page":"394","DOI":"10.1109\/TIP.2024.3523802","article-title":"Universal fine-Grained visual categorization by concept guided learning","volume":"34","author":"Bi","year":"2025","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.knosys.2026.115678_bib0047","series-title":"Proceedings of the 18Th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications, VISIGRAPP 2023, Volume 4: Visapp, Lisbon, Portugal, February 19\u201321, 2023","article-title":"Salient mask-Guided vision transformer for fine-Grained classification","author":"Demidov","year":"2023"},{"key":"10.1016\/j.knosys.2026.115678_bib0048","first-page":"1","article-title":"TransIFC: invariant cues-aware feature concentration learning for efficient fine-grained bird image classification","author":"Liu","year":"2023","journal-title":"IEEE Trans. Multimedia"},{"issue":"12","key":"10.1016\/j.knosys.2026.115678_bib0049","doi-asserted-by":"crossref","first-page":"6116","DOI":"10.1109\/TIP.2019.2924811","article-title":"Piecewise classifier mappings: learning fine-grained learners for novel categories with few examples","volume":"28","author":"Wei","year":"2019","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.knosys.2026.115678_bib0050","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"4136","article-title":"Cross-layer and cross-sample feature optimization network for few-shot fine-grained image classification","volume":"38","author":"Ma","year":"2024"},{"key":"10.1016\/j.knosys.2026.115678_bib0051","article-title":"Adaptive task-aware refining network for few-shot fine-grained image classification","author":"Yu","year":"2024","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10.1016\/j.knosys.2026.115678_bib0052","unstructured":"C. Wah, S. Branson, P. Welinder, P. Perona, S. Belongie, The caltech-ucsd birds-200-2011 dataset(2011)."},{"key":"10.1016\/j.knosys.2026.115678_bib0053","series-title":"Proc. CVPR Workshop on Fine-grained Visual Categorization (FGVC)","article-title":"Novel dataset for fine-grained image categorization: stanford dogs","volume":"2","author":"Khosla","year":"2011"},{"key":"10.1016\/j.knosys.2026.115678_bib0054","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"595","article-title":"Building a bird recognition app and large scale dataset with citizen scientists: the fine print in fine-grained dataset collection","author":"Van Horn","year":"2015"},{"issue":"1","key":"10.1016\/j.knosys.2026.115678_bib0055","doi-asserted-by":"crossref","first-page":"137","DOI":"10.1007\/s11263-023-01873-z","article-title":"Accurate fine-grained object recognition with structure-driven relation graph networks","volume":"132","author":"Wang","year":"2024","journal-title":"Int. J. Comput. Vis."},{"key":"10.1016\/j.knosys.2026.115678_bib0056","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2024.110265","article-title":"FET-FGVC: Feature-enhanced transformer for fine-grained visual classification","volume":"149","author":"Chen","year":"2024","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.knosys.2026.115678_bib0057","series-title":"Proceedings of the 31St ACM International Conference on Multimedia, MM 2023, Ottawa, on, Canada, 29 October 2023- 3 November 2023","first-page":"7648","article-title":"Multi-Part token transformer with dual contrastive learning for fine-grained image classification","author":"Wang","year":"2023"},{"key":"10.1016\/j.knosys.2026.115678_bib0058","unstructured":"S. Maji, E. Rahtu, J. Kannala, M. Blaschko, A. Vedaldi, Fine-grained visual classification of aircraft, arXiv preprint arXiv: 1306.5151(2013)."},{"key":"10.1016\/j.knosys.2026.115678_bib0059","series-title":"Proceedings of the IEEE International Conference on Computer Vision Workshops","first-page":"554","article-title":"3D object representations for fine-grained categorization","author":"Krause","year":"2013"},{"key":"10.1016\/j.knosys.2026.115678_bib0060","first-page":"12116","article-title":"Do vision transformers see like convolutional neural networks?","volume":"34","author":"Raghu","year":"2021","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.knosys.2026.115678_bib0061","series-title":"International Conference on Learning Representations","article-title":"Imagenet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness","author":"Geirhos","year":"2018"},{"key":"10.1016\/j.knosys.2026.115678_bib0062","doi-asserted-by":"crossref","first-page":"4409","DOI":"10.1109\/TMM.2021.3117064","article-title":"Enhancing mixture-of-experts by leveraging attention for fine-grained recognition","volume":"24","author":"Zhang","year":"2021","journal-title":"IEEE Trans. Multimedia"},{"key":"10.1016\/j.knosys.2026.115678_bib0063","doi-asserted-by":"crossref","first-page":"6017","DOI":"10.1109\/TIP.2022.3205215","article-title":"SR-GNN: Spatial relation-Aware graph neural network for fine-Grained image categorization","volume":"31","author":"Bera","year":"2022","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.knosys.2026.115678_bib0064","series-title":"IEEE International Conference on Computer Vision, ICCV 2017, Venice, Italy, October 22\u201329, 2017","first-page":"618","article-title":"Grad-CAM: visual explanations from deep networks via gradient-Based localization","author":"Selvaraju","year":"2017"},{"issue":"301","key":"10.1016\/j.knosys.2026.115678_bib0065","first-page":"1","article-title":"Theoretical foundations of t-SNE for visualizing high-Dimensional clustered data","volume":"23","author":"Cai","year":"2022","journal-title":"J. Mach. Learn. Res."}],"container-title":["Knowledge-Based Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0950705126004181?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0950705126004181?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,3,28]],"date-time":"2026-03-28T12:08:13Z","timestamp":1774699693000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0950705126004181"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,5]]},"references-count":65,"alternative-id":["S0950705126004181"],"URL":"https:\/\/doi.org\/10.1016\/j.knosys.2026.115678","relation":{},"ISSN":["0950-7051"],"issn-type":[{"value":"0950-7051","type":"print"}],"subject":[],"published":{"date-parts":[[2026,5]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"SPG-Mask: Fine-Grained visual classification with Self-Supervised foreground structural prior driven dynamic mask optimization","name":"articletitle","label":"Article Title"},{"value":"Knowledge-Based Systems","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.knosys.2026.115678","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2026 Elsevier B.V. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"115678"}}