{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,31]],"date-time":"2025-08-31T10:34:04Z","timestamp":1756636444967,"version":"3.41.0"},"reference-count":55,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2025,6,1]],"date-time":"2025-06-01T00:00:00Z","timestamp":1748736000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2025,6,1]],"date-time":"2025-06-01T00:00:00Z","timestamp":1748736000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2025,6,1]],"date-time":"2025-06-01T00:00:00Z","timestamp":1748736000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2025,6,1]],"date-time":"2025-06-01T00:00:00Z","timestamp":1748736000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2025,6,1]],"date-time":"2025-06-01T00:00:00Z","timestamp":1748736000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2025,6,1]],"date-time":"2025-06-01T00:00:00Z","timestamp":1748736000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,6,1]],"date-time":"2025-06-01T00:00:00Z","timestamp":1748736000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100005230","name":"Chongqing Natural Science Foundation","doi-asserted-by":"publisher","award":["cstc2021jcyj-msxmX0568"],"award-info":[{"award-number":["cstc2021jcyj-msxmX0568"]}],"id":[{"id":"10.13039\/501100005230","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62176030","62276033"],"award-info":[{"award-number":["62176030","62276033"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Image and Vision Computing"],"published-print":{"date-parts":[[2025,6]]},"DOI":"10.1016\/j.imavis.2025.105550","type":"journal-article","created":{"date-parts":[[2025,4,22]],"date-time":"2025-04-22T16:42:37Z","timestamp":1745340157000},"page":"105550","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":1,"special_numbering":"C","title":["Rethinking the sample relations for few-shot classification"],"prefix":"10.1016","volume":"159","author":[{"given":"Guowei","family":"Yin","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5610-0826","authenticated-orcid":false,"given":"Sheng","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Luwen","family":"Huangfu","sequence":"additional","affiliation":[]},{"given":"Yi","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Xiaohong","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"issue":"6","key":"10.1016\/j.imavis.2025.105550_b1","doi-asserted-by":"crossref","first-page":"84","DOI":"10.1145\/3065386","article-title":"Imagenet classification with deep convolutional neural networks","volume":"60","author":"Krizhevsky","year":"2017","journal-title":"Commun. ACM"},{"key":"10.1016\/j.imavis.2025.105550_b2","doi-asserted-by":"crossref","unstructured":"J. Redmon, S. Divvala, R. Girshick, A. Farhadi, You only look once: Unified, real-time object detection, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 779\u2013788.","DOI":"10.1109\/CVPR.2016.91"},{"key":"10.1016\/j.imavis.2025.105550_b3","article-title":"Faster r-cnn: Towards real-time object detection with region proposal networks","volume":"28","author":"Ren","year":"2015","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.imavis.2025.105550_b4","series-title":"International Conference on Machine Learning","first-page":"1126","article-title":"Model-agnostic meta-learning for fast adaptation of deep networks","author":"Finn","year":"2017"},{"key":"10.1016\/j.imavis.2025.105550_b5","unstructured":"S. Ravi, H. Larochelle, Optimization as a model for few-shot learning, in: International Conference on Learning Representations, 2017."},{"key":"10.1016\/j.imavis.2025.105550_b6","doi-asserted-by":"crossref","unstructured":"K. Lee, S. Maji, A. Ravichandran, S. Soatto, Meta-learning with differentiable convex optimization, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 10657\u201310665.","DOI":"10.1109\/CVPR.2019.01091"},{"key":"10.1016\/j.imavis.2025.105550_b7","article-title":"Prototypical networks for few-shot learning","volume":"30","author":"Snell","year":"2017","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.imavis.2025.105550_b8","article-title":"Matching networks for one shot learning","volume":"29","author":"Vinyals","year":"2016","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.imavis.2025.105550_b9","doi-asserted-by":"crossref","unstructured":"C. Zhang, Y. Cai, G. Lin, C. Shen, Deepemd: Few-shot image classification with differentiable earth mover\u2019s distance and structured classifiers, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 12203\u201312213.","DOI":"10.1109\/CVPR42600.2020.01222"},{"issue":"7","key":"10.1016\/j.imavis.2025.105550_b10","doi-asserted-by":"crossref","first-page":"7970","DOI":"10.1007\/s10489-022-03951-0","article-title":"Light transformer learning embedding for few-shot classification with task-based enhancement","volume":"53","author":"Zhu","year":"2023","journal-title":"Appl. Intell."},{"key":"10.1016\/j.imavis.2025.105550_b11","doi-asserted-by":"crossref","unstructured":"Z. Chen, Y. Fu, Y.-X. Wang, L. Ma, W. Liu, M. Hebert, Image deformation meta-networks for one-shot learning, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 8680\u20138689.","DOI":"10.1109\/CVPR.2019.00888"},{"issue":"9","key":"10.1016\/j.imavis.2025.105550_b12","doi-asserted-by":"crossref","first-page":"4594","DOI":"10.1109\/TIP.2019.2910052","article-title":"Multi-level semantic feature augmentation for one-shot learning","volume":"28","author":"Chen","year":"2019","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.imavis.2025.105550_b13","doi-asserted-by":"crossref","unstructured":"K. Li, Y. Zhang, K. Li, Y. Fu, Adversarial feature hallucination networks for few-shot learning, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 13470\u201313479.","DOI":"10.1109\/CVPR42600.2020.01348"},{"key":"10.1016\/j.imavis.2025.105550_b14","article-title":"Semi-identical twins variational AutoEncoder for few-shot learning","author":"Zhang","year":"2023","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"year":"2019","series-title":"A baseline for few-shot image classification","author":"Dhillon","key":"10.1016\/j.imavis.2025.105550_b15"},{"key":"10.1016\/j.imavis.2025.105550_b16","unstructured":"W.-Y. Chen, Y.-C. Liu, Z. Kira, Y.-C.F. Wang, J.-B. Huang, A Closer Look at Few-shot Classification, in: International Conference on Learning Representations."},{"key":"10.1016\/j.imavis.2025.105550_b17","series-title":"Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XIV 16","first-page":"266","article-title":"Rethinking few-shot image classification: a good embedding is all you need?","author":"Tian","year":"2020"},{"key":"10.1016\/j.imavis.2025.105550_b18","doi-asserted-by":"crossref","unstructured":"M.N. Rizve, S. Khan, F.S. Khan, M. Shah, Exploring complementary strengths of invariant and equivariant representations for few-shot learning, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 10836\u201310846.","DOI":"10.1109\/CVPR46437.2021.01069"},{"key":"10.1016\/j.imavis.2025.105550_b19","doi-asserted-by":"crossref","unstructured":"J. Ma, H. Xie, G. Han, S.-F. Chang, A. Galstyan, W. Abd-Almageed, Partner-assisted learning for few-shot image classification, in: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2021, pp. 10573\u201310582.","DOI":"10.1109\/ICCV48922.2021.01040"},{"key":"10.1016\/j.imavis.2025.105550_b20","doi-asserted-by":"crossref","unstructured":"Y. Zhang, S. Huang, F. Zhou, Generally Boosting Few-Shot Learning with HandCrafted Features, in: Proceedings of the 29th ACM International Conference on Multimedia, 2021, pp. 3143\u20133152.","DOI":"10.1145\/3474085.3475459"},{"key":"10.1016\/j.imavis.2025.105550_b21","series-title":"Machine Learning and Knowledge Discovery in Databases. Research Track: European Conference, ECML PKDD 2021, Bilbao, Spain, September 13\u201317, 2021, Proceedings, Part I 21","first-page":"671","article-title":"Spatial contrastive learning for few-shot classification","author":"Ouali","year":"2021"},{"key":"10.1016\/j.imavis.2025.105550_b22","series-title":"International Conference on Machine Learning","first-page":"5714","article-title":"Self-supervised label augmentation via input transformations","author":"Lee","year":"2020"},{"key":"10.1016\/j.imavis.2025.105550_b23","unstructured":"M. Zhang, J. Zhang, Z. Lu, T. Xiang, M. Ding, S. Huang, IEPT: Instance-level and episode-level pretext tasks for few-shot learning, in: International Conference on Learning Representations, 2021."},{"key":"10.1016\/j.imavis.2025.105550_b24","doi-asserted-by":"crossref","unstructured":"Y. Rong, X. Lu, Z. Sun, Y. Chen, S. Xiong, ESPT: A Self-Supervised Episodic Spatial Pretext Task for Improving Few-Shot Learning, in: Proceedings of the AAAI Conference on Artificial Intelligence, 2023.","DOI":"10.1609\/aaai.v37i8.26148"},{"key":"10.1016\/j.imavis.2025.105550_b25","first-page":"18661","article-title":"Supervised contrastive learning","volume":"33","author":"Khosla","year":"2020","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.imavis.2025.105550_b26","series-title":"International Conference on Machine Learning","first-page":"1597","article-title":"A simple framework for contrastive learning of visual representations","author":"Chen","year":"2020"},{"key":"10.1016\/j.imavis.2025.105550_b27","doi-asserted-by":"crossref","unstructured":"K. He, H. Fan, Y. Wu, S. Xie, R. Girshick, Momentum contrast for unsupervised visual representation learning, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 9729\u20139738.","DOI":"10.1109\/CVPR42600.2020.00975"},{"year":"2018","series-title":"Meta-learning for semi-supervised few-shot classification","author":"Ren","key":"10.1016\/j.imavis.2025.105550_b28"},{"year":"2018","series-title":"Meta-learning with differentiable closed-form solvers","author":"Bertinetto","key":"10.1016\/j.imavis.2025.105550_b29"},{"year":"2011","series-title":"The Caltech-Ucsd Birds-200\u20132011 Dataset","author":"Wah","key":"10.1016\/j.imavis.2025.105550_b30"},{"key":"10.1016\/j.imavis.2025.105550_b31","doi-asserted-by":"crossref","DOI":"10.1016\/j.imavis.2023.104754","article-title":"Contrastive learning with semantic consistency constraint","author":"Guo","year":"2023","journal-title":"Image Vis. Comput."},{"key":"10.1016\/j.imavis.2025.105550_b32","doi-asserted-by":"crossref","DOI":"10.1016\/j.imavis.2022.104570","article-title":"ArCo: Attention-reinforced transformer with contrastive learning for image captioning","volume":"128","author":"Wang","year":"2022","journal-title":"Image Vis. Comput."},{"key":"10.1016\/j.imavis.2025.105550_b33","doi-asserted-by":"crossref","DOI":"10.1016\/j.imavis.2022.104607","article-title":"Dual-level contrastive learning for unsupervised person re-identification","volume":"129","author":"Zhao","year":"2023","journal-title":"Image Vis. Comput."},{"key":"10.1016\/j.imavis.2025.105550_b34","doi-asserted-by":"crossref","DOI":"10.1016\/j.imavis.2023.104786","article-title":"Unsupervised person re-identification by dynamic hybrid contrastive learning","volume":"137","author":"Zhao","year":"2023","journal-title":"Image Vis. Comput."},{"key":"10.1016\/j.imavis.2025.105550_b35","series-title":"ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing","first-page":"1745","article-title":"Self-supervised learning for few-shot image classification","author":"Chen","year":"2021"},{"issue":"7","key":"10.1016\/j.imavis.2025.105550_b36","doi-asserted-by":"crossref","first-page":"1858","DOI":"10.1109\/TIT.2003.813506","article-title":"A new metric for probability distributions","volume":"49","author":"Endres","year":"2003","journal-title":"IEEE Trans. Inform. Theory"},{"key":"10.1016\/j.imavis.2025.105550_b37","series-title":"International Symposium OnInformation Theory, 2004. ISIT 2004. Proceedings","first-page":"31","article-title":"Jensen-Shannon divergence and Hilbert space embedding","author":"Fuglede","year":"2004"},{"key":"10.1016\/j.imavis.2025.105550_b38","doi-asserted-by":"crossref","unstructured":"Z. Wu, Y. Xiong, S.X. Yu, D. Lin, Unsupervised feature learning via non-parametric instance discrimination, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 3733\u20133742.","DOI":"10.1109\/CVPR.2018.00393"},{"key":"10.1016\/j.imavis.2025.105550_b39","doi-asserted-by":"crossref","unstructured":"D. Kang, H. Kwon, J. Min, M. Cho, Relational embedding for few-shot classification, in: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2021, pp. 8822\u20138833.","DOI":"10.1109\/ICCV48922.2021.00870"},{"key":"10.1016\/j.imavis.2025.105550_b40","series-title":"Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part V 16","first-page":"18","article-title":"Associative alignment for few-shot image classification","author":"Afrasiyabi","year":"2020"},{"key":"10.1016\/j.imavis.2025.105550_b41","doi-asserted-by":"crossref","DOI":"10.1016\/j.imavis.2020.104006","article-title":"GIFSL-grafting based improved few-shot learning","volume":"104","author":"Mazumder","year":"2020","journal-title":"Image Vis. Comput."},{"key":"10.1016\/j.imavis.2025.105550_b42","unstructured":"N. Fei, Z. Lu, T. Xiang, S. Huang, Melr: Meta-learning via modeling episode-level relationships for few-shot learning, in: International Conference on Learning Representations, 2021."},{"key":"10.1016\/j.imavis.2025.105550_b43","doi-asserted-by":"crossref","DOI":"10.1016\/j.imavis.2021.104164","article-title":"PDA: Proxy-based domain adaptation for few-shot image recognition","volume":"110","author":"Liu","year":"2021","journal-title":"Image Vis. Comput."},{"key":"10.1016\/j.imavis.2025.105550_b44","first-page":"3179","article-title":"Hybrid graph neural networks for few-shot learning","volume":"vol. 36","author":"Yu","year":"2022"},{"key":"10.1016\/j.imavis.2025.105550_b45","first-page":"1926","article-title":"Adaptive poincar\u00e9 point to set distance for few-shot classification","volume":"vol. 36","author":"Ma","year":"2022"},{"key":"10.1016\/j.imavis.2025.105550_b46","doi-asserted-by":"crossref","DOI":"10.1016\/j.imavis.2022.104574","article-title":"Dual global-aware propagation for few-shot learning","volume":"128","author":"Cui","year":"2022","journal-title":"Image Vis. Comput."},{"key":"10.1016\/j.imavis.2025.105550_b47","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2022.109235","article-title":"Meta-hallucinating prototype for few-shot learning promotion","volume":"136","author":"Zhang","year":"2023","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.imavis.2025.105550_b48","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2022.109170","article-title":"SAPENet: Self-attention based prototype enhancement network for few-shot learning","volume":"135","author":"Huang","year":"2023","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.imavis.2025.105550_b49","article-title":"Disentangled feature representation for few-shot image classification","author":"Cheng","year":"2023","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"10.1016\/j.imavis.2025.105550_b50","doi-asserted-by":"crossref","DOI":"10.1016\/j.imavis.2023.104869","article-title":"Few-shot classification with multisemantic information fusion network","volume":"141","author":"Gao","year":"2024","journal-title":"Image Vis. Comput."},{"key":"10.1016\/j.imavis.2025.105550_b51","first-page":"16687","article-title":"Metadiff: Meta-learning with conditional diffusion for few-shot learning","volume":"vol. 38","author":"Zhang","year":"2024"},{"key":"10.1016\/j.imavis.2025.105550_b52","doi-asserted-by":"crossref","unstructured":"M. Fu, K. Zhu, Instance-based Max-margin for Practical Few-shot Recognition, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2024, pp. 28674\u201328683.","DOI":"10.1109\/CVPR52733.2024.02709"},{"key":"10.1016\/j.imavis.2025.105550_b53","unstructured":"H.-J. Ye, H. Hu, D.-C. Zhan, F. Sha, Few-shot learning via embedding adaptation with set-to-set functions, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 8808\u20138817."},{"key":"10.1016\/j.imavis.2025.105550_b54","doi-asserted-by":"crossref","unstructured":"Y. Chen, Z. Liu, H. Xu, T. Darrell, X. Wang, Meta-baseline: Exploring simple meta-learning for few-shot learning, in: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2021, pp. 9062\u20139071.","DOI":"10.1109\/ICCV48922.2021.00893"},{"key":"10.1016\/j.imavis.2025.105550_b55","unstructured":"W. Xu, Y. Xu, H. Wang, Z. Tu, Attentional constellation nets for few-shot learning, in: International Conference on Learning Representations, 2021."}],"container-title":["Image and Vision Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0262885625001386?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0262885625001386?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2025,5,22]],"date-time":"2025-05-22T06:13:04Z","timestamp":1747894384000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0262885625001386"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6]]},"references-count":55,"alternative-id":["S0262885625001386"],"URL":"https:\/\/doi.org\/10.1016\/j.imavis.2025.105550","relation":{},"ISSN":["0262-8856"],"issn-type":[{"type":"print","value":"0262-8856"}],"subject":[],"published":{"date-parts":[[2025,6]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Rethinking the sample relations for few-shot classification","name":"articletitle","label":"Article Title"},{"value":"Image and Vision Computing","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.imavis.2025.105550","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2025 Elsevier B.V. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"105550"}}