{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,9]],"date-time":"2026-04-09T10:57:16Z","timestamp":1775732236385,"version":"3.50.1"},"reference-count":34,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2024,7,24]],"date-time":"2024-07-24T00:00:00Z","timestamp":1721779200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2024,7,24]],"date-time":"2024-07-24T00:00:00Z","timestamp":1721779200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100004735","name":"Natural Science Foundation of Hunan Province","doi-asserted-by":"publisher","award":["2022JJ30231"],"award-info":[{"award-number":["2022JJ30231"]}],"id":[{"id":"10.13039\/501100004735","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100014472","name":"Scientific Research Foundation of Hunan Provincial Education Department","doi-asserted-by":"publisher","award":["22B0559"],"award-info":[{"award-number":["22B0559"]}],"id":[{"id":"10.13039\/100014472","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100014472","name":"Scientific Research Foundation of Hunan Provincial Education Department","doi-asserted-by":"publisher","award":["2022JJ50051"],"award-info":[{"award-number":["2022JJ50051"]}],"id":[{"id":"10.13039\/100014472","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100014472","name":"Scientific Research Foundation of Hunan Provincial Education Department","doi-asserted-by":"publisher","award":["21A0350"],"award-info":[{"award-number":["21A0350"]}],"id":[{"id":"10.13039\/100014472","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100014472","name":"Scientific Research Foundation of Hunan Provincial Education Department","doi-asserted-by":"publisher","award":["21C0439"],"award-info":[{"award-number":["21C0439"]}],"id":[{"id":"10.13039\/100014472","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Complex Intell. Syst."],"published-print":{"date-parts":[[2024,12]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Humans excel at learning and recognizing objects, swiftly adapting to new concepts with just a few samples. However, current studies in computer vision on few-shot learning have not yet achieved human performance in integrating prior knowledge during the learning process. Humans utilize a hierarchical structure of object categories based on past experiences to facilitate learning and classification. Therefore, we propose a method named n-Hierarchy SEmantic Guided Attention (nHi-SEGA) that acquires abstract superclasses. This allows the model to associate with and pay attention to different levels of objects utilizing semantics and visual features embedded in the class hierarchy (e.g., house finch-bird-animal, goldfish-fish-animal, rose-flower-plant), resembling human cognition. We constructed an nHi-Tree using WordNet and Glove tools and devised two methods to extract hierarchical semantic features, which were then fused with visual features to improve sample feature prototypes.<\/jats:p>","DOI":"10.1007\/s40747-024-01546-5","type":"journal-article","created":{"date-parts":[[2024,7,24]],"date-time":"2024-07-24T09:02:45Z","timestamp":1721811765000},"page":"7577-7589","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["nHi-SEGA: n-Hierarchy SEmantic Guided Attention for few-shot learning"],"prefix":"10.1007","volume":"10","author":[{"given":"Xinpan","family":"Yuan","sequence":"first","affiliation":[]},{"given":"Shaojun","family":"Xie","sequence":"additional","affiliation":[]},{"given":"Zhigao","family":"Zeng","sequence":"additional","affiliation":[]},{"given":"Changyun","family":"Li","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0007-0797-9476","authenticated-orcid":false,"given":"Luda","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,7,24]]},"reference":[{"key":"1546_CR1","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2015) Deep residual learning for image recognition. In: 2016 IEEE conference on computer vision and pattern recognition (CVPR), pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"1546_CR2","doi-asserted-by":"crossref","unstructured":"Szegedy C, Ioffe S, Vanhoucke V, Alemi A (2017) Inception-v4, inception-resnet and the impact of residual connections on learning. In: AAAI conference on artificial intelligence, vol 3. AAAI Press, San Francisco, California, USA, pp 4278\u20134284","DOI":"10.1609\/aaai.v31i1.11231"},{"key":"1546_CR3","doi-asserted-by":"crossref","unstructured":"Fe-Fei L (2003) A Bayesian approach to unsupervised one-shot learning of object categories. In: The ninth IEEE international conference on computer vision (ICCV 2003), vol 2. IEEE Computer Society, USA, pp 1134\u20131141","DOI":"10.1109\/ICCV.2003.1238476"},{"key":"1546_CR4","doi-asserted-by":"crossref","unstructured":"Zhao F, Zhao J, Yan S, Feng J (2018) Dynamic conditional networks for few-shot learning. In: The European conference on computer vision (ECCV 2018). Springer International Publishing, pp 20\u201336 (ISBN: 978-3-030-01267-0)","DOI":"10.1007\/978-3-030-01267-0_2"},{"key":"1546_CR5","doi-asserted-by":"crossref","unstructured":"Yang F, Wang R, Chen X (2022) SEGA: semantic guided attention on visual prototype for few-shot learning. In: 2022 IEEE\/CVF winter conference on applications of computer vision (WACV), pp 1586-1596","DOI":"10.1109\/WACV51458.2022.00165"},{"key":"1546_CR6","doi-asserted-by":"crossref","unstructured":"Zhen L, Hu P, Wang X, Peng D (2019) Deep supervised cross-modal retrieval. In: 2019 IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp 10394\u201310403","DOI":"10.1109\/CVPR.2019.01064"},{"key":"1546_CR7","doi-asserted-by":"crossref","unstructured":"Malinowski M, Rohrbach M, Fritz M (2015) Ask your neurons: a neural-based approach to answering questions about images. In: 2015 IEEE international conference on computer vision (ICCV), pp 1\u20139","DOI":"10.1109\/ICCV.2015.9"},{"issue":"7","key":"1546_CR8","doi-asserted-by":"publisher","first-page":"1425","DOI":"10.1109\/TPAMI.2015.2487986","volume":"38","author":"Z Akata","year":"2015","unstructured":"Akata Z, Perronnin F, Harchaoui Z, Schmid C (2015) Label-embedding for image classification. IEEE Trans Pattern Anal Mach Intell 38(7):1425\u20131438","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"1546_CR9","doi-asserted-by":"crossref","unstructured":"Reed S, Akata Z, Lee H, Schiele B (2016) Learning deep representations of fine-grained visual descriptions. In: 2016 IEEE conference on computer vision and pattern recognition (CVPR), pp 49\u201358","DOI":"10.1109\/CVPR.2016.13"},{"key":"1546_CR10","doi-asserted-by":"crossref","unstructured":"Schonfeld E, Ebrahimi S, Sinha S, Darrell T, Akata Z (2019) Generalized zero-and few-shot learning via aligned variational autoencoders. In: 2019 IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp 8247\u20138255","DOI":"10.1109\/CVPR.2019.00844"},{"key":"1546_CR11","unstructured":"Tsai Y-HH, Salakhutdinov R (2017) Improving one-shot learning through fusing side information. arXiv preprint. arXiv:1710.08347"},{"key":"1546_CR12","doi-asserted-by":"crossref","unstructured":"Tokmakov P, Wang Y-X, Hebert M (2019) Learning compositional representations for few-shot recognition. In: 2019 IEEE\/CVF international conference on computer vision (ICCV), pp 6372\u20136381","DOI":"10.1109\/ICCV.2019.00647"},{"key":"1546_CR13","doi-asserted-by":"crossref","unstructured":"Pahde F, Nabi M, Klein T, Jahnichen P (2018) Discriminative hallucination for multi-modal few-shot learning. In: 2018 25th IEEE international conference on image processing (ICIP). IEEE, pp 156\u2013160","DOI":"10.1109\/ICIP.2018.8451372"},{"key":"1546_CR14","doi-asserted-by":"crossref","unstructured":"Zhao F, Zhao J, Yan S, Feng J (2018) Dynamic conditional networks for few-shot learning. In: Proceedings of the European conference on computer vision (ECCV), pp 19\u201335","DOI":"10.1007\/978-3-030-01267-0_2"},{"key":"1546_CR15","doi-asserted-by":"crossref","unstructured":"Peng Z, Li Z, Zhang J, Li Y, Qi G-J, Tang J (2019) Few-shot image recognition with knowledge transfer. In: 2019 IEEE\/CVF international conference on computer vision (ICCV), pp 441\u2013449","DOI":"10.1109\/ICCV.2019.00053"},{"key":"1546_CR16","doi-asserted-by":"crossref","unstructured":"Li A, Huang W, Lan X, Feng J, Li Z, Wang L (2020) Boosting few-shot learning with adaptive margin loss. In: 2020 IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp 12576\u201312584","DOI":"10.1109\/CVPR42600.2020.01259"},{"key":"1546_CR17","unstructured":"Xing C, Rostamzadeh N, Oreshkin B, O Pinheiro PO (2019) Adaptive cross-modal few-shot learning. Adv Neural Inf Process Syst 32:4847\u20134857"},{"key":"1546_CR18","doi-asserted-by":"crossref","unstructured":"Zhang B, Li X, Ye Y, Huang Z, Zhang L (2021) Prototype completion with primitive knowledge for few-shot learning. In: 2021 IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp 3754\u20133762","DOI":"10.1109\/CVPR46437.2021.00375"},{"key":"1546_CR19","doi-asserted-by":"publisher","first-page":"142","DOI":"10.1016\/j.patrec.2022.06.012","volume":"160","author":"E Schwartz","year":"2022","unstructured":"Schwartz E, Karlinsky L, Feris R, Giryes R, Bronstein A (2022) Baby steps towards few-shot learning with multiple semantics. Pattern Recognit Lett 160:142\u2013147","journal-title":"Pattern Recognit Lett"},{"key":"1546_CR20","doi-asserted-by":"crossref","unstructured":"Tokmakov P, Wang Y-X, Hebert M (2019) Learning compositional representations for few-shot recognition. In: 2019 IEEE\/CVF international conference on computer vision (ICCV), pp 6372\u20136381","DOI":"10.1109\/ICCV.2019.00647"},{"key":"1546_CR21","doi-asserted-by":"crossref","unstructured":"Li A, Luo T, Lu Z, Xiang T, Wang L (2019) Large-scale few-shot learning: knowledge transfer with class hierarchy. In: 2019 IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp 7212\u20137220","DOI":"10.1109\/CVPR.2019.00738"},{"issue":"5","key":"1546_CR22","doi-asserted-by":"publisher","first-page":"2293","DOI":"10.1109\/TKDE.2020.3004939","volume":"34","author":"L Liu","year":"2020","unstructured":"Liu L, Zhou T, Long G, Jiang J, Zhang C (2020) Many-class few-shot learning on multi-granularity class hierarchy. IEEE Trans Knowl Data Eng 34(5):2293\u20132305","journal-title":"IEEE Trans Knowl Data Eng"},{"issue":"3","key":"1546_CR23","doi-asserted-by":"publisher","first-page":"2011","DOI":"10.1007\/s10462-022-10223-3","volume":"56","author":"Z Wu","year":"2023","unstructured":"Wu Z, Zhao H (2023) Hierarchical few-shot learning based on coarse-and fine-grained relation network. Artif Intell Rev 56(3):2011\u20132030","journal-title":"Artif Intell Rev"},{"key":"1546_CR24","doi-asserted-by":"crossref","unstructured":"Zhou J, Ma C, Long D, Xu G, Ding N, Zhang H, Xie P, Liu G (2020) Hierarchy-aware global model for hierarchical text classification. In: Proceedings of the 58th annual meeting of the association for computational linguistics, pp 1106\u20131117","DOI":"10.18653\/v1\/2020.acl-main.104"},{"key":"1546_CR25","doi-asserted-by":"crossref","unstructured":"Guo Y, Xu M, Li J, Ni B, Zhu X, Sun Z, Xu Y (2022) HCSC: hierarchical contrastive selective coding. In: 2022 IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp 9706\u20139715","DOI":"10.1109\/CVPR52688.2022.00948"},{"issue":"11","key":"1546_CR26","doi-asserted-by":"publisher","first-page":"39","DOI":"10.1145\/219717.219748","volume":"38","author":"GA Miller","year":"1995","unstructured":"Miller GA (1995) WordNet: a lexical database for English. Commun ACM 38(11):39\u201341","journal-title":"Commun ACM"},{"key":"1546_CR27","doi-asserted-by":"crossref","unstructured":"Pennington J, Socher R, Manning CD (2014) Glove: global vectors for word representation. In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), pp 1532\u20131543","DOI":"10.3115\/v1\/D14-1162"},{"key":"1546_CR28","unstructured":"Snell J, Swersky K, Zemel R (2017) Prototypical networks for few-shot learning. Adv Neural Inf Process Syst 30:4077\u20134087"},{"key":"1546_CR29","unstructured":"Bahdanau D, Cho K, Bengio Y (2014) Neural machine translation by jointly learning to align and translate. arXiv preprint. arXiv:1409.0473"},{"key":"1546_CR30","unstructured":"Vinyals O, Blundell C, Lillicrap T, Wierstra D et al (2016) Matching networks for one shot learning. In: Advances in Neural Information Processing Systems (NeurlIPS), vol 29. Curran Associates Inc., Red Hook, NY, USA pp 3637\u20133645 (ISBN: 9781510838819)"},{"key":"1546_CR31","unstructured":"Ren M, Triantafillou E, Ravi S, Snell J, Swersky K, Tenenbaum JB, Larochelle H, Zemel RS (2018) Meta-learning for semi-supervised few-shot classification. In: Proceedings of 6th international conference on learning representations (ICLR)"},{"key":"1546_CR32","unstructured":"Finn C, Abbeel P, Levine S (2017) Model-agnostic meta-learning for fast adaptation of deep networks. In: International conference on machine learning. PMLR, pp 1126\u20131135"},{"key":"1546_CR33","doi-asserted-by":"crossref","unstructured":"Zhang C, Cai Y, Lin, G, Shen C (2020) DeepEMD: few-shot image classification with differentiable earth mover\u2019s distance and structured classifiers. In: 2020 IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp 12203\u201312213","DOI":"10.1109\/CVPR42600.2020.01222"},{"key":"1546_CR34","doi-asserted-by":"crossref","unstructured":"Ye H-J, Hu H, Zhan D-C, Sha F (2020) Few-shot learning via embedding adaptation with set-to-set functions. In: 2020 IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp 8805-8814","DOI":"10.1109\/CVPR42600.2020.00883"}],"container-title":["Complex &amp; Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-024-01546-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s40747-024-01546-5\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-024-01546-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,16]],"date-time":"2024-10-16T22:09:34Z","timestamp":1729116574000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s40747-024-01546-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,7,24]]},"references-count":34,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2024,12]]}},"alternative-id":["1546"],"URL":"https:\/\/doi.org\/10.1007\/s40747-024-01546-5","relation":{},"ISSN":["2199-4536","2198-6053"],"issn-type":[{"value":"2199-4536","type":"print"},{"value":"2198-6053","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,7,24]]},"assertion":[{"value":"17 November 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"15 May 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"24 July 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors do not have any possible conflicts of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}