{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,8]],"date-time":"2025-10-08T16:01:19Z","timestamp":1759939279934,"version":"3.37.3"},"reference-count":36,"publisher":"Wiley","license":[{"start":{"date-parts":[[2022,1,10]],"date-time":"2022-01-10T00:00:00Z","timestamp":1641772800000},"content-version":"unspecified","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61402350","61103143","U1404620","U1404622","182102310034","172102310124","212102210400","20A520046"],"award-info":[{"award-number":["61402350","61103143","U1404620","U1404622","182102310034","172102310124","212102210400","20A520046"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Key Scientific and Technological Project of Henan Province","award":["61402350","61103143","U1404620","U1404622","182102310034","172102310124","212102210400","20A520046"],"award-info":[{"award-number":["61402350","61103143","U1404620","U1404622","182102310034","172102310124","212102210400","20A520046"]}]},{"DOI":"10.13039\/501100009101","name":"Education Department of Henan Province","doi-asserted-by":"publisher","award":["61402350","61103143","U1404620","U1404622","182102310034","172102310124","212102210400","20A520046"],"award-info":[{"award-number":["61402350","61103143","U1404620","U1404622","182102310034","172102310124","212102210400","20A520046"]}],"id":[{"id":"10.13039\/501100009101","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Scientific Programming"],"published-print":{"date-parts":[[2022,1,10]]},"abstract":"<jats:p>In view of the complexity of the multimodal environment and the existing shallow network structure that cannot achieve high-precision image and text retrieval, a cross-modal image and text retrieval method combining efficient feature extraction and interactive learning convolutional autoencoder (CAE) is proposed. First, the residual network convolution kernel is improved by incorporating two-dimensional principal component analysis (2DPCA) to extract image features and extracting text features through long short-term memory (LSTM) and word vectors to efficiently extract graphic features. Then, based on interactive learning CAE, cross-modal retrieval of images and text is realized. Among them, the image and text features are respectively input to the two input terminals of the dual-modal CAE, and the image-text relationship model is obtained through the interactive learning of the middle layer to realize the image-text retrieval. Finally, based on Flickr30K, MSCOCO, and Pascal VOC 2007 datasets, the proposed method is experimentally demonstrated. The results show that the proposed method can complete accurate image retrieval and text retrieval. Moreover, the mean average precision (MAP) has reached more than 0.3, the area of precision-recall rate (PR) curves are better than other comparison methods, and they are applicable.<\/jats:p>","DOI":"10.1155\/2022\/7314599","type":"journal-article","created":{"date-parts":[[2022,1,11]],"date-time":"2022-01-11T00:50:12Z","timestamp":1641862212000},"page":"1-12","source":"Crossref","is-referenced-by-count":2,"title":["A Cross-Modal Image and Text Retrieval Method Based on Efficient Feature Extraction and Interactive Learning CAE"],"prefix":"10.1155","volume":"2022","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-4413-8995","authenticated-orcid":true,"given":"Xiuye","family":"Yin","sequence":"first","affiliation":[{"name":"School of Computer Science and Technology, Zhoukou Normal University, Henan, Zhoukou 466001, China"}]},{"given":"Liyong","family":"Chen","sequence":"additional","affiliation":[{"name":"School of Network Engineering, Zhoukou Normal University, Henan, Zhoukou 466001, China"}]}],"member":"311","reference":[{"key":"1","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2018.05.052"},{"key":"2","doi-asserted-by":"publisher","DOI":"10.1109\/tmm.2019.2942494"},{"key":"3","doi-asserted-by":"publisher","DOI":"10.1109\/tgrs.2018.2841808"},{"key":"4","doi-asserted-by":"publisher","DOI":"10.1109\/tmm.2019.2922128"},{"key":"5","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2019.12.086"},{"key":"6","doi-asserted-by":"publisher","DOI":"10.1109\/tip.2020.2963957"},{"key":"7","doi-asserted-by":"publisher","DOI":"10.4018\/ijghpc.2018070103"},{"key":"8","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2019.05.017"},{"key":"9","doi-asserted-by":"publisher","DOI":"10.1145\/3362065"},{"key":"10","doi-asserted-by":"publisher","DOI":"10.1016\/j.patrec.2020.02.006"},{"key":"11","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2019.12.096"},{"key":"12","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2019.02.004"},{"key":"13","doi-asserted-by":"publisher","DOI":"10.5555\/3288251.3288318"},{"key":"14","doi-asserted-by":"publisher","DOI":"10.1007\/s00530-016-0532-7"},{"key":"15","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-018-5767-1"},{"key":"16","doi-asserted-by":"publisher","DOI":"10.3969\/j.issn.0253-2778.2018.04.008"},{"key":"17","doi-asserted-by":"publisher","DOI":"10.14311\/nnw.2018.28.018"},{"key":"18","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-018-7068-0"},{"key":"19","doi-asserted-by":"publisher","DOI":"10.1109\/tnnls.2018.2856253"},{"key":"20","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-019-7605-5"},{"key":"21","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-018-6556-6"},{"key":"22","doi-asserted-by":"publisher","DOI":"10.1109\/tbiom.2020.2983467"},{"first-page":"614","article-title":"Attention-aware deep adversarial hashing for cross-modal retrieval","author":"Z. Xi","key":"23"},{"key":"24","doi-asserted-by":"publisher","DOI":"10.1109\/tie.2018.2873547"},{"key":"25","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-019-07909-2"},{"key":"26","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-018-5970-0"},{"key":"27","doi-asserted-by":"publisher","DOI":"10.1145\/2647868"},{"key":"28","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2018.11.042"},{"key":"29","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR35066.2017"},{"article-title":"Learning Semantic Concepts and Order for Image and Sentence matching","year":"2017","author":"Y. Huang","key":"30"},{"key":"31","doi-asserted-by":"publisher","DOI":"10.1109\/access.2019.2939650"},{"key":"32","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2018.2882155"},{"key":"33","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-019-08202-y"},{"key":"34","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-018-7072-4"},{"key":"35","doi-asserted-by":"publisher","DOI":"10.1109\/jbhi.2018.2882647"},{"key":"36","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2018.2860898"}],"container-title":["Scientific Programming"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/downloads.hindawi.com\/journals\/sp\/2022\/7314599.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/downloads.hindawi.com\/journals\/sp\/2022\/7314599.xml","content-type":"application\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/downloads.hindawi.com\/journals\/sp\/2022\/7314599.pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,1,11]],"date-time":"2022-01-11T00:50:19Z","timestamp":1641862219000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.hindawi.com\/journals\/sp\/2022\/7314599\/"}},"subtitle":[],"editor":[{"given":"Le","family":"Sun","sequence":"additional","affiliation":[]}],"short-title":[],"issued":{"date-parts":[[2022,1,10]]},"references-count":36,"alternative-id":["7314599","7314599"],"URL":"https:\/\/doi.org\/10.1155\/2022\/7314599","relation":{},"ISSN":["1875-919X","1058-9244"],"issn-type":[{"type":"electronic","value":"1875-919X"},{"type":"print","value":"1058-9244"}],"subject":[],"published":{"date-parts":[[2022,1,10]]}}}