{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,8]],"date-time":"2025-09-08T06:08:36Z","timestamp":1757311716907,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":27,"publisher":"ACM","license":[{"start":{"date-parts":[[2021,9,24]],"date-time":"2021-09-24T00:00:00Z","timestamp":1632441600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2021,9,24]]},"DOI":"10.1145\/3488933.3488978","type":"proceedings-article","created":{"date-parts":[[2022,2,25]],"date-time":"2022-02-25T11:36:59Z","timestamp":1645789019000},"page":"128-136","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":3,"title":["Image Caption Based on Bigru and Attention Hybrid Model"],"prefix":"10.1145","author":[{"given":"Qiao","family":"Yucong","sequence":"first","affiliation":[{"name":"Xi'an University of Posts and Telecommunications, China"}]},{"given":"Ma","family":"Li","sequence":"additional","affiliation":[{"name":"Xi'an University of Posts and Telecommunications, China"}]}],"member":"320","published-online":{"date-parts":[[2022,2,25]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"ORIOL V AlEXANDER T SAMY B etal Show and Tell:A Neural Image Caption Generator.[J]. IEICE Transactions on Fundamentals of Electronics Communications and Computer Sciences 2014 abs\/1411.4555  ORIOL V AlEXANDER T SAMY B et al. Show and Tell:A Neural Image Caption Generator.[J]. IEICE Transactions on Fundamentals of Electronics Communications and Computer Sciences 2014 abs\/1411.4555"},{"key":"e_1_3_2_1_2_1","volume-title":"Deep Visual-Semantic Alignments for Generating Image Descriptions.[J].IEICE Transactionson Fundamenta of Electronics,Communications and Computer Sciences","author":"ANDREJ K","year":"2014","unstructured":"ANDREJ K ,Li F, F. Deep Visual-Semantic Alignments for Generating Image Descriptions.[J].IEICE Transactionson Fundamenta of Electronics,Communications and Computer Sciences , 2014 , abs\/1412.2306 ANDREJ K,Li F,F. Deep Visual-Semantic Alignments for Generating Image Descriptions.[J].IEICE Transactionson Fundamenta of Electronics,Communications and Computer Sciences, 2014, abs\/1412.2306"},{"key":"e_1_3_2_1_3_1","first-page":"1233","article-title":"Visual story telling [C]\/\/Proceeding soft the 2016 Conference of the North American Chapter of the Association for Computational Linguistics","volume":"2016","author":"HUANG T H","unstructured":"HUANG T H , FERRARO F , Visual story telling [C]\/\/Proceeding soft the 2016 Conference of the North American Chapter of the Association for Computational Linguistics : Human Language Technologies , 2016 : 1233 - 1239 . HUANG T H,FERRARO F,et al.Visual story telling [C]\/\/Proceeding soft the 2016 Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies, 2016:1233-1239.","journal-title":"Human Language Technologies"},{"key":"e_1_3_2_1_4_1","first-page":"1485","volume-title":"Proceedings of the IEEE,2010","author":"YAO B Z","unstructured":"YAO B Z , YANG X , LIN L , Image 2text:Image parsing to text description[J] . Proceedings of the IEEE,2010 ,98)8(: 1485 - 1508 . YAO B Z,YANG X,LIN L,et al. Image 2text:Image parsing to text description[J]. Proceedings of the IEEE,2010,98)8(:1485-1508."},{"key":"e_1_3_2_1_5_1","first-page":"155","article-title":"of image annotation based on keywords[J].","volume":"2011","author":"GUO Q J","unstructured":"GUO Q J ,Ding J,Li N,Overview of image annotation based on keywords[J]. Computer Engineering and Applications , 2011 ,47 (30): 155 - 158 . GUO Q J,Ding J,Li N,Overview of image annotation based on keywords[J].Computer Engineering and Applications,2011,47 (30):155-158.","journal-title":"Computer Engineering and Applications"},{"key":"e_1_3_2_1_6_1","first-page":"1292","article-title":"Image description using visual dependency representations[C].Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing.Seattle,Washington","volume":"2013","author":"ELLIOTT D","unstructured":"ELLIOTT D , KELLER F . Image description using visual dependency representations[C].Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing.Seattle,Washington , USA:ACL , 2013 : 1292 - 1302 . ELLIOTT D, KELLER F. Image description using visual dependency representations[C].Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing.Seattle,Washington,USA:ACL,2013:1292-1302.","journal-title":"USA:ACL"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298932"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"crossref","unstructured":"BO C CUI S A and LI T \"Tensor Locality Preserving Projections Based Urban Building Areas Extraction from High-Resolution SAR Images \" Vol. 7 No. 4 pp. 291-296 November 2016. doi: 10.12720\/jait.7.4.291-296.  BO C CUI S A and LI T \"Tensor Locality Preserving Projections Based Urban Building Areas Extraction from High-Resolution SAR Images \" Vol. 7 No. 4 pp. 291-296 November 2016. doi: 10.12720\/jait.7.4.291-296.","DOI":"10.12720\/jait.7.4.291-296"},{"key":"e_1_3_2_1_9_1","first-page":"2017","article-title":"Show and tell:Lessons learned from the 2015 mscoco image captioning challenge[J]","volume":"0162","author":"VINYALS O","unstructured":"VINYALS O , TOSHEV A , BENGIO S , Show and tell:Lessons learned from the 2015 mscoco image captioning challenge[J] . IEEE Transactions on Pattern Analysis And Machine Intelligence (S 0162-8828 ), 2017 , 39(4): 652-663 VINYALS O, TOSHEV A, BENGIO S, Show and tell:Lessons learned from the 2015 mscoco image captioning challenge[J]. IEEE Transactions on Pattern Analysis And Machine Intelligence (S0162-8828), 2017, 39(4): 652-663","journal-title":"IEEE Transactions on Pattern Analysis And Machine Intelligence (S"},{"key":"e_1_3_2_1_10_1","volume-title":"Explain images with multi modal recurrent neural networks","author":"MAO J","year":"2014","unstructured":"MAO J , XU W , YANG Y , WANG J ,and YUILLE A L . \u201c Explain images with multi modal recurrent neural networks .\u201d arXiv preprint arXiv:1410.1090 ( 2014 ). MAO J,XU W,YANG Y,WANG J,and YUILLE A L. \u201cExplain images with multi modal recurrent neural networks.\u201d arXiv preprint arXiv:1410.1090 (2014)."},{"volume-title":"Proceedings of the IEEE conference on computer vision and pattern recognition. 2015","author":"VINYALS O","key":"e_1_3_2_1_11_1","unstructured":"VINYALS O , TOSHEV A , BENGIO S , and ERHAN D. \u201c Show and tell:A neural image caption generator .\u201d Proceedings of the IEEE conference on computer vision and pattern recognition. 2015 VINYALS O, TOSHEV A, BENGIO S, and ERHAN D. \u201cShow and tell:A neural image caption generator.\u201d Proceedings of the IEEE conference on computer vision and pattern recognition. 2015"},{"key":"e_1_3_2_1_12_1","unstructured":"TENG J B KONG W W TIAN Q X WANG Z Q.Text Classification Method Based on LSTM-Attention and CNN Mixed Model[J\/OL].Computer Engineering and Applications1-9[2021-04-10].  TENG J B KONG W W TIAN Q X WANG Z Q.Text Classification Method Based on LSTM-Attention and CNN Mixed Model[J\/OL].Computer Engineering and Applications1-9[2021-04-10]."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"crossref","unstructured":"WANG G Z LI A HE G J LIU J B Zhang Z M and WANG M M \"Classification of High Spatial Resolution Remote Sensing Images Based on Decision Fusion \" Vol. 8 No. 1 pp. 42-46 February 2017. doi: 10.12720\/jait.8.1.42-46  WANG G Z LI A HE G J LIU J B Zhang Z M and WANG M M \"Classification of High Spatial Resolution Remote Sensing Images Based on Decision Fusion \" Vol. 8 No. 1 pp. 42-46 February 2017. doi: 10.12720\/jait.8.1.42-46","DOI":"10.12720\/jait.8.1.42-46"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.12720\/jait.11.4.192-199"},{"issue":"4","key":"e_1_3_2_1_15_1","first-page":"194","article-title":"Features Extraction from High Frequency Domain for Retina Digital Images Classification","volume":"4","author":"SALIM L","year":"2013","unstructured":"SALIM L , \" Features Extraction from High Frequency Domain for Retina Digital Images Classification ,\" Journal of Advances in Information Technology , Vol. 4 , No. 4 , pp. 194 - 198 , November , 2013 .doi:10.4304\/jait.4.4.194-198 SALIM L, \"Features Extraction from High Frequency Domain for Retina Digital Images Classification,\" Journal of Advances in Information Technology, Vol. 4, No. 4, pp. 194-198, November, 2013.doi:10.4304\/jait.4.4.194-198","journal-title":"Journal of Advances in Information Technology"},{"key":"e_1_3_2_1_16_1","unstructured":"J BA J MNIH V KAVUKCUOGLU K. Multiple Object Recognition with Visual Attention[J]. Computer ence 2014.  J BA J MNIH V KAVUKCUOGLU K. Multiple Object Recognition with Visual Attention[J]. Computer ence 2014."},{"key":"e_1_3_2_1_17_1","first-page":"2204","article-title":"Recurrent models of visual attention\/\/Advances","volume":"2014","author":"MNIH V","unstructured":"C MNIH V , HEESS N ,Graves A . Recurrent models of visual attention\/\/Advances in Neural Information Processing Systems. 2014 : 2204 - 2212 . C MNIH V,HEESS N,Graves A. Recurrent models of visual attention\/\/Advances in Neural Information Processing Systems. 2014: 2204-2212.","journal-title":"Neural Information Processing Systems."},{"key":"e_1_3_2_1_18_1","volume-title":"IEEE","author":"FU J","year":"2017","unstructured":"C FU J , ZHENG H , MEI T. Look Closer to See Better:Recurrent Attention Convolutional Neural Network for Fine-Grained Image Recognition\/\/ IEEE Conference on Computer Vision & Pattern Recognition . IEEE , 2017 . C FU J, ZHENG H, MEI T. Look Closer to See Better:Recurrent Attention Convolutional Neural Network for Fine-Grained Image Recognition\/\/ IEEE Conference on Computer Vision & Pattern Recognition. IEEE, 2017."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"crossref","unstructured":"K CHO B VAN MERRIENBOER C GULCEHRE D BAHDANAU F BOUGARES H SCHWENK and Y BENGIO \u201cLearning phrase representations using RNN encoder-decoder for statistical machine translation \u201d2014 arXiv: 1406.1078.[Online].Available:http:\/\/arxiv.org\/abs\/1406.1078  K CHO B VAN MERRIENBOER C GULCEHRE D BAHDANAU F BOUGARES H SCHWENK and Y BENGIO \u201cLearning phrase representations using RNN encoder-decoder for statistical machine translation \u201d2014 arXiv:1406.1078.[Online].Available:http:\/\/arxiv.org\/abs\/1406.1078","DOI":"10.3115\/v1\/D14-1179"},{"key":"e_1_3_2_1_20_1","volume-title":"Empiricalevaluation of gated recurrent neural networks on sequence modeling [J]. arXiv","author":"CHUNG J","year":"2014","unstructured":"CHUNG J , GULCEHRE C , CHO K H , Empiricalevaluation of gated recurrent neural networks on sequence modeling [J]. arXiv , 2014 :1412. 3555. CHUNG J,GULCEHRE C,CHO K H,et al. Empiricalevaluation of gated recurrent neural networks on sequence modeling [J]. arXiv, 2014:1412. 3555."},{"key":"e_1_3_2_1_21_1","volume-title":"Microsoft coco:Common objects in context[C]\/\/Proceedings of the European Conference on Computer Vision","author":"LIN T Y","year":"2014","unstructured":"LIN T Y , MAIRE M , BELONGIE S , Microsoft coco:Common objects in context[C]\/\/Proceedings of the European Conference on Computer Vision . Springer ,Cham, 2014 :730-755. LIN T Y,MAIRE M,BELONGIE S,et al. Microsoft coco:Common objects in context[C]\/\/Proceedings of the European Conference on Computer Vision.Springer,Cham,2014:730-755."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00166"},{"key":"e_1_3_2_1_23_1","first-page":"2641","article-title":"Flickr30k entities:collecting region-to-phrase correspondences for richer image-to-sentence models[C]\/\/","volume":"2015","author":"PLUMMER B A","unstructured":"PLUMMER B A , WANG L ,Cervantes C M, Flickr30k entities:collecting region-to-phrase correspondences for richer image-to-sentence models[C]\/\/ Proceedings of the IEEE International Conferenceon Computer Vision , 2015 : 2641 - 2649 . PLUMMER B A, WANG L,Cervantes C M,et al. Flickr30k entities:collecting region-to-phrase correspondences for richer image-to-sentence models[C]\/\/ Proceedings of the IEEE International Conferenceon Computer Vision,2015:2641-2649.","journal-title":"Proceedings of the IEEE International Conferenceon Computer Vision"},{"key":"e_1_3_2_1_24_1","volume-title":"Bleu:a method for automatic evaluation of machine translation [C]\/\/Proceedings of the 40th Annual Meeting on Association for Computational Linguistics","author":"PAPINENI K","year":"2002","unstructured":"PAPINENI K , ROUKOS S , WARD T , Bleu:a method for automatic evaluation of machine translation [C]\/\/Proceedings of the 40th Annual Meeting on Association for Computational Linguistics . Association for Computationa lLinguistics, 2002 :311-318. PAPINENI K,ROUKOS S,WARD T,et al.Bleu:a method for automatic evaluation of machine translation [C]\/\/Proceedings of the 40th Annual Meeting on Association for Computational Linguistics. Association for Computationa lLinguistics, 2002:311-318."},{"key":"e_1_3_2_1_25_1","unstructured":"BANERJEE S LAVIE A.Meteor:An automatic metric for mt evaluation with improved correlation with human judgments[C]\/\/Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and\/or Summarization 2005:65-72.  BANERJEE S LAVIE A.Meteor:An automatic metric for mt evaluation with improved correlation with human judgments[C]\/\/Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and\/or Summarization 2005:65-72."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"crossref","unstructured":"LIN C Y HOVY E.Automatic evaluation of summaries using n-gram co-occurrence statistics[C]\/\/Pro- ceedings of the 2003 Human Language Technology Conference of the North American Chapter of the Association for Computational Linguistics 2003:150- 157.  LIN C Y HOVY E.Automatic evaluation of summaries using n-gram co-occurrence statistics[C]\/\/Pro- ceedings of the 2003 Human Language Technology Conference of the North American Chapter of the Association for Computational Linguistics 2003:150- 157.","DOI":"10.3115\/1073445.1073465"},{"key":"e_1_3_2_1_27_1","first-page":"4566","article-title":"consensus-based imaged escription evaluation[C]\/\/","volume":"2015","author":"VEDANTAM R","unstructured":"VEDANTAM R , LAWRENCE Z , ITNICK C , PARIKH D. Cider : consensus-based imaged escription evaluation[C]\/\/ Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition , 2015 : 4566 - 4575 . VEDANTAM R,LAWRENCE Z,ITNICK C,PARIKH D.Cider:consensus-based imaged escription evaluation[C]\/\/ Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,2015:4566-4575.","journal-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition"}],"event":{"name":"AIPR 2021: 2021 4th International Conference on Artificial Intelligence and Pattern Recognition","acronym":"AIPR 2021","location":"Xiamen China"},"container-title":["2021 4th International Conference on Artificial Intelligence and Pattern Recognition"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3488933.3488978","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3488933.3488978","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:31:28Z","timestamp":1750188688000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3488933.3488978"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,9,24]]},"references-count":27,"alternative-id":["10.1145\/3488933.3488978","10.1145\/3488933"],"URL":"https:\/\/doi.org\/10.1145\/3488933.3488978","relation":{},"subject":[],"published":{"date-parts":[[2021,9,24]]},"assertion":[{"value":"2022-02-25","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}