{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,10,30]],"date-time":"2024-10-30T00:57:45Z","timestamp":1730249865404,"version":"3.28.0"},"reference-count":27,"publisher":"IEEE","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2018,7]]},"DOI":"10.1109\/icme.2018.8486437","type":"proceedings-article","created":{"date-parts":[[2018,10,11]],"date-time":"2018-10-11T22:34:18Z","timestamp":1539297258000},"page":"1-6","source":"Crossref","is-referenced-by-count":3,"title":["Refining Attention: A Sequential Attention Model for Image Captioning"],"prefix":"10.1109","author":[{"given":"Fang","family":"Fang","sequence":"first","affiliation":[]},{"given":"Qinyu","family":"Li","sequence":"additional","affiliation":[]},{"given":"Hanli","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Pengjie","family":"Tang","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"ref11","first-page":"2048","article-title":"Show, attend and tell: Neural image caption generation with visual attention","author":"xu","year":"2015","journal-title":"Proc ICML'15"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.667"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.29"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.503"},{"key":"ref15","doi-asserted-by":"crossref","first-page":"2321","DOI":"10.1109\/TPAMI.2016.2642953","article-title":"Aligning where to see and what to tell: image caption with region-based attention and scene factorization","volume":"39","author":"jin","year":"2017","journal-title":"IEEE Trans Pattern Analysis and Machine Intelligence"},{"key":"ref16","first-page":"740","article-title":"Microsoft coco: Common objects in context","author":"lin","year":"2014","journal-title":"Proc ECCV'14"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1613\/jair.3994"},{"key":"ref18","first-page":"311","article-title":"BLEU: a method for automatic evaluation of machine translation","author":"papineni","year":"2002","journal-title":"Proc ACL'02"},{"key":"ref19","first-page":"65","article-title":"METEOR: An automatic metric for MT evaluation with improved correlation with human judgments","volume":"29","author":"banerjee","year":"2005","journal-title":"Proc ACL Workshop IEEMMTS'05"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00177"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2017\/563"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-10593-2_35"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298935"},{"key":"ref5","first-page":"3104","article-title":"Sequence to sequence learning with neural networks","author":"sutskever","year":"2014","journal-title":"Proc NIP-S'14"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298932"},{"key":"ref2","first-page":"1292","article-title":"Image description using visual dependency representations","author":"elliott","year":"2013","journal-title":"Proc EMNLP&#x2019;11"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref1","first-page":"15","article-title":"Every picture tells a story: Generating sentences from images","author":"farhadi","year":"2010","journal-title":"Proceedings of the ECCV'10"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7299087"},{"key":"ref22","article-title":"Deep captioning with multimodal recurrent neural networks (m-RNN)","author":"mao","year":"2015","journal-title":"Proc ICLR'15"},{"key":"ref21","first-page":"1929","article-title":"Dropout: a simple way to prevent neural networks from overfitting","volume":"15","author":"srivastava","year":"2014","journal-title":"J Machine Learning Research"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.140"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.277"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.100"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.524"}],"event":{"name":"2018 IEEE International Conference on Multimedia and Expo (ICME)","start":{"date-parts":[[2018,7,23]]},"location":"San Diego, CA","end":{"date-parts":[[2018,7,27]]}},"container-title":["2018 IEEE International Conference on Multimedia and Expo (ICME)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/8472825\/8486434\/08486437.pdf?arnumber=8486437","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,1,26]],"date-time":"2022-01-26T09:01:46Z","timestamp":1643187706000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8486437\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2018,7]]},"references-count":27,"URL":"https:\/\/doi.org\/10.1109\/icme.2018.8486437","relation":{},"subject":[],"published":{"date-parts":[[2018,7]]}}}