{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,9]],"date-time":"2026-03-09T23:12:09Z","timestamp":1773097929974,"version":"3.50.1"},"reference-count":67,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"11","license":[{"start":{"date-parts":[[2018,11,1]],"date-time":"2018-11-01T00:00:00Z","timestamp":1541030400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61572108"],"award-info":[{"award-number":["61572108"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61632007"],"award-info":[{"award-number":["61632007"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100013314","name":"111 Project","doi-asserted-by":"crossref","award":["B17008"],"award-info":[{"award-number":["B17008"]}],"id":[{"id":"10.13039\/501100013314","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. on Image Process."],"published-print":{"date-parts":[[2018,11]]},"DOI":"10.1109\/tip.2018.2855422","type":"journal-article","created":{"date-parts":[[2018,7,12]],"date-time":"2018-07-12T19:05:00Z","timestamp":1531422300000},"page":"5600-5611","source":"Crossref","is-referenced-by-count":190,"title":["Video Captioning by Adversarial LSTM"],"prefix":"10.1109","volume":"27","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-5070-4511","authenticated-orcid":false,"given":"Yang","family":"Yang","sequence":"first","affiliation":[]},{"given":"Jie","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Jiangbo","family":"Ai","sequence":"additional","affiliation":[]},{"given":"Yi","family":"Bin","sequence":"additional","affiliation":[]},{"given":"Alan","family":"Hanjalic","sequence":"additional","affiliation":[]},{"given":"Heng Tao","family":"Shen","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9122-6141","authenticated-orcid":false,"given":"Yanli","family":"Ji","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.323"},{"key":"ref38","author":"press","year":"2017","journal-title":"Language generation with recurrent generative adversarial networks without pre-training"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.108"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2017.2729019"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.127"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/TCYB.2018.2831447"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D17-1230"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.496"},{"key":"ref35","article-title":"Adaptive feature abstraction for translating video to text","author":"pu","year":"2018","journal-title":"Proc 32nd AAAI Conf Artif Intell"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.111"},{"key":"ref60","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298940"},{"key":"ref62","first-page":"1","article-title":"Rouge: A package for automatic evaluation of summaries","volume":"8","author":"lin","year":"2004","journal-title":"Proc ACL"},{"key":"ref61","first-page":"311","article-title":"BLEU: A method for automatic evaluation of machine translation","author":"papineni","year":"2002","journal-title":"Proc ACL"},{"key":"ref63","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/W14-3348"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2013.337"},{"key":"ref64","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298754"},{"key":"ref27","doi-asserted-by":"crossref","first-page":"541","DOI":"10.1609\/aaai.v27i1.8679","article-title":"Generating natural-language video descriptions using text-mined knowledge","volume":"1","author":"krishnamoorthy","year":"2013","journal-title":"Proc AAAI"},{"key":"ref65","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298932"},{"key":"ref66","author":"chen","year":"2015","journal-title":"Microsoft COCO captions Data collection and evaluation server"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1145\/2964284.2967258"},{"key":"ref67","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-24947-6_17"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298878"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.512"},{"key":"ref20","author":"yu","year":"2016","journal-title":"Seqgan sequence generative adversarial nets with policy gradient"},{"key":"ref22","author":"husz\u00e1r","year":"2015","journal-title":"How (not) to Train your Generative Model Scheduled Sampling Likelihood Adversary?"},{"key":"ref21","author":"kwak","year":"2016","journal-title":"Generating images part by part with composite generative adversarial networks"},{"key":"ref24","first-page":"1","article-title":"Generating text via adversarial training","author":"zhang","year":"2016","journal-title":"Proc NIPS Workshop on Adversarial Training"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2017.2749147"},{"key":"ref26","first-page":"2493","article-title":"Natural language processing (almost) from scratch","volume":"12","author":"collobert","year":"2011","journal-title":"J Mach Learn Res"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/D14-1181"},{"key":"ref50","author":"maddison","year":"2016","journal-title":"The concrete distribution A continuous relaxation of discrete random variables"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/D14-1162"},{"key":"ref59","author":"torabi","year":"2015","journal-title":"Using descriptive video services to create a large data source for video annotation research"},{"key":"ref58","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.571"},{"key":"ref57","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.503"},{"key":"ref56","author":"long","year":"2016","journal-title":"Video captioning with multi-faceted attention"},{"key":"ref55","first-page":"2234","article-title":"Improved techniques for training GANs","author":"salimans","year":"2016","journal-title":"Proc NIPS"},{"key":"ref54","author":"zhang","year":"2015","journal-title":"Text understanding from scratch"},{"key":"ref53","first-page":"2267","article-title":"Recurrent convolutional neural networks for text classification","volume":"333","author":"lai","year":"2015","journal-title":"Proc AAAI"},{"key":"ref52","doi-asserted-by":"crossref","first-page":"2345","DOI":"10.21437\/Interspeech.2013-548","article-title":"Sequence-discriminative training of deep neural networks","author":"vesel\u00fd","year":"2013","journal-title":"Proc INTERSPEECH"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.497"},{"key":"ref11","author":"venugopalan","year":"2014","journal-title":"Translating videos to natural language using deep recurrent neural networks"},{"key":"ref40","article-title":"Long short-term memory in recurrent neural networks","author":"gers","year":"2001"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1145\/2733373.2806314"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2017.2717185"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/TBDATA.2016.2516024"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2017.2701825"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2016.2601260"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2017.2699863"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2017.02.034"},{"key":"ref19","first-page":"2672","article-title":"Generative adversarial nets","author":"goodfellow","year":"2014","journal-title":"Proc NIPS"},{"key":"ref4","author":"xu","year":"2015","journal-title":"A multi-scale multiple instance video description network"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.515"},{"key":"ref6","first-page":"77","article-title":"Show, attend and tell: Neural image caption generation with visual attention","volume":"14","author":"xu","year":"2015","journal-title":"Proc ICML"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/TCYB.2014.2344015"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1145\/243199.243261"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298935"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1007\/BF00992696"},{"key":"ref9","first-page":"190","article-title":"Collecting highly parallel data for paraphrase evaluation","author":"chen","year":"2011","journal-title":"Proc ACL-HLT"},{"key":"ref46","author":"bowman","year":"2015","journal-title":"Generating sentences from a continuous space"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46454-1_36"},{"key":"ref48","author":"simonyan","year":"2014","journal-title":"Very Deep Convolutional Networks for Large-scale Image Recognition"},{"key":"ref47","first-page":"3104","article-title":"Sequence to sequence learning with neural networks","author":"sutskever","year":"2014","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1109\/5.58337"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"ref44","first-page":"1488","article-title":"Enabling dark energy science with deep generative models of galaxy images","author":"ravanbakhsh","year":"2017","journal-title":"Proc AAAI"},{"key":"ref43","author":"s\u00f8nderby","year":"2016","journal-title":"Amortised map inference for image super-resolution"}],"container-title":["IEEE Transactions on Image Processing"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/83\/8421670\/08410586.pdf?arnumber=8410586","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,7,8]],"date-time":"2024-07-08T04:44:24Z","timestamp":1720413864000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8410586\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2018,11]]},"references-count":67,"journal-issue":{"issue":"11"},"URL":"https:\/\/doi.org\/10.1109\/tip.2018.2855422","relation":{},"ISSN":["1057-7149","1941-0042"],"issn-type":[{"value":"1057-7149","type":"print"},{"value":"1941-0042","type":"electronic"}],"subject":[],"published":{"date-parts":[[2018,11]]}}}