{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T16:23:27Z","timestamp":1775579007192,"version":"3.50.1"},"reference-count":72,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U20B2062"],"award-info":[{"award-number":["U20B2062"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62172036"],"award-info":[{"award-number":["62172036"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62227801"],"award-info":[{"award-number":["62227801"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"National Key Research and Development Program of China","award":["2022ZD0116300"],"award-info":[{"award-number":["2022ZD0116300"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Multimedia"],"published-print":{"date-parts":[[2023]]},"DOI":"10.1109\/tmm.2023.3251095","type":"journal-article","created":{"date-parts":[[2023,3,1]],"date-time":"2023-03-01T18:25:41Z","timestamp":1677695141000},"page":"9370-9382","source":"Crossref","is-referenced-by-count":8,"title":["MotionVideoGAN: A Novel Video Generator Based on the Motion Space Learned From Image Pairs"],"prefix":"10.1109","volume":"25","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3995-8382","authenticated-orcid":false,"given":"Jingyuan","family":"Zhu","sequence":"first","affiliation":[{"name":"Department of Electronic Engineering, Tsinghua University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5383-5667","authenticated-orcid":false,"given":"Huimin","family":"Ma","sequence":"additional","affiliation":[{"name":"School of Computer and Communication Engineering, University of Science and Technology, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2040-7938","authenticated-orcid":false,"given":"Jiansheng","family":"Chen","sequence":"additional","affiliation":[{"name":"School of Computer and Communication Engineering, University of Science and Technology, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9734-6056","authenticated-orcid":false,"given":"Jian","family":"Yuan","sequence":"additional","affiliation":[{"name":"Department of Electronic Engineering, Tsinghua University, Beijing, China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00453"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"ref3","first-page":"852","article-title":"Alias-free generative adversarial networks","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Karras","year":"2021"},{"key":"ref4","first-page":"1","article-title":"Progressive growing of GANs for improved quality, stability, and variation","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Karras","year":"2018"},{"key":"ref5","first-page":"7354","article-title":"Self-attention generative adversarial networks","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Zhang","year":"2019"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/tmm.2022.3147425"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2021.3109419"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.632"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01246-5_3"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.244"},{"key":"ref11","article-title":"Exemplar guided unsupervised image-to-image translation with semantic consistency","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Ma","year":"2019"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2021.3121987"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/ICAECT54875.2022.9807935"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/DSN-W.2019.00012"},{"key":"ref15","first-page":"1","article-title":"StyleNeRF: A style-based 3D aware generator for high-resolution image synthesis","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Gu","year":"2022"},{"key":"ref16","first-page":"20154","article-title":"GRAF: Generative radiance fields for 3D-aware image synthesis","volume-title":"Proc. Proc. Adv. Neural Inf. Process. Syst.","author":"Schwarz","year":"2020"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01565"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00574"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1145\/3503927"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW.2019.00190"},{"key":"ref21","first-page":"1","article-title":"Generating videos with scene dynamics","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"29","author":"Vondrick","year":"2016"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00165"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-020-01333-y"},{"key":"ref24","first-page":"1","article-title":"Generative adversarial nets","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"27","author":"Goodfellow","year":"2014"},{"key":"ref25","first-page":"1","article-title":"Large scale GAN training for high fidelity natural image synthesis","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Brock","year":"2019"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00361"},{"key":"ref27","article-title":"Faceforensics: A large-scale video dataset for forgery detection in human faces","author":"Rssler","year":"2018"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00251"},{"key":"ref29","first-page":"1","article-title":"A good image generator is what you need for high-resolution video synthesis","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Tian","year":"2021"},{"key":"ref30","first-page":"16648","article-title":"Low-rank subspaces in GANs","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Zhu","year":"2021"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"ref32","article-title":"Ucf101: A dataset of 101 human actions classes from videos in the wild","author":"Soomro","year":"2012"},{"key":"ref33","article-title":"Towards accurate generative models of video: A new metric and challenges","author":"Unterthiner","year":"2018"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.308"},{"key":"ref35","first-page":"1","article-title":"Unsupervised representation learning with deep convolutional generative adversarial networks","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Radford","year":"2016"},{"key":"ref36","first-page":"1","article-title":"AdaGAN: Boosting generative models","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"30","author":"Tolstikhin","year":"2017"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00332"},{"key":"ref38","first-page":"1","article-title":"You only need adversarial supervision for semantic image synthesis","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Schnfeld","year":"2021"},{"key":"ref39","first-page":"1","article-title":"Improved techniques for training GANs","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"29","author":"Salimans","year":"2016"},{"key":"ref40","first-page":"1","article-title":"Improved training of wasserstein GANs","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"30","author":"Gulrajani","year":"2017"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1145\/3343031.3350944"},{"key":"ref42","first-page":"1","article-title":"Large scale adversarial representation learning","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"32","author":"Donahue","year":"2019"},{"key":"ref43","article-title":"Adversarial video generation on complex\ndatasets","author":"Clark","year":"2019"},{"key":"ref44","article-title":"Towards high resolution video generation with progressive growing of sliced wasserstein gans","author":"Acharya","year":"2018"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00613"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1109\/WACV48630.2021.00322"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01068"},{"key":"ref48","article-title":"VideoGPT: Video generation using VQ-VAE and transformers","author":"Yan","year":"2021"},{"key":"ref49","first-page":"1","article-title":"StylevideoGAN: A temporal generative model using a pretrained styleGAN","volume-title":"Proc. Brit. Mach. Vis. Conf.","author":"Fox","year":"2021"},{"key":"ref50","first-page":"214","article-title":"Wasserstein generative adversarial networks","volume-title":"Proc. 34th Int. Conf. Mach. Learn.","author":"Arjovsky","year":"2017"},{"key":"ref51","first-page":"1","article-title":"Generating videos with dynamics-aware implicit generative adversarial networks","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Yu","year":"2022"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2020.09.016"},{"key":"ref53","first-page":"1","article-title":"GAN dissection: Visualizing and understanding generative adversarial networks","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Bau","year":"2019"},{"key":"ref54","first-page":"9786","article-title":"Unsupervised discovery of interpretable directions in the gan latent space","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Voynov","year":"2020"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33018901"},{"key":"ref56","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00584"},{"key":"ref57","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00926"},{"key":"ref58","first-page":"1","article-title":"On the steerability of generative adversarial networks","volume-title":"Int. Conf. Learn. Representations","author":"Jahanian","year":"2020"},{"key":"ref59","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01267"},{"key":"ref60","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00453"},{"key":"ref61","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00832"},{"key":"ref62","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2020.3034267"},{"key":"ref63","first-page":"1","article-title":"Controlling generative models with continuous factors of variations","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Plumerault","year":"2020"},{"key":"ref64","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00158"},{"key":"ref65","first-page":"16331","article-title":"EditGAN: High-precision semantic image editing","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Ling","year":"2021"},{"key":"ref66","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01001"},{"key":"ref67","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00820"},{"key":"ref68","doi-asserted-by":"publisher","DOI":"10.1561\/2200000016"},{"key":"ref69","doi-asserted-by":"publisher","DOI":"10.1016\/j.jsb.2012.10.010"},{"key":"ref70","first-page":"1","article-title":"Gans trained by a two time-scale update rule converge to a local nash equilibrium","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"30","author":"Heusel","year":"2017"},{"key":"ref71","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58589-1_42"},{"key":"ref72","first-page":"12104","article-title":"Training generative adversarial networks with limited data","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Karras","year":"2020"}],"container-title":["IEEE Transactions on Multimedia"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/6046\/10016790\/10057062.pdf?arnumber=10057062","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,11]],"date-time":"2024-01-11T23:45:23Z","timestamp":1705016723000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10057062\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"references-count":72,"URL":"https:\/\/doi.org\/10.1109\/tmm.2023.3251095","relation":{},"ISSN":["1520-9210","1941-0077"],"issn-type":[{"value":"1520-9210","type":"print"},{"value":"1941-0077","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023]]}}}