{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T16:20:00Z","timestamp":1775578800986,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":38,"publisher":"ACM","license":[{"start":{"date-parts":[[2019,10,15]],"date-time":"2019-10-15T00:00:00Z","timestamp":1571097600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"National Natural Science Foundation of China","award":["61572451"],"award-info":[{"award-number":["61572451"]}]},{"name":"National Natural Science Foundation of China","award":["61872329"],"award-info":[{"award-number":["61872329"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2019,10,15]]},"DOI":"10.1145\/3343031.3350937","type":"proceedings-article","created":{"date-parts":[[2019,10,21]],"date-time":"2019-10-21T16:32:26Z","timestamp":1571675546000},"page":"647-655","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":68,"title":["Mocycle-GAN"],"prefix":"10.1145","author":[{"given":"Yang","family":"Chen","sequence":"first","affiliation":[{"name":"University of Science and Technology of China &amp; JD AI Research, Hefei, China"}]},{"given":"Yingwei","family":"Pan","sequence":"additional","affiliation":[{"name":"University of Science and Technology of China &amp; JD AI Research, Beijing, China"}]},{"given":"Ting","family":"Yao","sequence":"additional","affiliation":[{"name":"University of Science and Technology of China &amp; JD AI Research, Beijing, China"}]},{"given":"Xinmei","family":"Tian","sequence":"additional","affiliation":[{"name":"University of Science and Technology of China &amp; JD AI Research, Hefei, China"}]},{"given":"Tao","family":"Mei","sequence":"additional","affiliation":[{"name":"University of Science and Technology of China &amp; JD AI Research, Beijing, China"}]}],"member":"320","published-online":{"date-parts":[[2019,10,15]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Deepmovie: Using optical flow and deep neural networks to stylize movies. arXiv preprint arXiv:1605.08153","author":"Anderson Alexander G","year":"2016","unstructured":"Alexander G Anderson , Cory P Berg , Daniel P Mossing , and Bruno A Olshausen . 2016 . Deepmovie: Using optical flow and deep neural networks to stylize movies. arXiv preprint arXiv:1605.08153 (2016). Alexander G Anderson, Cory P Berg, Daniel P Mossing, and Bruno A Olshausen. 2016. Deepmovie: Using optical flow and deep neural networks to stylize movies. arXiv preprint arXiv:1605.08153 (2016)."},{"key":"e_1_3_2_1_2_1","volume-title":"Recycle-gan: Unsupervised video retargeting. In ECCV .","author":"Bansal Aayush","year":"2018","unstructured":"Aayush Bansal , Shugao Ma , Deva Ramanan , and Yaser Sheikh . 2018 . Recycle-gan: Unsupervised video retargeting. In ECCV . Aayush Bansal, Shugao Ma, Deva Ramanan, and Yaser Sheikh. 2018. Recycle-gan: Unsupervised video retargeting. In ECCV ."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"crossref","unstructured":"Dongdong Chen Jing Liao Lu Yuan Nenghai Yu and Gang Hua. 2017. Coherent online video style transfer. In ICCV .  Dongdong Chen Jing Liao Lu Yuan Nenghai Yu and Gang Hua. 2017. Coherent online video style transfer. In ICCV .","DOI":"10.1109\/ICCV.2017.126"},{"key":"e_1_3_2_1_4_1","volume-title":"Stargan: Unified generative adversarial networks for multi-domain image-to-image translation. In CVPR .","author":"Choi Yunjey","year":"2018","unstructured":"Yunjey Choi , Minje Choi , Munyoung Kim , Jung-Woo Ha , Sunghun Kim , and Jaegul Choo . 2018 . Stargan: Unified generative adversarial networks for multi-domain image-to-image translation. In CVPR . Yunjey Choi, Minje Choi, Munyoung Kim, Jung-Woo Ha, Sunghun Kim, and Jaegul Choo. 2018. Stargan: Unified generative adversarial networks for multi-domain image-to-image translation. In CVPR ."},{"key":"e_1_3_2_1_5_1","volume-title":"ReCoNet: Real-time Coherent Video Style Transfer Network. arXiv preprint arXiv:1807.01197","author":"Gao Chang","year":"2018","unstructured":"Chang Gao , Derun Gu , Fangjun Zhang , and Yizhou Yu. 2018. ReCoNet: Real-time Coherent Video Style Transfer Network. arXiv preprint arXiv:1807.01197 ( 2018 ). Chang Gao, Derun Gu, Fangjun Zhang, and Yizhou Yu. 2018. ReCoNet: Real-time Coherent Video Style Transfer Network. arXiv preprint arXiv:1807.01197 (2018)."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"crossref","unstructured":"Leon A Gatys Alexander S Ecker and Matthias Bethge. 2016. Image style transfer using convolutional neural networks. In CVPR .  Leon A Gatys Alexander S Ecker and Matthias Bethge. 2016. Image style transfer using convolutional neural networks. In CVPR .","DOI":"10.1109\/CVPR.2016.265"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"crossref","unstructured":"Golnaz Ghiasi Honglak Lee Manjunath Kudlur Vincent Dumoulin and Jonathon Shlens. 2017. Exploring the structure of a real-time arbitrary neural artistic stylization network. In BMVC .  Golnaz Ghiasi Honglak Lee Manjunath Kudlur Vincent Dumoulin and Jonathon Shlens. 2017. Exploring the structure of a real-time arbitrary neural artistic stylization network. In BMVC .","DOI":"10.5244\/C.31.114"},{"key":"e_1_3_2_1_8_1","unstructured":"Ian Goodfellow Jean Pouget-Abadie Mehdi Mirza Bing Xu David Warde-Farley Sherjil Ozair Aaron Courville and Yoshua Bengio. 2014. Generative adversarial nets. In NIPS .  Ian Goodfellow Jean Pouget-Abadie Mehdi Mirza Bing Xu David Warde-Farley Sherjil Ozair Aaron Courville and Yoshua Bengio. 2014. Generative adversarial nets. In NIPS ."},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"crossref","unstructured":"Agrim Gupta Justin Johnson Alexandre Alahi and Li Fei-Fei. 2017. Characterizing and improving stability in neural style transfer. In ICCV .  Agrim Gupta Justin Johnson Alexandre Alahi and Li Fei-Fei. 2017. Characterizing and improving stability in neural style transfer. In ICCV .","DOI":"10.1109\/ICCV.2017.438"},{"key":"e_1_3_2_1_10_1","unstructured":"Kaiming He Xiangyu Zhang Shaoqing Ren and Jian Sun. 2016. Deep residual learning for image recognition. In CVPR .  Kaiming He Xiangyu Zhang Shaoqing Ren and Jian Sun. 2016. Deep residual learning for image recognition. In CVPR ."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"crossref","unstructured":"Haozhi Huang Hao Wang Wenhan Luo Lin Ma Wenhao Jiang Xiaolong Zhu Zhifeng Li and Wei Liu. 2017. Real-time neural style transfer for videos. In CVPR .  Haozhi Huang Hao Wang Wenhan Luo Lin Ma Wenhao Jiang Xiaolong Zhu Zhifeng Li and Wei Liu. 2017. Real-time neural style transfer for videos. In CVPR .","DOI":"10.1109\/CVPR.2017.745"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"crossref","unstructured":"Eddy Ilg Nikolaus Mayer Tonmoy Saikia Margret Keuper Alexey Dosovitskiy and Thomas Brox. 2017. Flownet 2.0: Evolution of optical flow estimation with deep networks. In CVPR .  Eddy Ilg Nikolaus Mayer Tonmoy Saikia Margret Keuper Alexey Dosovitskiy and Thomas Brox. 2017. Flownet 2.0: Evolution of optical flow estimation with deep networks. In CVPR .","DOI":"10.1109\/CVPR.2017.179"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"crossref","unstructured":"Phillip Isola Jun-Yan Zhu Tinghui Zhou and Alexei A Efros. 2017. Image-to-image translation with conditional adversarial networks. In CVPR .  Phillip Isola Jun-Yan Zhu Tinghui Zhou and Alexei A Efros. 2017. Image-to-image translation with conditional adversarial networks. In CVPR .","DOI":"10.1109\/CVPR.2017.632"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"crossref","unstructured":"Justin Johnson Alexandre Alahi and Li Fei-Fei. 2016. Perceptual losses for real-time style transfer and super-resolution. In ECCV .  Justin Johnson Alexandre Alahi and Li Fei-Fei. 2016. Perceptual losses for real-time style transfer and super-resolution. In ECCV .","DOI":"10.1007\/978-3-319-46475-6_43"},{"key":"e_1_3_2_1_15_1","volume-title":"Jung Kwon Lee, and Jiwon Kim","author":"Kim Taeksoo","year":"2017","unstructured":"Taeksoo Kim , Moonsu Cha , Hyunsoo Kim , Jung Kwon Lee, and Jiwon Kim . 2017 . Learning to discover cross-domain relations with generative adversarial networks. In ICML . Taeksoo Kim, Moonsu Cha, Hyunsoo Kim, Jung Kwon Lee, and Jiwon Kim. 2017. Learning to discover cross-domain relations with generative adversarial networks. In ICML ."},{"key":"e_1_3_2_1_16_1","volume-title":"Adam: A method for stochastic optimization. In ICLR .","author":"Kingma Diederik P","year":"2014","unstructured":"Diederik P Kingma and Jimmy Ba . 2014 . Adam: A method for stochastic optimization. In ICLR . Diederik P Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. In ICLR ."},{"key":"e_1_3_2_1_17_1","unstructured":"Yehao Li Ting Yao Yingwei Pan Hongyang Chao and Tao Mei. 2018. Jointly localizing and describing events for dense video captioning. In CVPR .  Yehao Li Ting Yao Yingwei Pan Hongyang Chao and Tao Mei. 2018. Jointly localizing and describing events for dense video captioning. In CVPR ."},{"key":"e_1_3_2_1_18_1","unstructured":"Ming-Yu Liu and Oncel Tuzel. 2016. Coupled generative adversarial networks. In NIPS .  Ming-Yu Liu and Oncel Tuzel. 2016. Coupled generative adversarial networks. In NIPS ."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"crossref","unstructured":"Jonathan Long Evan Shelhamer and Trevor Darrell. 2015. Fully convolutional networks for semantic segmentation. In CVPR .  Jonathan Long Evan Shelhamer and Trevor Darrell. 2015. Fully convolutional networks for semantic segmentation. In CVPR .","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"e_1_3_2_1_20_1","unstructured":"Youssef Alami Mejjati Christian Richardt James Tompkin Darren Cosker and Kwang In Kim. 2018. Unsupervised Attention-guided Image-to-Image Translation. In NIPS .  Youssef Alami Mejjati Christian Richardt James Tompkin Darren Cosker and Kwang In Kim. 2018. Unsupervised Attention-guided Image-to-Image Translation. In NIPS ."},{"key":"e_1_3_2_1_21_1","unstructured":"Yingwei Pan Yehao Li Ting Yao Tao Mei Houqiang Li and Yong Rui. 2016a. Learning Deep Intrinsic Video Representation by Exploring Temporal Coherence and Graph Structure. In IJCAI .  Yingwei Pan Yehao Li Ting Yao Tao Mei Houqiang Li and Yong Rui. 2016a. Learning Deep Intrinsic Video Representation by Exploring Temporal Coherence and Graph Structure. In IJCAI ."},{"key":"e_1_3_2_1_22_1","unstructured":"Yingwei Pan Tao Mei Ting Yao Houqiang Li and Yong Rui. 2016b. Jointly modeling embedding and translation to bridge video and language. In CVPR .  Yingwei Pan Tao Mei Ting Yao Houqiang Li and Yong Rui. 2016b. Jointly modeling embedding and translation to bridge video and language. In CVPR ."},{"key":"e_1_3_2_1_23_1","unstructured":"Yingwei Pan Zhaofan Qiu Ting Yao Houqiang Li and Tao Mei. 2017a. To create what you tell: Generating videos from captions. In ACM MM .  Yingwei Pan Zhaofan Qiu Ting Yao Houqiang Li and Tao Mei. 2017a. To create what you tell: Generating videos from captions. In ACM MM ."},{"key":"e_1_3_2_1_24_1","unstructured":"Yingwei Pan Ting Yao Houqiang Li and Tao Mei. 2017b. Video captioning with transferred semantic attributes. In CVPR .  Yingwei Pan Ting Yao Houqiang Li and Tao Mei. 2017b. Video captioning with transferred semantic attributes. In CVPR ."},{"key":"e_1_3_2_1_25_1","volume-title":"NIPS Workshop .","author":"Paszke Adam","year":"2017","unstructured":"Adam Paszke , Sam Gross , Soumith Chintala , Gregory Chanan , Edward Yang , 2017 . Automatic differentiation in PyTorch . In NIPS Workshop . Adam Paszke, Sam Gross, Soumith Chintala, Gregory Chanan, Edward Yang, et almbox. 2017. Automatic differentiation in PyTorch. In NIPS Workshop ."},{"key":"e_1_3_2_1_26_1","unstructured":"Zhaofan Qiu Yingwei Pan Ting Yao and Tao Mei. 2017. Deep semantic hashing with generative adversarial networks. In SIGIR .  Zhaofan Qiu Yingwei Pan Ting Yao and Tao Mei. 2017. Deep semantic hashing with generative adversarial networks. In SIGIR ."},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"crossref","unstructured":"Stephan R Richter Zeeshan Hayder and Vladlen Koltun. 2017. Playing for benchmarks. In ICCV .  Stephan R Richter Zeeshan Hayder and Vladlen Koltun. 2017. Playing for benchmarks. In ICCV .","DOI":"10.1109\/ICCV.2017.243"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"crossref","unstructured":"Manuel Ruder Alexey Dosovitskiy and Thomas Brox. 2016. Artistic style transfer for videos. In GCPR .  Manuel Ruder Alexey Dosovitskiy and Thomas Brox. 2016. Artistic style transfer for videos. In GCPR .","DOI":"10.1007\/978-3-319-45886-1_3"},{"key":"e_1_3_2_1_29_1","volume-title":"Texture Networks: Feed-forward Synthesis of Textures and Stylized Images.. In ICML .","author":"Ulyanov Dmitry","year":"2016","unstructured":"Dmitry Ulyanov , Vadim Lebedev , Andrea Vedaldi , and Victor S Lempitsky . 2016 . Texture Networks: Feed-forward Synthesis of Textures and Stylized Images.. In ICML . Dmitry Ulyanov, Vadim Lebedev, Andrea Vedaldi, and Victor S Lempitsky. 2016. Texture Networks: Feed-forward Synthesis of Textures and Stylized Images.. In ICML ."},{"key":"e_1_3_2_1_30_1","unstructured":"Carl Vondrick Hamed Pirsiavash and Antonio Torralba. 2016. Generating videos with scene dynamics. In NIPS .  Carl Vondrick Hamed Pirsiavash and Antonio Torralba. 2016. Generating videos with scene dynamics. In NIPS ."},{"key":"e_1_3_2_1_31_1","unstructured":"Ting-Chun Wang Ming-Yu Liu Jun-Yan Zhu Guilin Liu Andrew Tao Jan Kautz and Bryan Catanzaro. 2018a. Video-to-video synthesis. In NIPS .  Ting-Chun Wang Ming-Yu Liu Jun-Yan Zhu Guilin Liu Andrew Tao Jan Kautz and Bryan Catanzaro. 2018a. Video-to-video synthesis. In NIPS ."},{"key":"e_1_3_2_1_32_1","unstructured":"Ting-Chun Wang Ming-Yu Liu Jun-Yan Zhu Andrew Tao Jan Kautz and Bryan Catanzaro. 2018b. High-resolution image synthesis and semantic manipulation with conditional gans. In CVPR .  Ting-Chun Wang Ming-Yu Liu Jun-Yan Zhu Andrew Tao Jan Kautz and Bryan Catanzaro. 2018b. High-resolution image synthesis and semantic manipulation with conditional gans. In CVPR ."},{"key":"e_1_3_2_1_33_1","unstructured":"Xingxing Wei Jun Zhu Sitong Feng and Hang Su. 2018. Video-to-video translation with global temporal consistency. In ACM MM .  Xingxing Wei Jun Zhu Sitong Feng and Hang Su. 2018. Video-to-video translation with global temporal consistency. In ACM MM ."},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"crossref","unstructured":"Xuewen Yang Dongliang Xie and Xin Wang. 2018. Crossing-Domain Generative Adversarial Networks for Unsupervised Multi-Domain Image-to-Image Translation. In ACM MM .  Xuewen Yang Dongliang Xie and Xin Wang. 2018. Crossing-Domain Generative Adversarial Networks for Unsupervised Multi-Domain Image-to-Image Translation. In ACM MM .","DOI":"10.1145\/3240508.3240716"},{"key":"e_1_3_2_1_35_1","volume-title":"Dualgan: Unsupervised dual learning for image-to-image translation. In ICCV .","author":"Yi Zili","year":"2017","unstructured":"Zili Yi , Hao Zhang , Ping Tan , and Minglun Gong . 2017 . Dualgan: Unsupervised dual learning for image-to-image translation. In ICCV . Zili Yi, Hao Zhang, Ping Tan, and Minglun Gong. 2017. Dualgan: Unsupervised dual learning for image-to-image translation. In ICCV ."},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"crossref","unstructured":"Hang Zhang and Kristin Dana. 2018. Multi-style generative network for real-time transfer. In ECCV .  Hang Zhang and Kristin Dana. 2018. Multi-style generative network for real-time transfer. In ECCV .","DOI":"10.1007\/978-3-030-11018-5_32"},{"key":"e_1_3_2_1_37_1","unstructured":"Jun-Yan Zhu Taesung Park Phillip Isola and Alexei A Efros. 2017a. Unpaired image-to-image translation using cycle-consistent adversarial networks. In ICCV .  Jun-Yan Zhu Taesung Park Phillip Isola and Alexei A Efros. 2017a. Unpaired image-to-image translation using cycle-consistent adversarial networks. In ICCV ."},{"key":"e_1_3_2_1_38_1","unstructured":"Jun-Yan Zhu Richard Zhang Deepak Pathak Trevor Darrell Alexei A Efros Oliver Wang and Eli Shechtman. 2017b. Toward multimodal image-to-image translation. In NIPS .  Jun-Yan Zhu Richard Zhang Deepak Pathak Trevor Darrell Alexei A Efros Oliver Wang and Eli Shechtman. 2017b. Toward multimodal image-to-image translation. In NIPS ."}],"event":{"name":"MM '19: The 27th ACM International Conference on Multimedia","location":"Nice France","acronym":"MM '19","sponsor":["SIGMM ACM Special Interest Group on Multimedia"]},"container-title":["Proceedings of the 27th ACM International Conference on Multimedia"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3343031.3350937","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3343031.3350937","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T23:13:17Z","timestamp":1750201997000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3343031.3350937"}},"subtitle":["Unpaired Video-to-Video Translation"],"short-title":[],"issued":{"date-parts":[[2019,10,15]]},"references-count":38,"alternative-id":["10.1145\/3343031.3350937","10.1145\/3343031"],"URL":"https:\/\/doi.org\/10.1145\/3343031.3350937","relation":{},"subject":[],"published":{"date-parts":[[2019,10,15]]},"assertion":[{"value":"2019-10-15","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}