{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T16:04:19Z","timestamp":1743091459870,"version":"3.40.3"},"publisher-location":"Cham","reference-count":30,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783031138690"},{"type":"electronic","value":"9783031138706"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-13870-6_36","type":"book-chapter","created":{"date-parts":[[2022,8,14]],"date-time":"2022-08-14T09:03:13Z","timestamp":1660467793000},"page":"430-443","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Two-Channel VAE-GAN Based Image-To-Video Translation"],"prefix":"10.1007","author":[{"given":"Shengli","family":"Wang","sequence":"first","affiliation":[]},{"given":"Mulin","family":"Xieshi","sequence":"additional","affiliation":[]},{"given":"Zhangpeng","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Xiang","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Xujie","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Zeyi","family":"Tang","sequence":"additional","affiliation":[]},{"given":"Yuxing","family":"Dai","sequence":"additional","affiliation":[]},{"given":"Xuexin","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Pingyuan","family":"Lin","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,8,15]]},"reference":[{"key":"36_CR1","unstructured":"Aifanti, N., Papachristou, C., Delopoulos, A.: The mug facial expression database. In: 11th International Workshop on Image Analysis for Multimedia Interactive Services WIAMIS 10, pp. 1\u20134 (2010)"},{"key":"36_CR2","unstructured":"Arjovsky, M., Bottou, L.: Towards principled methods for training generative adversarial networks. In: 5th International Conference on Learning Representations, ICLR, Toulon, France, 24\u201326 April 2017 (2017)"},{"key":"36_CR3","unstructured":"Babaeizadeh, M., Finn, C., Erhan, D., Campbell, R.H., Levine, S.: Stochastic variational video prediction. In: 6th International Conference on Learning Representations, ICLR 2018 (2018)"},{"key":"36_CR4","doi-asserted-by":"crossref","unstructured":"Baltrusaitis, T., Robinson, P., Morency, L.: Openface: an open source facial behavior analysis toolkit. In: 2016 IEEE Winter Conference on Applications of Compute Vision, WACV, Lake Placid, NY, USA, 7\u201310 March 2016, pp. 1\u201310 (2016)","DOI":"10.1109\/WACV.2016.7477553"},{"key":"36_CR5","doi-asserted-by":"crossref","unstructured":"Carreira, J., Zisserman, A.: Quo vadis, action recognition? A new model and the kinetics dataset. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR, Honolulu, HI, USA, 21\u201326 July 2017, pp. 4724\u20134733 (2017)","DOI":"10.1109\/CVPR.2017.502"},{"key":"36_CR6","doi-asserted-by":"crossref","unstructured":"Fan, L., Huang, W., Gan, C., Huang, J., Gong, B.: Controllable image-to-video translation: a case study on facial expression generation. In: The Thirty-Third AAAI Conference on Artificial Intelligence, AAAI 2019, pp. 3510\u20133517 (2019)","DOI":"10.1609\/aaai.v33i01.33013510"},{"key":"36_CR7","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. In: Advances in Neural Information Processing Systems, vol. 27, pp. 2672\u20132680. Curran Associates, Inc. (2014)"},{"key":"36_CR8","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: GANs trained by a two time-scale update rule converge to a local nash equilibrium. In: Advances in Neural Information Processing Systems, vol. 30, pp. 6626\u20136637 (2017)"},{"key":"36_CR9","doi-asserted-by":"publisher","unstructured":"Johson, J., Alahi, A., Fei-Fei, L.: Perceptual losses for real-time style transfer and super-resolution. Computer Vision \u2013 ECCV 2016, pp. 694\u2013711 (2016). https:\/\/doi.org\/10.1007\/978-3-319-46475-6_43","DOI":"10.1007\/978-3-319-46475-6_43"},{"key":"36_CR10","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. In: 3rd International Conference on Learning Representations, ICLR, San Diego, CA, USA, 7\u20139 May 2015 (2015)"},{"key":"36_CR11","unstructured":"Lee, A.X., Zhang, R., Ebert, F., Abbeel, P., Finn, C., Levine, S.: Stochastic adversarial video prediction. CoRR (2018)"},{"key":"36_CR12","doi-asserted-by":"publisher","unstructured":"Li, Y., Fang, C., Yang, J., Wang, Z., Lu, X., Yang, M.: Flow-grounded spatial-temporal video prediction from still images. In: Computer Vision - ECCV 2018 - 15th European Conference, pp. 609\u2013625 (2018). https:\/\/doi.org\/10.1007\/978-3-030-01240-3_37","DOI":"10.1007\/978-3-030-01240-3_37"},{"key":"36_CR13","doi-asserted-by":"crossref","unstructured":"Li, Y., Min, M.R., Shen, D., Carlson, D.E., Carin, L.: Video generation from text. In: McIlraith, S.A., Weinberger, K.Q. (eds.) Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, (AAAI-18), pp. 7065\u20137072 (2018)","DOI":"10.1609\/aaai.v32i1.12233"},{"key":"36_CR14","doi-asserted-by":"crossref","unstructured":"Mao, X., Li, Q., Xie, H., Lau, R.Y.K., Wang, Z., Smolley, S.P.: Least squares generative adversarial networks. In: IEEE International Conference on Computer Vision, ICCV 2017, pp. 2813\u20132821 (2017)","DOI":"10.1109\/ICCV.2017.304"},{"key":"36_CR15","doi-asserted-by":"crossref","unstructured":"Nam, S., Ma, C., Chai, M., Brendel, W., Xu, N., Kim, S.J.: End-to-end time-lapse video synthesis from a single outdoor image. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR, Long Beach, CA, USA, June 16\u201320, 2019, pp. 1409\u20131418 (2019)","DOI":"10.1109\/CVPR.2019.00150"},{"key":"36_CR16","doi-asserted-by":"crossref","unstructured":"Pan, J., et al.: Video generation from single semantic label map. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, pp. 3733\u20133742 (2019)","DOI":"10.1109\/CVPR.2019.00385"},{"key":"36_CR17","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., P.Fischer, Brox, T.: U-net: convolutional networks for biomedical image segmentation. In: Medical Image Computing and Computer-Assisted Intervention (MICCAI), pp. 234\u2013241 (2015)","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"36_CR18","doi-asserted-by":"crossref","unstructured":"Saito, M., Matsumoto, E., Saito, S.: Temporal generative adversarial nets with singular value clipping. In: IEEE International Conference on Computer Vision ICCV Venice, Italy, 22\u201329 October 2017, pp. 2849\u20132858 (2017)","DOI":"10.1109\/ICCV.2017.308"},{"key":"36_CR19","unstructured":"Salimans, T., et al.: Improved techniques for training GANs. In: Advances in Neural Information Processing Systems, vol. 29, pp. 2234\u20132242 (2016)"},{"key":"36_CR20","doi-asserted-by":"crossref","unstructured":"Shen, G., et al.: Facial image-to-video translation by a hidden affine transformation. In: Proceedings of the 27th ACM International Conference on Multimedia, pp. 2505\u20132513 (2019)","DOI":"10.1145\/3343031.3350981"},{"key":"36_CR21","unstructured":"Shi, X., Chen, Z., Wang, H., Yeung, D.Y., Wong, W.K., WOO, W.C.: Convolutional LSTM network: A machine learning approach for precipitation nowcasting. In: Advances in Neural Information Processing Systems, vol. 28, pp. 802\u2013810 (2015)"},{"key":"36_CR22","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition, pp. 1\u201314. Computational and Biological Learning Society (2015)"},{"key":"36_CR23","doi-asserted-by":"crossref","unstructured":"Tulyakov, S., Liu, M., Yang, X., Kautz, J.: Mocogan: decomposing motion and content for video generation. In: 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR, Salt Lake City, UT, USA, 18\u201322 June 2018, pp. 1526\u20131535 (2018)","DOI":"10.1109\/CVPR.2018.00165"},{"key":"36_CR24","unstructured":"Vondrick, C., Pirsiavash, H., Torralba, A.: Generating videos with scene dynamics. In: Advances in Neural Information Processing Systems, vol. 29, pp. 613\u2013621. Curran Associates, Inc. (2016)"},{"key":"36_CR25","doi-asserted-by":"publisher","unstructured":"Walker, J., Doersch, C., Gupta, A., Hebert, M.: An uncertain future: Forecasting from static images using variational autoencoders. In: Computer Vision \u2013 ECCV 2016 - 14th European Conference, pp. 835\u2013851 (2016).https:\/\/doi.org\/10.1007\/978-3-319-46478-7_51","DOI":"10.1007\/978-3-319-46478-7_51"},{"key":"36_CR26","unstructured":"Wang, T.C., et al.: Video- to-video synthesis. In: Advances in Neural Information Processing Systems, vol. 31, pp. 1144\u20131156. Curran Associates, Inc. (2018)"},{"key":"36_CR27","doi-asserted-by":"crossref","unstructured":"Wang, T., Cheng, Y., Lin, C.H., Chen, H., Sun, M.: Point-to-point video generation. In: 2019 IEEE\/CVF International Conference on Computer Vision, ICCV, Seoul, Korea (South), 27 October\u20132 November 2019, pp. 10490\u201310499 (2019)","DOI":"10.1109\/ICCV.2019.01059"},{"key":"36_CR28","doi-asserted-by":"crossref","unstructured":"Xie, S., Girshick, R.B., Doll\u00e1r, P., Tu, Z., He, K.: Aggregated residual transformations for deep neural networks. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR, Honolulu, HI, USA, 21\u201326 July 2017, pp. 5987\u20135995 (2017)","DOI":"10.1109\/CVPR.2017.634"},{"key":"36_CR29","unstructured":"Xue, T., Wu, J., Bouman, K., Freeman, B.: Visual dynamics: Probabilistic future frame synthesis via cross convolutional networks. In: Advances in Neural Information Processing Systems, vol. 29, pp. 91\u201399. Curran Associates, Inc. (2016)"},{"key":"36_CR30","doi-asserted-by":"crossref","unstructured":"Zhang, C., Peng, Y.: Stacking VAE and GAN for context-aware text-to-image generation. In: Fourth IEEE International Conference on Multimedia Big Data, BigMM, Xi\u2019an, China, 13\u201316 September 2018, pp. 1\u20135 (2018)","DOI":"10.1109\/BigMM.2018.8499439"}],"container-title":["Lecture Notes in Computer Science","Intelligent Computing Theories and Application"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-13870-6_36","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,7]],"date-time":"2024-03-07T16:57:47Z","timestamp":1709830667000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-13870-6_36"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031138690","9783031138706"],"references-count":30,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-13870-6_36","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"15 August 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICIC","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Intelligent Computing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Xi'an","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"7 August 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"11 August 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icic2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.ic-icc.cn\/2022\/index.htm","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Open","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"IC-ICC-CN","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"449","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"209","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"47% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}