{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T00:33:37Z","timestamp":1769042017210,"version":"3.49.0"},"reference-count":30,"publisher":"Springer Science and Business Media LLC","issue":"11","license":[{"start":{"date-parts":[[2022,10,12]],"date-time":"2022-10-12T00:00:00Z","timestamp":1665532800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2022,10,12]],"date-time":"2022-10-12T00:00:00Z","timestamp":1665532800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62102320"],"award-info":[{"award-number":["62102320"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"the Fundamental Research Funds for the Central Universities","award":["No. D5000210737"],"award-info":[{"award-number":["No. D5000210737"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"published-print":{"date-parts":[[2023,5]]},"DOI":"10.1007\/s11042-022-14040-2","type":"journal-article","created":{"date-parts":[[2022,10,12]],"date-time":"2022-10-12T11:23:39Z","timestamp":1665573819000},"page":"16057-16079","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["Controllable smoke image generation network based on smoke imaging principle"],"prefix":"10.1007","volume":"82","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-1453-2468","authenticated-orcid":false,"given":"Huanjie","family":"Tao","sequence":"first","affiliation":[]},{"given":"Jing","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Zhouxin","family":"Xin","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,12]]},"reference":[{"key":"14040_CR1","unstructured":"Berthelot D, Schumm T, Metz L (2017) BEGAN: boundary equilibrium generative adversarial networks[J]. arXiv preprint, arXiv:1703.10717"},{"key":"14040_CR2","unstructured":"Bounliphone W, Belilovsky E, Blaschko MB, Antonoglou I, Gretton A (2015) A test of relative similarity for model selection in generative models[J]. arXiv preprint, arXiv:1511.04581"},{"key":"14040_CR3","unstructured":"Chen X, Duan Y, Houthooft R, Schulman J, Sutskever I, Abbeel P (2017) InfoGAN: interpretable representation learning by information maximizing generative adversarial Nets. In Proceedings of the annual conference on neural information processing systems (NIPS), 2180\u20132188"},{"key":"14040_CR4","doi-asserted-by":"publisher","first-page":"3041","DOI":"10.1109\/TIP.2021.3055936","volume":"30","author":"X Chen","year":"2021","unstructured":"Chen X, Luo X, Weng J, Luo W, Li H, Tian Q (2021) Multi-view gait image generation for cross-view gait recognition. IEEE Trans Image Process 30:3041\u20133055","journal-title":"IEEE Trans Image Process"},{"key":"14040_CR5","doi-asserted-by":"crossref","unstructured":"Choi Y, Choi M, Kim M, Ha JW, Kim S, Choo J (2018) StarGAN: unified generative adversarial networks for multi-domain image-to-image translation. In proc. IEEE CVPR, 8789\u20138797","DOI":"10.1109\/CVPR.2018.00916"},{"key":"14040_CR6","doi-asserted-by":"publisher","first-page":"101294","DOI":"10.1016\/j.jup.2021.101294","volume":"73","author":"GF Fan","year":"2021","unstructured":"Fan GF, Yu M, Dong SQ, Yeh YH, Hong WC (2021) Forecasting short-term electricity load using hybrid support vector regression with grey catastrophe and random forest modeling[J]. Util Policy 73:101294","journal-title":"Util Policy"},{"issue":"2","key":"14040_CR7","doi-asserted-by":"publisher","first-page":"725","DOI":"10.1109\/TII.2017.2757457","volume":"14","author":"A Filonenko","year":"2018","unstructured":"Filonenko A, Hern\u00e1ndez DC, Jo KH (2018) Fast smoke detection for video surveillance using CUDA[J]. IEEE Trans Ind Inf 14(2):725\u2013733","journal-title":"IEEE Trans Ind Inf"},{"key":"14040_CR8","unstructured":"Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, Bengio Y (2014) Generative adversarial nets. In proceedings of the international conference on neural information processing systems (NIPS), 2:2672\u20132680"},{"key":"14040_CR9","doi-asserted-by":"crossref","unstructured":"Gu K, Xia Z, Qiao J, Lin W (2019) Deep dual-channel neural network for image-based smoke detection. IEEE Trans Multimed 22(2):311\u2013323","DOI":"10.1109\/TMM.2019.2929009"},{"key":"14040_CR10","unstructured":"Gulrajani I, Ahmed F, Arjovsky M, Dumoulin V, Courville AC (2017) Improved training of Wasserstein GANs. In proceedings of the annual conference on neural information processing systems (NIPS), 5768\u20135778"},{"key":"14040_CR11","unstructured":"Higgins I, Matthey L, Pal A, Burgess C, Glorot X, Botvinick M, Mohamed S, Lerchner A (2017) \u03b2-VAE: learning basic visual concepts with a constrained variational framework. In proc. of the 5th international conference on learning representations (ICLR), 1\u201310"},{"key":"14040_CR12","unstructured":"Huang H, Li Z, He R, Sun Z, Tan T (2018) IntroVAE: introspective variational autoencoders for photographic image synthesis. In Proc of the conference on neural information processing systems (NIPS), 31\u201343"},{"issue":"6","key":"14040_CR13","doi-asserted-by":"publisher","first-page":"9237","DOI":"10.1109\/JIOT.2019.2896120","volume":"6","author":"S Khan","year":"2019","unstructured":"Khan S, Muhammad K, Mumtaz S, Baik SW, de Albuquerque VHC (2019) Energy-efficient deep CNN for smoke detection in foggy IoT environment. IEEE Internet Things J 6(6):9237\u20139245","journal-title":"IEEE Internet Things J"},{"key":"14040_CR14","unstructured":"Kingma DP, Welling M (2014) Auto-encoding variational bayes. arXiv preprint, arXiv:1312.6114"},{"key":"14040_CR15","unstructured":"LopezPaz D, Oquab M (2017) Revisiting classifier two-sample tests. arXiv preprint, arXiv:1610.06545"},{"key":"14040_CR16","doi-asserted-by":"crossref","unstructured":"Nguyen MD, Kim D, Ro S (2018) A video smoke detection algorithm based on cascade classification and deep learning. KSII Trans Int Inf Syst 12(12):6018\u20136033","DOI":"10.3837\/tiis.2018.12.022"},{"key":"14040_CR17","unstructured":"Radford A, Metz L, Chintala S (2016) Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint, arXiv:1511.06434"},{"key":"14040_CR18","doi-asserted-by":"crossref","unstructured":"Ronneberger O, Fischer P, Brox T (2015) U-net: convolutional networks for biomedical image segmentation. In proc. of the conference on medical image computing and computer-assisted intervention (MICCAI), 9351:234\u2013241","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"14040_CR19","doi-asserted-by":"crossref","unstructured":"Shaham TR, Dekel T, Michaeli T (2019) SinGAN: learning a generative model from a single natural image[C]. In proceedings of ICCV, 4569-4579","DOI":"10.1109\/ICCV.2019.00467"},{"issue":"12","key":"14040_CR20","doi-asserted-by":"publisher","first-page":"17785","DOI":"10.1007\/s11042-021-10601-z","volume":"80","author":"C Sheng","year":"2021","unstructured":"Sheng C, Hu B, Meng F, Yin D (2021) Lightweight dual-branch network for vehicle exhausts segmentation. Multimed Tools Appl 80(12):17785\u201317806","journal-title":"Multimed Tools Appl"},{"issue":"10","key":"14040_CR21","doi-asserted-by":"publisher","first-page":"3301","DOI":"10.1109\/TCSVT.2019.2920657","volume":"30","author":"H Tao","year":"2020","unstructured":"Tao H, Lu X (2020) Smoke vehicle detection based on spatiotemporal bag-of-features and professional convolutional neural network[J]. IEEE Trans Circuits Syst Vid Technol 30(10):3301\u20133316","journal-title":"IEEE Trans Circuits Syst Vid Technol"},{"key":"14040_CR22","doi-asserted-by":"crossref","unstructured":"Tao H, Lu M, Hu Z, Xin Z, Wang J (2022) Attention-aggregated attribute-aware network with redundancy reduction convolution for video-based industrial smoke emission recognition. IEEE Trans Industrial Inf 18(11):7653\u20137664","DOI":"10.1109\/TII.2022.3146142"},{"key":"14040_CR23","doi-asserted-by":"crossref","unstructured":"Tao H, Xie C, Wang J, Xin Z. (2022) CENet: a channel-enhanced spatiotemporal network with sufficient supervision information for recognizing industrial smoke emissions[J]. IEEE internet of things journal. 1-10","DOI":"10.1109\/JIOT.2022.3162016"},{"key":"14040_CR24","unstructured":"Tero K, Samuli L, Timo A (2019) A style-based generator architecture for generative adversarial networks. In proceedings of CVPR, 4396\u20134405"},{"key":"14040_CR25","doi-asserted-by":"publisher","first-page":"1385","DOI":"10.1007\/s00371-019-01738-y","volume":"36","author":"J Wen","year":"2020","unstructured":"Wen J, Ma H, Luo X (2020) Deep generative smoke simulator: connecting simulated and real data[J]. Vis Comput 36:1385\u20131399","journal-title":"Vis Comput"},{"key":"14040_CR26","doi-asserted-by":"publisher","first-page":"201418","DOI":"10.1109\/ACCESS.2020.3036105","volume":"8","author":"C Xie","year":"2020","unstructured":"Xie C, Tao H (2020) Generating realistic smoke images with controllable smoke components[J]. IEEE Access 8:201418\u2013201427","journal-title":"IEEE Access"},{"key":"14040_CR27","doi-asserted-by":"crossref","unstructured":"Yuan F, Zhang L, Xia X, et al. (2019) A wave-shaped deep neural network for smoke density estimation[J]. IEEE transactions on image processing, 2301-2313","DOI":"10.1109\/TIP.2019.2946126"},{"key":"14040_CR28","doi-asserted-by":"publisher","first-page":"345","DOI":"10.1007\/s00138-018-0990-3","volume":"30","author":"F Yuan","year":"2019","unstructured":"Yuan F, Zhang L, Wan B, Xia X, Shi J (2019) Convolutional neural networks based on multi-scale additive merging layers for visual smoke recognition. Mach Vis Appl 30:345\u2013358","journal-title":"Mach Vis Appl"},{"key":"14040_CR29","first-page":"441","volume":"211","author":"QX Zhang","year":"2018","unstructured":"Zhang QX, Lin GH, Zhang YM, Xu G, Wang JJ (2018) Wildland forest fire smoke detection based on faster R-CNN using synthetic smoke images. Proc Conf Fire Sci Fire Protect Eng 211:441\u2013446","journal-title":"Proc Conf Fire Sci Fire Protect Eng"},{"key":"14040_CR30","doi-asserted-by":"crossref","unstructured":"Zhu JY, Park T, Isola P, et al. (2017) Unpaired image-to-image translation using cycle-consistent adversarial networks[C]. In proceedings of ICCV, 2242-2251","DOI":"10.1109\/ICCV.2017.244"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-022-14040-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-022-14040-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-022-14040-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,4,15]],"date-time":"2023-04-15T09:18:39Z","timestamp":1681550319000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-022-14040-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,10,12]]},"references-count":30,"journal-issue":{"issue":"11","published-print":{"date-parts":[[2023,5]]}},"alternative-id":["14040"],"URL":"https:\/\/doi.org\/10.1007\/s11042-022-14040-2","relation":{},"ISSN":["1380-7501","1573-7721"],"issn-type":[{"value":"1380-7501","type":"print"},{"value":"1573-7721","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,10,12]]},"assertion":[{"value":"1 September 2021","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 May 2022","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 October 2022","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 October 2022","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"Huanjie Tao declares that he has no conflict of interest. Jing Wang declares that she has no conflict of interest. Zhouxin Xin declares that he has no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"This article does not contain any studies with human participants performed by any of the authors.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}]}}