{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,29]],"date-time":"2025-12-29T22:16:10Z","timestamp":1767046570644,"version":"3.45.0"},"reference-count":39,"publisher":"Tech Science Press","issue":"1","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["CMC"],"published-print":{"date-parts":[[2025]]},"DOI":"10.32604\/cmc.2025.062712","type":"journal-article","created":{"date-parts":[[2025,4,29]],"date-time":"2025-04-29T23:09:07Z","timestamp":1745968147000},"page":"1123-1144","source":"Crossref","is-referenced-by-count":2,"title":["Image Style Transfer for Exhibition Hall Design Based on Multimodal Semantic-Enhanced Algorithm"],"prefix":"10.32604","volume":"84","author":[{"given":"Qing","family":"Xie","sequence":"first","affiliation":[]},{"given":"Ruiyun","family":"Yu","sequence":"additional","affiliation":[]}],"member":"17807","published-online":{"date-parts":[[2025]]},"reference":[{"key":"ref1","series-title":"2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)","first-page":"2414","article-title":"Image style transfer using convolutional neural networks","author":"Gatys","year":"2016 Jun 27\u201330"},{"key":"ref2","unstructured":"Yeh MC, Tang S, Bhattad A, Forsyth DA, Yeh MC, Tang S, et al. Quantitative evaluation of style transfer. arXiv: 1804.00118. 2018."},{"key":"ref3","doi-asserted-by":"crossref","first-page":"3274","DOI":"10.1109\/TFUZZ.2024.3373792","article-title":"A novel fuzzy neural network architecture search framework for defect recognition with uncertainties","volume":"32","author":"Ma","year":"2024","journal-title":"IEEE Trans Fuzzy Syst"},{"key":"ref4","doi-asserted-by":"crossref","unstructured":"Liu S, Lin T, He D, Li F, Wang M, Li X, et al. AdaAttN: revisit attention mechanism in arbitrary neural style transfer. arXiv:2108.03647. 2021.","DOI":"10.1109\/ICCV48922.2021.00658"},{"key":"ref5","unstructured":"Luo X, Han Z, Yang L, Zhang L. Consistent style transfer. arXiv:2201.02233. 2022."},{"key":"ref6","series-title":"Proceedings of the 7th IEEE International Conference on Computer Vision","first-page":"1033","article-title":"Texture synthesis by non-parametric sampling","author":"Efros","year":"1999 Sep 20\u201327"},{"key":"ref7","doi-asserted-by":"crossref","DOI":"10.1201\/9781439864173","author":"Gooch","year":"2001","journal-title":"Non-photorealistic rendering"},{"key":"ref8","doi-asserted-by":"crossref","first-page":"2338","DOI":"10.1109\/TIP.2017.2678168","article-title":"Style transfer via texture synthesis","volume":"26","author":"Elad","year":"2017","journal-title":"IEEE Trans Image Process"},{"key":"ref9","unstructured":"Mordvintsev A, Olah C, Tyka M. Inceptionism: going deeper into neural networks. [cited 2025 Jan 1]. Available from: https:\/\/research.google\/blog\/inceptionism-going-deeper-into-neural-networks\/."},{"key":"ref10","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR)","first-page":"2479","article-title":"Combining Markov random fields and convolutional neural networks for image synthesis","author":"Li","year":"2016 Jun 27\u201330"},{"key":"ref11","unstructured":"Radford A, Metz L, Chintala S, Dinakaran R, Easom P, Zhang L, et al. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv:1511.06434. 2015."},{"key":"ref12","series-title":"Proceedings of the International Conference on Machine Learning","first-page":"214","article-title":"Wasserstein generative adversarial networks","author":"Arjovsky","year":"2017 Aug 6\u201311"},{"key":"ref13","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","first-page":"4396","article-title":"A style-based generator architecture for generative adversarial networks","author":"Karras","year":"2019 Jun 15\u201320"},{"key":"ref14","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","first-page":"1467","article-title":"Attention-aware multi-stroke style transfer","author":"Yao","year":"2019 Jun 15\u201320"},{"key":"ref15","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","first-page":"11316","article-title":"StyTr2: image style transfer with transformers","author":"Deng","year":"2022 Jun 18\u201324"},{"key":"ref16","unstructured":"Delbracio M, Milanfar P. Inversion by direct iteration: an alternative to denoising diffusion for image restoration. arXiv:2303.11435. 2023."},{"key":"ref17","first-page":"6840","article-title":"Denoising diffusion probabilistic models","volume":"33","author":"Ho","year":"2020","journal-title":"Adv Neural Inf Process Syst (NeurIPS)"},{"key":"ref18","doi-asserted-by":"crossref","unstructured":"Rombach R, Blattmann A, Lorenz D, Esser P, Ommer B, Mantri KSI, et al. High-resolution image synthesis with latent diffusion models. arXiv:2112.10752. 2021.","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"ref19","unstructured":"Zhang X, Song Y, Song T, Yang D, Ye Y, Zhou J, et al. AKConv: convolutional kernel with arbitrary sampled shapes and arbitrary number of parameters. arXiv:2311.11587. 2023."},{"key":"ref20","unstructured":"Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, et al. Attention is all you need. arXiv:1706.03762. 2017."},{"key":"ref21","series-title":"International Conference on Learning Representations","article-title":"SeaFormer: squeeze-enhanced axial transformer for mobile semantic segmentation","author":"Wan","year":"2023 May 1\u20135"},{"key":"ref22","series-title":"IEEE International Conference on Computer Vision","article-title":"Swin transformer: hierarchical vision transformer using shifted windows","author":"Liu","year":"2021 Oct 10\u201317"},{"key":"ref23","series-title":"IEEE International Conference on Computer Vision","article-title":"CCNet: crisscross attention for semantic segmentation","author":"Huang","year":"2019 Oct 27\u2013Nov 2"},{"key":"ref24","unstructured":"Ho J, Kalchbrenner N, Weissenborn D, Salimans T. Axial attention in multidimensional transformers. arXiv:1912.12180. 2019."},{"key":"ref25","unstructured":"Summer2Winter Yosemite: CycleGAN's Summer Winter Images Dataset at Yosemite.[cited 2025 Jan 1].Available from: https:\/\/www.kaggle.com\/datasets\/balraj98\/summer2winter-yosemite."},{"key":"ref26","unstructured":"The Cityscapes Dataset: 5 000 images with high quality annotations \u2022 20 000 images with coarse annotations \u2022 50 different cities.[cited 2025 Jan 1].Available from: https:\/\/www.cityscapes-dataset.com\/."},{"key":"ref27","unstructured":"Larger Google Sat2Map dataset.[cited 2025 Jan 1].Available from: https:\/\/github.com\/taesungp\/larger-google-sat2maps-dataset."},{"key":"ref28","unstructured":"Heusel M, Ramsauer H, Unterthiner T, Nessler B, Hochreiter S. GANs trained by a two time-scale update rule converge to a local Nash equilibrium. arXiv:1706.08500. 2017."},{"key":"ref29","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","first-page":"8165","article-title":"Reusing discriminators for encoding: towards unsupervised image-to-image translation","author":"Chen","year":"2020 Jun 13\u201319"},{"key":"ref30","series-title":"International Conference on Learning Representations","article-title":"Neural optimal transport","author":"Korotin","year":"2023 May 1\u20135"},{"key":"ref31","series-title":"Proceedings of the IEEE International Conference on Computer Vision (ICCV)","article-title":"Unpaired image-to-image translation using cycle-consistent adversarial networks","author":"Zhu","year":"2017 Oct 22\u201329"},{"key":"ref32","series-title":"Proceedings of the European Conference on Computer Vision (ECCV)","article-title":"Multimodal unsupervised image-to-image translation","author":"Huang","year":"2018 Sep 8\u201314"},{"key":"ref33","unstructured":"Benaim S, Wolf L. One-sided unsupervised domain mapping. arXiv:1706.00826. 2017."},{"key":"ref34","series-title":"2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","first-page":"2422","article-title":"Geometry-consistent generative adversarial networks for one-sided unsupervised domain mapping","author":"Fu","year":"2019 Jun 15\u201320"},{"key":"ref35","series-title":"European Conference on Computer Vision","first-page":"319","article-title":"Contrastive learning for unpaired image-to-image translation","author":"Park","year":"2020 Aug 23\u201328"},{"key":"ref36","series-title":"International Conference on Learning Representations","article-title":"SDEdit: guided image synthesis and editing with stochastic differential equations","author":"Meng","year":"2022 Apr 25\u201329"},{"key":"ref37","unstructured":"Hertz A, Mokady R, Tenenbaum J, Aberman K, Pritch Y, Cohen-Or D. Prompt-to-prompt image editing with cross attention control. arXiv:2208.01626. 2022."},{"key":"ref38","series-title":"International Conference on Learning Representations","article-title":"Unpaired image-to-image translation via neural schr\u00f6dinger bridge","author":"Kim","year":"2024 May 7\u201311"},{"key":"ref39","unstructured":"Katherine C. Clip-guided diffusion. [cited 2025 Jan 1]. Available from: https:\/\/github.com\/afiaka87\/clip-guided-diffusion."}],"container-title":["Computers, Materials &amp; Continua"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/cdn.techscience.cn\/files\/cmc\/2025\/TSP_CMC-84-1\/TSP_CMC_62712\/TSP_CMC_62712.pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,17]],"date-time":"2025-11-17T01:35:27Z","timestamp":1763343327000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.techscience.com\/cmc\/v84n1\/61717"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"references-count":39,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2025]]},"published-print":{"date-parts":[[2025]]}},"URL":"https:\/\/doi.org\/10.32604\/cmc.2025.062712","relation":{},"ISSN":["1546-2226"],"issn-type":[{"type":"electronic","value":"1546-2226"}],"subject":[],"published":{"date-parts":[[2025]]}}}