{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,1]],"date-time":"2025-10-01T15:55:59Z","timestamp":1759334159620,"version":"build-2065373602"},"reference-count":44,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/100004358","name":"Samsung","doi-asserted-by":"publisher","award":["IO201218-08232-01"],"award-info":[{"award-number":["IO201218-08232-01"]}],"id":[{"id":"10.13039\/100004358","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Ministry of Science"},{"name":"Information Technology Research Center","award":["IITP-2025-RS-2023-00260091"],"award-info":[{"award-number":["IITP-2025-RS-2023-00260091"]}]},{"name":"Institute for Information Communications Technology Planning Evaluation"},{"name":"Graduate School of Metaverse Convergence support Program","award":["IITP-RS-2022-00156318"],"award-info":[{"award-number":["IITP-RS-2022-00156318"]}]},{"DOI":"10.13039\/501100003725","name":"National Research Foundation of Korea","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100003725","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Korea Government","award":["RS-2024-00414230"],"award-info":[{"award-number":["RS-2024-00414230"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Multimedia"],"published-print":{"date-parts":[[2025]]},"DOI":"10.1109\/tmm.2025.3581748","type":"journal-article","created":{"date-parts":[[2025,9,24]],"date-time":"2025-09-24T17:34:02Z","timestamp":1758735242000},"page":"6358-6368","source":"Crossref","is-referenced-by-count":0,"title":["Programmable-Room: Interactive Textured 3D Room Meshes Generation Empowered by Large Language Models"],"prefix":"10.1109","volume":"27","author":[{"given":"Jihyun","family":"Kim","sequence":"first","affiliation":[{"name":"Department of Electronic Engineering, Sogang University, Seoul, South Korea"}]},{"given":"Junho","family":"Park","sequence":"additional","affiliation":[{"name":"Department of Electronic Engineering, Sogang University, Seoul, South Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1135-7502","authenticated-orcid":false,"given":"Kyeongbo","family":"Kong","sequence":"additional","affiliation":[{"name":"Department of Electrical and Electronics Engineering, Pusan National University, Pusan, South Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4809-956X","authenticated-orcid":false,"given":"Suk-Ju","family":"Kang","sequence":"additional","affiliation":[{"name":"Department of Electronic Engineering, Sogang University, Seoul, South Korea"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01436"},{"article-title":"GPT-4 technical report","year":"2023","author":"Achiam","key":"ref2"},{"key":"ref3","first-page":"7909","article-title":"Text2Room: Extracting textured 3D meshes from 2D text-to-image models","volume-title":"Proc. IEEE\/CVF Int. Conf. Comput. Vis.","author":"Hllein","year":"2023"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2024.3361502"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/3dv66043.2025.00069"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.01536"},{"key":"ref7","first-page":"39897","article-title":"SceneScape: Text-driven consistent scene generation","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"36","author":"Fridman","year":"2024"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.00607"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1145\/3550454.3555447"},{"key":"ref11","article-title":"Mvdiffusion: Enabling holistic multi-view image generation with correspondence-aware diffusion","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Tang","year":"2023"},{"key":"ref12","first-page":"11127","article-title":"Uni-ControlNet: All-in-one control to text-to-image diffusion models","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"36","author":"Zhao","year":"2023"},{"article-title":"Deep bidirectional and unidirectional LSTM recurrent neural network for network-wide traffic speed prediction","year":"2018","author":"Cui","key":"ref13"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/3DV53792.2021.00021"},{"key":"ref15","first-page":"12013","article-title":"Atiss: Autoregressive transformers for indoor scene synthesis","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"34","author":"Paschalidou","year":"2021"},{"key":"ref16","first-page":"30026","article-title":"Commonscenes: Generating commonsense 3D indoor scenes with scene graphs","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"36","author":"Zhai","year":"2024"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/cvpr52733.2024.01938"},{"key":"ref18","first-page":"25102","article-title":"GAUDI: A neural architect for immersive 3D scene generation","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"35","author":"Bautista","year":"2022"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00814"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00659"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1145\/3581783.3611800"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/3DV62453.2024.00132"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01825"},{"key":"ref24","first-page":"18225","article-title":"LayoutGPT: Compositional visual planning and generation with large language models","volume-title":"Adv. Neural Inf. Process. Syst.","volume":"36","author":"Feng","year":"2023"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.5573\/ieiespc.2023.12.4.329"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2021.3060291"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2021.3116416"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2023.3278992"},{"issue":"3","key":"ref29","first-page":"209","article-title":"Controlling 3D human action with transformer variational autoencoder in latent space","volume":"13","author":"Kim","year":"2024","journal-title":"IEIE Trans. Smart Process. Comput."},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00355"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i5.28226"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-73027-6_19"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58545-7_30"},{"article-title":"Qwen-vl: A frontier large vision-language model with versatile abilities","year":"2023","author":"Bai","key":"ref34"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00244"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01268"},{"key":"ref37","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","volume-title":"Proc. 38th Int. Conf. Mach. Learn.","author":"Radford","year":"2021"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1706.03762"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-021-01534-z"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01075"},{"key":"ref42","article-title":"GANs trained by a two time-scale update rule converge to a local nash equilibrium","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"30","author":"Heusel","year":"2017"},{"article-title":"Demystifying MMD GANs","year":"2018","author":"Bikowski","key":"ref43"},{"article-title":"Adam: A method for stochastic optimization","year":"2014","author":"Kingma","key":"ref44"}],"container-title":["IEEE Transactions on Multimedia"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/6046\/10844992\/11178224.pdf?arnumber=11178224","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,30]],"date-time":"2025-09-30T13:14:45Z","timestamp":1759238085000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11178224\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"references-count":44,"URL":"https:\/\/doi.org\/10.1109\/tmm.2025.3581748","relation":{},"ISSN":["1520-9210","1941-0077"],"issn-type":[{"type":"print","value":"1520-9210"},{"type":"electronic","value":"1941-0077"}],"subject":[],"published":{"date-parts":[[2025]]}}}