{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,1]],"date-time":"2025-10-01T16:16:55Z","timestamp":1759335415700,"version":"3.37.3"},"reference-count":62,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/legalcode"}],"funder":[{"name":"Institute of Information and Communications Technology Planning and Evaluation"},{"name":"Korea Government [Ministry of Science and ICT (MSIT)]","award":["2021-0-02068"],"award-info":[{"award-number":["2021-0-02068"]}]},{"name":"IITP"},{"DOI":"10.13039\/501100014188","name":"Korea Government","doi-asserted-by":"publisher","award":["2022-0-00759"],"award-info":[{"award-number":["2022-0-00759"]}],"id":[{"id":"10.13039\/501100014188","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100010418","name":"IITP","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100010418","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003621","name":"Korea government","doi-asserted-by":"publisher","award":["RS-2022-00155911"],"award-info":[{"award-number":["RS-2022-00155911"]}],"id":[{"id":"10.13039\/501100003621","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Artificial Intelligence Convergence Innovation Human Resources Development"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2022]]},"DOI":"10.1109\/access.2022.3231339","type":"journal-article","created":{"date-parts":[[2022,12,21]],"date-time":"2022-12-21T18:51:49Z","timestamp":1671648709000},"page":"133969-133980","source":"Crossref","is-referenced-by-count":1,"title":["DenSE SwinHDR: SDRTV to HDRTV Conversion Using Densely Connected Swin Transformer With Squeeze and Excitation Module"],"prefix":"10.1109","volume":"10","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-0214-6962","authenticated-orcid":false,"given":"Joon-Ki","family":"Bae","sequence":"first","affiliation":[{"name":"Department of Computer Science and Engineering, Kyung Hee University, Yongin-si, Republic of Korea"}]},{"given":"Subin","family":"Yang","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Engineering, Kyung Hee University, Yongin-si, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2677-3186","authenticated-orcid":false,"given":"Sung-Ho","family":"Bae","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Engineering, Kyung Hee University, Yongin-si, Republic of Korea"}]}],"member":"263","reference":[{"key":"ref13","first-page":"2802","article-title":"Image restoration using very deep convolutional encoder&#x2013;decoder networks with symmetric skip connections","volume":"29","author":"mao","year":"2016","journal-title":"Proc Adv Neural Inf Process Syst"},{"year":"2017","journal-title":"Colour Gamut Conversion From Recommendation ITU-R BT","key":"ref57"},{"doi-asserted-by":"publisher","key":"ref12","DOI":"10.1007\/978-3-030-01234-2_18"},{"key":"ref56","first-page":"2058","article-title":"Learning image-adaptive 3D lookup tables for high performance photo enhancement in real-time","volume":"44","author":"zeng","year":"2022","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"doi-asserted-by":"publisher","key":"ref15","DOI":"10.1007\/978-3-030-01237-3_32"},{"key":"ref59","article-title":"Adam: A method for stochastic optimization","author":"kingma","year":"2014","journal-title":"arXiv 1412 6980"},{"doi-asserted-by":"publisher","key":"ref14","DOI":"10.1109\/CVPR.2018.00178"},{"key":"ref58","first-page":"303","article-title":"Early convolutions help transformers see better","volume":"34","author":"xiao","year":"0","journal-title":"Proc Adv Neural Inf Process Syst"},{"doi-asserted-by":"publisher","key":"ref53","DOI":"10.1109\/ICCV.2017.97"},{"key":"ref52","article-title":"How do vision transformers work?","author":"park","year":"2022","journal-title":"arXiv 2202 06709"},{"doi-asserted-by":"publisher","key":"ref11","DOI":"10.1109\/CVPR.2018.00082"},{"key":"ref55","article-title":"Very lightweight photo retouching network with conditional sequential modulation","author":"liu","year":"2021","journal-title":"arXiv 2104 06279"},{"doi-asserted-by":"publisher","key":"ref10","DOI":"10.1109\/CVPR.2017.298"},{"key":"ref54","first-page":"10347","article-title":"Training data-efficient image transformers & distillation through attention","author":"touvron","year":"2021","journal-title":"Proc Int Conf Mach Learn"},{"doi-asserted-by":"publisher","key":"ref17","DOI":"10.1007\/978-3-030-01249-6_16"},{"doi-asserted-by":"publisher","key":"ref16","DOI":"10.1109\/ICCV.2017.514"},{"key":"ref19","first-page":"63","article-title":"ESRGAN: Enhanced super-resolution generative adversarial networks","author":"wang","year":"2018","journal-title":"Proc Eur Conf Comput Vis (ECCV) Workshops"},{"doi-asserted-by":"publisher","key":"ref18","DOI":"10.1109\/CVPR.2018.00262"},{"doi-asserted-by":"publisher","key":"ref51","DOI":"10.1109\/CVPR.2018.00745"},{"doi-asserted-by":"publisher","key":"ref50","DOI":"10.1109\/CVPR.2017.243"},{"doi-asserted-by":"publisher","key":"ref46","DOI":"10.1109\/ICCVW54120.2021.00210"},{"doi-asserted-by":"publisher","key":"ref45","DOI":"10.1109\/ICCV48922.2021.00986"},{"doi-asserted-by":"publisher","key":"ref48","DOI":"10.1109\/CVPR46437.2021.01212"},{"doi-asserted-by":"publisher","key":"ref47","DOI":"10.1109\/ICCV48922.2021.00060"},{"year":"2012","journal-title":"Parameter Values for Ultra-high Definition Television Systems for Production and International Programme Exchange","key":"ref42"},{"year":"2002","journal-title":"Parameter Values for the HDTV Standards for Production and International Programme Exchange","first-page":"709","key":"ref41"},{"key":"ref44","article-title":"An image is worth 16 &#x00D7; 16 words: Transformers for image recognition at scale","author":"dosovitskiy","year":"2020","journal-title":"arXiv 2010 11929"},{"key":"ref43","first-page":"5998","article-title":"Attention is all you need","author":"vaswani","year":"2017","journal-title":"Proc Adv Neural Inf Process Syst"},{"doi-asserted-by":"publisher","key":"ref49","DOI":"10.1109\/CVPR.2016.90"},{"doi-asserted-by":"publisher","key":"ref8","DOI":"10.1109\/TPAMI.2015.2439281"},{"doi-asserted-by":"publisher","key":"ref7","DOI":"10.1109\/CVPR42600.2020.00172"},{"doi-asserted-by":"publisher","key":"ref9","DOI":"10.1109\/CVPR.2016.182"},{"doi-asserted-by":"publisher","key":"ref4","DOI":"10.1145\/3130800.3130816"},{"doi-asserted-by":"publisher","key":"ref3","DOI":"10.1109\/ICCV48922.2021.00446"},{"doi-asserted-by":"publisher","key":"ref6","DOI":"10.1007\/978-3-030-01216-8_37"},{"doi-asserted-by":"publisher","key":"ref5","DOI":"10.1111\/cgf.13340"},{"doi-asserted-by":"publisher","key":"ref40","DOI":"10.1111\/cgf.13340"},{"doi-asserted-by":"publisher","key":"ref35","DOI":"10.1109\/GlobalSIP45357.2019.8969167"},{"key":"ref34","article-title":"End-to-end differentiable learning to HDR image synthesis for multi-exposure images","author":"kim","year":"2020","journal-title":"arXiv 2006 15833"},{"doi-asserted-by":"publisher","key":"ref37","DOI":"10.1109\/CVPR42600.2020.00172"},{"doi-asserted-by":"publisher","key":"ref36","DOI":"10.1145\/3386569.3392403"},{"doi-asserted-by":"publisher","key":"ref31","DOI":"10.1117\/12.706472"},{"doi-asserted-by":"publisher","key":"ref30","DOI":"10.1145\/3130800.3130816"},{"doi-asserted-by":"publisher","key":"ref33","DOI":"10.1007\/978-3-030-01216-8_37"},{"doi-asserted-by":"publisher","key":"ref32","DOI":"10.1109\/ACCESS.2018.2868246"},{"doi-asserted-by":"publisher","key":"ref2","DOI":"10.1609\/aaai.v34i07.6789"},{"doi-asserted-by":"publisher","key":"ref1","DOI":"10.1109\/ICCV.2019.00321"},{"doi-asserted-by":"publisher","key":"ref39","DOI":"10.1109\/TCSVT.2019.2892438"},{"key":"ref38","first-page":"349","article-title":"Inverse tone mapping","author":"francesco","year":"2006","journal-title":"Proc 4th Int Conf Comput Graph Interact Techn Australasia Southeast Asia"},{"doi-asserted-by":"publisher","key":"ref24","DOI":"10.1109\/TIP.2018.2839891"},{"doi-asserted-by":"publisher","key":"ref23","DOI":"10.1515\/9783110524116"},{"doi-asserted-by":"publisher","key":"ref26","DOI":"10.1109\/ICTAI.2017.00192"},{"doi-asserted-by":"publisher","key":"ref25","DOI":"10.1109\/CVPR.2018.00344"},{"doi-asserted-by":"publisher","key":"ref20","DOI":"10.1109\/CVPR.2018.00179"},{"doi-asserted-by":"publisher","key":"ref22","DOI":"10.1109\/CVPR.2017.19"},{"doi-asserted-by":"publisher","key":"ref21","DOI":"10.1109\/CVPR.2019.01132"},{"doi-asserted-by":"publisher","key":"ref28","DOI":"10.1109\/LRA.2018.2812225"},{"doi-asserted-by":"publisher","key":"ref27","DOI":"10.1109\/ICCV.2015.55"},{"key":"ref29","first-page":"649","article-title":"Colorful image colorization","author":"zhang","year":"2016","journal-title":"Proc Eur Conf Comput Vis"},{"doi-asserted-by":"publisher","key":"ref60","DOI":"10.1145\/2010324.1964935"},{"key":"ref62","first-page":"139","article-title":"Generative adversarial nets","volume":"27","author":"goodfellow","year":"2014","journal-title":"Proc Adv Neural Inf Process Syst"},{"doi-asserted-by":"publisher","key":"ref61","DOI":"10.1109\/PCS50896.2021.9477471"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/6287639\/9668973\/09996403.pdf?arnumber=9996403","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,1,23]],"date-time":"2023-01-23T20:05:50Z","timestamp":1674504350000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9996403\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"references-count":62,"URL":"https:\/\/doi.org\/10.1109\/access.2022.3231339","relation":{},"ISSN":["2169-3536"],"issn-type":[{"type":"electronic","value":"2169-3536"}],"subject":[],"published":{"date-parts":[[2022]]}}}