{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,27]],"date-time":"2026-01-27T13:37:24Z","timestamp":1769521044733,"version":"3.49.0"},"reference-count":38,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"funder":[{"name":"Korea Health Technology Resaerch and Development Project through the Korea Health Industry Development Institute"},{"DOI":"10.13039\/501100003710","name":"Ministry of Health and Welfare, Republic of Korea","doi-asserted-by":"publisher","award":["HI22C1518"],"award-info":[{"award-number":["HI22C1518"]}],"id":[{"id":"10.13039\/501100003710","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2023]]},"DOI":"10.1109\/access.2023.3342838","type":"journal-article","created":{"date-parts":[[2023,12,14]],"date-time":"2023-12-14T19:46:43Z","timestamp":1702583203000},"page":"143628-143634","source":"Crossref","is-referenced-by-count":4,"title":["An Effective Representation Learning Approach: The Integrated Self-Supervised Pre-Training Models of StyleGAN2-ADA and DINO for Colon Polyp Images"],"prefix":"10.1109","volume":"11","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1230-9307","authenticated-orcid":false,"given":"Jong-Yeup","family":"Kim","sequence":"first","affiliation":[{"name":"Department of Biomedical Informatics, College of Medicine, Konyang University, Daejeon, Republic of Korea"}]},{"given":"Gayrat","family":"Tangriberganov","sequence":"additional","affiliation":[{"name":"Healthcare Data Science Center, Konyang University Hospital, Daejeon, Republic of Korea"}]},{"given":"Woochul","family":"Jung","sequence":"additional","affiliation":[{"name":"Healthcare Data Science Center, Konyang University Hospital, Daejeon, Republic of Korea"}]},{"given":"Dae Sung","family":"Kim","sequence":"additional","affiliation":[{"name":"Department of Internal Medicine, College of Medicine, Konyang University, Daejeon, Republic of Korea"}]},{"given":"Hoon Sup","family":"Koo","sequence":"additional","affiliation":[{"name":"Department of Internal Medicine, College of Medicine, Konyang University, Daejeon, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2273-0915","authenticated-orcid":false,"given":"Suehyun","family":"Lee","sequence":"additional","affiliation":[{"name":"College of IT Convergence, Gachon University, Seongnam, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0436-3381","authenticated-orcid":false,"given":"Sun Moon","family":"Kim","sequence":"additional","affiliation":[{"name":"Department of Internal Medicine, College of Medicine, Konyang University, Daejeon, Republic of Korea"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1016\/j.gie.2014.12.022"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0250632"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01234-2_49"},{"key":"ref4","article-title":"Generative adversarial networks","author":"Goodfellow","year":"2014","journal-title":"arXiv:1406.2661"},{"key":"ref5","first-page":"12104","article-title":"Training generative adversarial networks with limited data","volume-title":"Proc. NeurIPS","author":"Karras"},{"key":"ref6","article-title":"An image is worth 16\u00d716 words: Transformers for image recognition at scale","volume-title":"Proc. Int. Conf. Learn. Represent.","author":"Dosovitskiy"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00681"},{"key":"ref8","article-title":"UNETR: Transformers for 3D medical image segmentation","author":"Hatamizadeh","year":"2021","journal-title":"arXiv:2103.10504"},{"key":"ref9","article-title":"BERT: Pre-training of deep bidirectional transformers for language understanding","author":"Devlin","year":"2018","journal-title":"arXiv:1810.04805"},{"key":"ref10","article-title":"Do vision transformers see like convolutional neural networks?","author":"Raghu","year":"2021","journal-title":"arXiv:2108.08810"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00951"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00194"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.492"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00086"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00453"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01001"},{"key":"ref17","article-title":"Image GANs meet differentiable rendering for inverse graphics and interpretable 3D neural rendering","volume-title":"Proc. Int. Conf. Learn. Represent.","author":"Zhang"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2021.107090"},{"key":"ref19","article-title":"Attention is all you need","volume-title":"Proc. NeurIPS","author":"Vaswani"},{"key":"ref20","volume-title":"Improving language understanding by generative pre-training","author":"Radford","year":"2018"},{"issue":"8","key":"ref21","first-page":"9","article-title":"Language models are unsupervised multitask learners","volume":"1","author":"Radford","year":"2019","journal-title":"OpenAI Blog"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00813"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.01009"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/ISBI.2019.8759553"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-32245-8_60"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1007\/s11548-018-1772-0"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/NICS54270.2021.9701580"},{"key":"ref28","article-title":"Towards fine-grained visual representations by combining contrastive learning with image reconstruction and attention-weighted pooling","author":"Dippel","year":"2021","journal-title":"arXiv:2104.04323"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00502"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-37734-2_37"},{"key":"ref31","article-title":"GANs trained by a two time-scale update rule converge to a local Nash equilibrium","volume-title":"Proc. NIPS","author":"Heusel"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/ISM46123.2019.00049"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1038\/s41746-020-00353-9"},{"key":"ref34","article-title":"Training data-efficient image transformers & distillation through attention","author":"Touvron","year":"2020","journal-title":"arXiv:2012.12877"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00060"},{"key":"ref36","article-title":"When vision transformers outperform ResNets without pre-training or strong data augmentations","author":"Chen","year":"2021","journal-title":"arXiv:2106.01548"},{"key":"ref37","article-title":"Sharpness-aware minimization for efficiently improving generalization","author":"Foret","year":"2020","journal-title":"arXiv:2010.01412"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.308"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/6287639\/10005208\/10359525.pdf?arnumber=10359525","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,12]],"date-time":"2024-01-12T20:34:36Z","timestamp":1705091676000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10359525\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"references-count":38,"URL":"https:\/\/doi.org\/10.1109\/access.2023.3342838","relation":{},"ISSN":["2169-3536"],"issn-type":[{"value":"2169-3536","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023]]}}}