{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T20:22:31Z","timestamp":1740169351150,"version":"3.37.3"},"reference-count":82,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"funder":[{"name":"This work was supported and funded by the ETRI Research and Development Support Program of MSIT\/IITP","award":["Project Number:RS-2023-00216821"],"award-info":[{"award-number":["Project Number:RS-2023-00216821"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2024]]},"DOI":"10.1109\/access.2024.3355814","type":"journal-article","created":{"date-parts":[[2024,1,18]],"date-time":"2024-01-18T18:34:01Z","timestamp":1705602841000},"page":"22866-22879","source":"Crossref","is-referenced-by-count":0,"title":["CorrFractal: High-Resolution Correspondence Method Using Fractal Affinity on Self-Supervised Learning"],"prefix":"10.1109","volume":"12","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7800-099X","authenticated-orcid":false,"given":"Jin-Mo","family":"Choi","sequence":"first","affiliation":[{"name":"Department of Computer Software, University of Science and Technology, Daejeon, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-8197-1330","authenticated-orcid":false,"given":"Blagovest I.","family":"Vladimirov","sequence":"additional","affiliation":[{"name":"Defence and Safety Convergence Research Division, Electronics and Telecommunications Research Institute, Daejeon, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4765-8905","authenticated-orcid":false,"given":"Sangjoon","family":"Park","sequence":"additional","affiliation":[{"name":"Defence and Safety Convergence Research Division, Electronics and Telecommunications Research Institute, Daejeon, Republic of Korea"}]}],"member":"263","reference":[{"doi-asserted-by":"publisher","key":"ref1","DOI":"10.1109\/ICCV.2009.5459322"},{"doi-asserted-by":"publisher","key":"ref2","DOI":"10.1109\/CVPR.2013.361"},{"doi-asserted-by":"publisher","key":"ref3","DOI":"10.1007\/978-3-642-12297-2_44"},{"doi-asserted-by":"publisher","key":"ref4","DOI":"10.1109\/TPAMI.2007.70769"},{"doi-asserted-by":"publisher","key":"ref5","DOI":"10.1109\/TPAMI.2006.80"},{"doi-asserted-by":"publisher","key":"ref6","DOI":"10.1109\/7.366332"},{"doi-asserted-by":"publisher","key":"ref7","DOI":"10.1037\/a0033117"},{"doi-asserted-by":"publisher","key":"ref8","DOI":"10.1109\/34.677275"},{"doi-asserted-by":"publisher","key":"ref9","DOI":"10.1109\/CVPR42600.2020.00584"},{"doi-asserted-by":"publisher","key":"ref10","DOI":"10.1109\/CVPR.2019.00267"},{"doi-asserted-by":"publisher","key":"ref11","DOI":"10.1007\/978-3-030-01261-8_24"},{"doi-asserted-by":"publisher","key":"ref12","DOI":"10.1109\/CVPR42600.2020.00651"},{"volume-title":"Proc. BMVC","author":"Lai","article-title":"Self-supervised learning for video correspondence flow","key":"ref13"},{"key":"ref14","article-title":"Space-time correspondence as a contrastive random walk","author":"Jabri","year":"2020","journal-title":"arXiv:2006.14613"},{"key":"ref15","first-page":"317","article-title":"Joint-task self-supervised learning for temporal correspondence","volume-title":"Proc. Adv. Neural Inf. Process. Syst. (NeurIPS)","volume":"32","author":"Li"},{"doi-asserted-by":"publisher","key":"ref16","DOI":"10.1007\/978-3-030-01219-9_25"},{"key":"ref17","article-title":"Learning to downsample for segmentation of ultra-high resolution images","author":"Jin","year":"2021","journal-title":"arXiv:2109.11071"},{"doi-asserted-by":"publisher","key":"ref18","DOI":"10.1109\/CVPR.2019.00584"},{"key":"ref19","article-title":"Deep high-resolution representation learning for visual recognition","author":"Wang","year":"2019","journal-title":"arXiv:1908.07919"},{"doi-asserted-by":"publisher","key":"ref20","DOI":"10.1109\/TIP.2021.3120054"},{"doi-asserted-by":"publisher","key":"ref21","DOI":"10.1609\/aaai.v35i4.16432"},{"doi-asserted-by":"publisher","key":"ref22","DOI":"10.1109\/TC.1976.1674569"},{"doi-asserted-by":"publisher","key":"ref23","DOI":"10.1609\/aaai.v34i07.7008"},{"key":"ref24","article-title":"Multiple object tracking with motion and appearance cues","author":"Li","year":"2017","journal-title":"arXiv:1909.00318"},{"doi-asserted-by":"publisher","key":"ref25","DOI":"10.1109\/CVPR46437.2021.01068"},{"doi-asserted-by":"publisher","key":"ref26","DOI":"10.1016\/j.bandl.2012.08.003"},{"key":"ref27","first-page":"287","article-title":"Learning video correspondence using appearance module for target tracking","volume-title":"Proc. IEEE Int. Conf. Big Data Smart Comput. (BigComp)","author":"Son"},{"key":"ref28","article-title":"Feature pyramid networks for object detection","author":"Lin","year":"2016","journal-title":"arXiv:1612.03144"},{"key":"ref29","article-title":"Deep layer aggregation","author":"Yu","year":"2017","journal-title":"arXiv:1707.06484"},{"doi-asserted-by":"publisher","key":"ref30","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"ref31","article-title":"Densely connected convolutional networks","author":"Huang","year":"2016","journal-title":"arXiv:1608.06993"},{"doi-asserted-by":"publisher","key":"ref32","DOI":"10.1007\/978-3-030-01258-8_12"},{"key":"ref33","article-title":"FractalNet: Ultra-deep neural networks without residuals","author":"Larsson","year":"2016","journal-title":"arXiv:1605.07648"},{"key":"ref34","article-title":"The 2017 Davis challenge on video object segmentation","author":"Pont-Tuset","year":"2017","journal-title":"arXiv:1704.00675"},{"key":"ref35","article-title":"A survey on deep learning technique for video segmentation","author":"Zhou","year":"2021","journal-title":"arXiv:2107.01153"},{"doi-asserted-by":"publisher","key":"ref36","DOI":"10.1109\/ICCV.2019.00502"},{"doi-asserted-by":"publisher","key":"ref37","DOI":"10.1007\/s10462-022-10176-7"},{"doi-asserted-by":"publisher","key":"ref38","DOI":"10.1109\/CVPR.2017.565"},{"doi-asserted-by":"publisher","key":"ref39","DOI":"10.1109\/CVPR.2018.00125"},{"doi-asserted-by":"publisher","key":"ref40","DOI":"10.1109\/TPAMI.2018.2838670"},{"doi-asserted-by":"publisher","key":"ref41","DOI":"10.1109\/TPAMI.2018.2890659"},{"doi-asserted-by":"publisher","key":"ref42","DOI":"10.1109\/CVPR.2017.372"},{"doi-asserted-by":"publisher","key":"ref43","DOI":"10.1007\/978-3-030-01219-9_6"},{"doi-asserted-by":"publisher","key":"ref44","DOI":"10.1109\/CVPR.2018.00152"},{"doi-asserted-by":"publisher","key":"ref45","DOI":"10.1109\/CVPR.2018.00946"},{"doi-asserted-by":"publisher","key":"ref46","DOI":"10.1109\/CVPR.2018.00770"},{"doi-asserted-by":"publisher","key":"ref47","DOI":"10.1109\/ICCV.2019.00568"},{"doi-asserted-by":"publisher","key":"ref48","DOI":"10.1109\/ICCV.2019.00405"},{"doi-asserted-by":"publisher","key":"ref49","DOI":"10.1109\/CVPR.2017.336"},{"doi-asserted-by":"publisher","key":"ref50","DOI":"10.1109\/CVPR.2019.00147"},{"doi-asserted-by":"publisher","key":"ref51","DOI":"10.1109\/ICCV.2017.480"},{"doi-asserted-by":"publisher","key":"ref52","DOI":"10.1007\/978-3-030-01228-1_36"},{"doi-asserted-by":"publisher","key":"ref53","DOI":"10.1109\/CVPR.2019.00542"},{"doi-asserted-by":"publisher","key":"ref54","DOI":"10.1109\/ICCV.2017.238"},{"doi-asserted-by":"publisher","key":"ref55","DOI":"10.1109\/ICCV.2019.00932"},{"key":"ref56","first-page":"11781","article-title":"Rethinking space-time networks with improved memory coverage for efficient video object segmentation","volume-title":"Proc. NeurIPS","author":"Cheng"},{"doi-asserted-by":"publisher","key":"ref57","DOI":"10.1007\/978-3-030-58580-8_39"},{"key":"ref58","first-page":"3430","article-title":"Video object segmentation with adaptive feature bank and uncertain-region refinement","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Liang"},{"doi-asserted-by":"publisher","key":"ref59","DOI":"10.1007\/978-3-031-19815-1_37"},{"doi-asserted-by":"publisher","key":"ref60","DOI":"10.1109\/CVPR.2018.00774"},{"doi-asserted-by":"publisher","key":"ref61","DOI":"10.1109\/ICCV48922.2021.00992"},{"doi-asserted-by":"publisher","key":"ref62","DOI":"10.1016\/j.visres.2008.03.006"},{"doi-asserted-by":"publisher","key":"ref63","DOI":"10.1016\/j.conb.2007.03.002"},{"doi-asserted-by":"publisher","key":"ref64","DOI":"10.1016\/S0960-9822(03)00052-6"},{"doi-asserted-by":"publisher","key":"ref65","DOI":"10.1016\/j.neuroimage.2022.118941"},{"doi-asserted-by":"publisher","key":"ref66","DOI":"10.5555\/2969033.2969125"},{"key":"ref67","article-title":"Progressive growing of GANs for improved quality, stability, and variation","author":"Karras","year":"2017","journal-title":"arXiv:1710.10196"},{"doi-asserted-by":"publisher","key":"ref68","DOI":"10.1109\/CVPR.2019.00453"},{"doi-asserted-by":"publisher","key":"ref69","DOI":"10.1109\/ICCV.2017.322"},{"volume-title":"Proc. CVPR","author":"He","article-title":"Deep residual learning for image recognition. in computer vision and pattern recognition","key":"ref70"},{"key":"ref71","article-title":"YouTube-VOS: A large-scale video object segmentation benchmark","author":"Xu","year":"2018","journal-title":"arXiv:1809.03327"},{"doi-asserted-by":"publisher","key":"ref72","DOI":"10.1007\/978-3-030-01219-9_41"},{"doi-asserted-by":"publisher","key":"ref73","DOI":"10.1145\/3240508.3240660"},{"doi-asserted-by":"publisher","key":"ref74","DOI":"10.1109\/ICCV.2013.396"},{"key":"ref75","article-title":"The kinetics human action video dataset","author":"Kay","year":"2017","journal-title":"arXiv:1705.06950"},{"doi-asserted-by":"publisher","key":"ref76","DOI":"10.1109\/CVPR.2018.00524"},{"doi-asserted-by":"publisher","key":"ref77","DOI":"10.1109\/ICME52920.2022.9859966"},{"doi-asserted-by":"publisher","key":"ref78","DOI":"10.1109\/CVPR52688.2022.01427"},{"doi-asserted-by":"publisher","key":"ref79","DOI":"10.1109\/CVPR52688.2022.00852"},{"doi-asserted-by":"publisher","key":"ref80","DOI":"10.1109\/ICCV51070.2023.00883"},{"doi-asserted-by":"publisher","key":"ref81","DOI":"10.1109\/CVPR.2016.438"},{"doi-asserted-by":"publisher","key":"ref82","DOI":"10.1007\/978-3-031-19821-2_6"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/6287639\/10380310\/10403912.pdf?arnumber=10403912","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,14]],"date-time":"2024-03-14T12:16:44Z","timestamp":1710418604000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10403912\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"references-count":82,"URL":"https:\/\/doi.org\/10.1109\/access.2024.3355814","relation":{},"ISSN":["2169-3536"],"issn-type":[{"type":"electronic","value":"2169-3536"}],"subject":[],"published":{"date-parts":[[2024]]}}}