{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,18]],"date-time":"2026-03-18T14:12:32Z","timestamp":1773843152753,"version":"3.50.1"},"reference-count":42,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/legalcode"}],"funder":[{"name":"Grant from the Institute of Information and Communications Technology Planning and Evaluation (IITP) by the Korean Government through MSIT","award":["2020-0-00994"],"award-info":[{"award-number":["2020-0-00994"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2022]]},"DOI":"10.1109\/access.2022.3160214","type":"journal-article","created":{"date-parts":[[2022,3,16]],"date-time":"2022-03-16T19:38:47Z","timestamp":1647459527000},"page":"30730-30742","source":"Crossref","is-referenced-by-count":25,"title":["Content-Based Video Retrieval With Prototypes of Deep Features"],"prefix":"10.1109","volume":"10","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-9384-1284","authenticated-orcid":false,"given":"Hyeok","family":"Yoon","sequence":"first","affiliation":[{"name":"Department of Computer Science and Engineering, Seoul National University of Science and Technology, Seoul, South Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8391-6898","authenticated-orcid":false,"given":"Ji-Hyeong","family":"Han","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Engineering, Seoul National University of Science and Technology, Seoul, South Korea"}]}],"member":"263","reference":[{"key":"ref1","volume-title":"YouTube During COVID-19","year":"2021"},{"key":"ref2","volume-title":"Cisco Annual Internet Report (2018\u20132023) White Paper","year":"2021"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/TSMCC.2011.2109710"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2020.103557"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/COMSNETS.2016.7440005"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/TNSM.2021.3051381"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/DICTA.2009.36"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.14445\/22312803\/ijctt-v19p102"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/ICCSNT.2011.6182331"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-23508-5_161"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/MSP.2006.1621446"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ICCP51029.2020.9266253"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2015.2391674"},{"key":"ref14","first-page":"1","article-title":"IBM research TRECVID-2003 video retrieval system","volume-title":"Proc. TREC Video Retr. Eval.","author":"Amir"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1145\/2671188.2749398"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1016\/j.dsp.2020.102729"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.3029834"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.510"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00685"},{"key":"ref20","article-title":"Would mega-scale datasets further enhance spatiotemporal 3D CNNs?","author":"Kataoka","year":"2020","journal-title":"arXiv:2004.04968"},{"key":"ref21","article-title":"A closer look at few-shot classification","volume-title":"Proc. Int. Conf. Learn. Represent. (ICLR)","author":"Chen"},{"key":"ref22","article-title":"Cross-domain few-shot classification via learned feature-wise transformation","volume-title":"Proc. Int. Conf. Learn. Represent. (ICLR)","author":"Tseng"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-015-0816-y"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.223"},{"key":"ref25","article-title":"YouTube-8M: A large-scale video classification benchmark","author":"Abu-El-Haija","year":"2016","journal-title":"arXiv:1609.08675"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.3169\/mta.4.227"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/ICCSN.2019.8905258"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1145\/509907.509965"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00675"},{"key":"ref30","article-title":"The kinetics human action video dataset","author":"Kay","year":"2017","journal-title":"arXiv:1705.06950"},{"key":"ref31","article-title":"A short note on the kinetics-700 human action dataset","author":"Carreira","year":"2019","journal-title":"arXiv:1907.06987"},{"key":"ref32","article-title":"ConvNet architecture search for spatiotemporal feature learning","author":"Tran","year":"2017","journal-title":"arXiv:1708.05038"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW.2017.373"},{"key":"ref34","first-page":"4077","article-title":"Prototypical networks for few-shot learning","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Snell"},{"key":"ref35","first-page":"3630","article-title":"Matching networks for one shot learning","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Vinyals"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00131"},{"key":"ref37","article-title":"UCF101: A dataset of 101 human actions classes from videos in the wild","author":"Soomro","year":"2012","journal-title":"arXiv:1212.0402"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2011.6126543"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298698"},{"key":"ref40","first-page":"318","article-title":"Salient video frames sampling method using the mean of deep features for efficient model training","volume-title":"Proc. Conf. Korean Inst. Broadcast Media Eng.","author":"Yoon"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-021-11217-z"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-12900-1_1"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/6287639\/9668973\/09737137.pdf?arnumber=9737137","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,18]],"date-time":"2024-01-18T00:25:05Z","timestamp":1705537505000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9737137\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"references-count":42,"URL":"https:\/\/doi.org\/10.1109\/access.2022.3160214","relation":{},"ISSN":["2169-3536"],"issn-type":[{"value":"2169-3536","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]}}}