{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T22:22:22Z","timestamp":1775082142890,"version":"3.50.1"},"reference-count":40,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/legalcode"}],"funder":[{"name":"Regional Innovation System & Education (RISE) program through the Jeju RISE center"},{"name":"Ministry of Education (MOE) and the Jeju Special Self-Governing Province, Republic of Korea","award":["2025-RISE-17-001"],"award-info":[{"award-number":["2025-RISE-17-001"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2026]]},"DOI":"10.1109\/access.2026.3674427","type":"journal-article","created":{"date-parts":[[2026,3,17]],"date-time":"2026-03-17T20:24:47Z","timestamp":1773779087000},"page":"42064-42085","source":"Crossref","is-referenced-by-count":0,"title":["Edge-VIS: Knowledge-Distilled Video Instance Segmentation for Energy-Efficient Real-Time Robotic Scene Understanding"],"prefix":"10.1109","volume":"14","author":[{"ORCID":"https:\/\/orcid.org\/0009-0004-1005-4229","authenticated-orcid":false,"given":"Hasnain","family":"Hyder","sequence":"first","affiliation":[{"name":"Department of Electronic Engineering, Institute of Information Science and Technology, Jeju National University, Jeju-si, South Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4759-0138","authenticated-orcid":false,"given":"Yong-Woon","family":"Kim","sequence":"additional","affiliation":[{"name":"Department of Computer Engineering, Jeju National University, Jeju-si, South Korea"}]},{"given":"Chulung","family":"Kang","sequence":"additional","affiliation":[{"name":"Department of Mechanical Systems Engineering, Jeju National University, Jeju-si, South Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1107-9941","authenticated-orcid":false,"given":"Yung-Cheol","family":"Byun","sequence":"additional","affiliation":[{"name":"Department of Computer Engineering, Major of Electronic Engineering, Food Tech Center (FTC), Jeju National University, Jeju, South Korea"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01828"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2020.2984887"},{"key":"ref3","article-title":"Benchmarking TinyML systems: Challenges and direction","author":"Banbury","year":"2020","journal-title":"arXiv:2003.04821"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01405"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-022-01629-1"},{"key":"ref6","volume-title":"Ultralytics YOLOv8","author":"Jocher","year":"2023"},{"issue":"2","key":"ref7","first-page":"345","article-title":"Lean deep neural networks: A survey on energy-efficient architectures and algorithms","volume":"42","author":"Venieris","year":"2023","journal-title":"IEEE Trans. Comput.-Aided Design Integr. Circuits Syst."},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00529"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.52202\/068431-1679"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00124"},{"key":"ref12","article-title":"DVIS++: Improved decoupled framework for universal video segmentation","author":"Zhang","year":"2023","journal-title":"arXiv:2312.13305"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.52202\/068431-2267"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00089"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00107"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-72667-5_6"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.02147"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19815-1_32"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01797"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1503.02531"},{"key":"ref21","first-page":"1","article-title":"FitNets: Hints for thin deep nets","volume-title":"Proc. Int. Conf. Learn. Represent. (ICLR)","author":"Romero"},{"key":"ref22","first-page":"1","article-title":"Paying more attention to attention: Improving the performance of convolutional neural networks via attention transfer","volume-title":"Proc. Int. Conf. Learn. Represent. (ICLR)","author":"Zagoruyko"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-021-01453-z"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00460"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00919"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2023.3248583"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00271"},{"key":"ref28","first-page":"1","article-title":"Learning efficient object detection models with knowledge distillation","volume-title":"Proc. Adv. Neural Inf. Process. Syst. (NeurIPS)","volume":"30","author":"Chen"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2022.3151999"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00035"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2025.3600622"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2025.3585288"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1007\/978-981-96-2395-2_8"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2024.111915"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1145\/3699518"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2024.3387328"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-025-20808-z"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.3390\/ai5020042"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1117\/12.3045691"},{"key":"ref40","first-page":"17675","article-title":"MC-bench: A benchmark for multi-context visual grounding in the era of MLLMs","volume-title":"Proc. IEEE\/CVF Int. Conf. Comput. Vis.","author":"Xu"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/6287639\/11323511\/11435949.pdf?arnumber=11435949","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T20:13:45Z","timestamp":1775074425000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11435949\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"references-count":40,"URL":"https:\/\/doi.org\/10.1109\/access.2026.3674427","relation":{},"ISSN":["2169-3536"],"issn-type":[{"value":"2169-3536","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]}}}