{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,5]],"date-time":"2026-02-05T22:59:26Z","timestamp":1770332366680,"version":"3.49.0"},"reference-count":28,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,1,19]],"date-time":"2025-01-19T00:00:00Z","timestamp":1737244800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,1,19]],"date-time":"2025-01-19T00:00:00Z","timestamp":1737244800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100003725","name":"National Research Foundation of Korea","doi-asserted-by":"publisher","award":["2021R1A2C2014557,RS-2024-00397293"],"award-info":[{"award-number":["2021R1A2C2014557,RS-2024-00397293"]}],"id":[{"id":"10.13039\/501100003725","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100010418","name":"IITP","doi-asserted-by":"publisher","award":["RS-2022-00167143"],"award-info":[{"award-number":["RS-2022-00167143"]}],"id":[{"id":"10.13039\/501100010418","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,1,19]]},"DOI":"10.1109\/iceic64972.2025.10879767","type":"proceedings-article","created":{"date-parts":[[2025,2,18]],"date-time":"2025-02-18T18:17:22Z","timestamp":1739902642000},"page":"1-6","source":"Crossref","is-referenced-by-count":1,"title":["Lightweight Look-ahead Multi-context Vision Transformer"],"prefix":"10.1109","author":[{"given":"Keunho","family":"Byeon","sequence":"first","affiliation":[{"name":"School of Electrical Engineering Korea University,Seoul,South Korea"}]},{"given":"Jeewoo","family":"Lim","sequence":"additional","affiliation":[{"name":"School of Electrical Engineering Korea University,Seoul,South Korea"}]},{"given":"Jaeung","family":"Lee","sequence":"additional","affiliation":[{"name":"School of Electrical Engineering Korea University,Seoul,South Korea"}]},{"given":"Chiwon","family":"Han","sequence":"additional","affiliation":[{"name":"Sejong University,Department of Computer Science and Engineering,Seoul,South Korea"}]},{"given":"SungYun","family":"Bay","sequence":"additional","affiliation":[{"name":"Sejong University,Department of Computer Science and Engineering,Seoul,South Korea"}]},{"given":"Giho","family":"Park","sequence":"additional","affiliation":[{"name":"Sejong University,Department of Computer Science and Engineering,Seoul,South Korea"}]},{"given":"Jin Tae","family":"Kwak","sequence":"additional","affiliation":[{"name":"School of Electrical Engineering Korea University,Seoul,South Korea"}]}],"member":"263","reference":[{"key":"ref1","first-page":"6105","article-title":"EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks","author":"Mingxing","journal-title":"International Conference on Machine Learning"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00474"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2010.11929"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"ref6","article-title":"ViDT: An Efficient and Effective Fully Transformer-Based Object Detector","author":"Song","year":"2021","journal-title":"arXiv preprint"},{"key":"ref7","first-page":"10347","article-title":"Training Data-Efficient Image Transformers & Distillation Through Attention","volume-title":"International Conference on Machine Learning","author":"Touvron"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00060"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00520"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01186"},{"key":"ref11","article-title":"MobileViT: Light-weight, General-Purpose, and Mobile-Friendly Vision Transformer","author":"Mehta","year":"2021","journal-title":"arXiv preprint"},{"key":"ref12","article-title":"Separable Self-Attention for Mobile Vision Transformers","author":"Mehta","year":"2022","journal-title":"arXiv preprint"},{"key":"ref13","article-title":"MobileViTv3: Mobile-Friendly Vision Transformer with Simple and Effective Fusion of Local, Global and Input Features","author":"Wadekar","year":"2022","journal-title":"arXiv preprint"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i16.17664"},{"key":"ref15","first-page":"30392","article-title":"Early Convolutions Help Transformers See Better","volume":"34","author":"Xiao","year":"2021","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00009"},{"key":"ref17","first-page":"2286","article-title":"ConViT: Improving Vision Transformers with Soft Con-volutional Inductive Biases","volume-title":"International Conference on Machine Learning","author":"d\u2019Ascoli"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01172"},{"key":"ref19","first-page":"9355","article-title":"Twins: Revisiting the Design of Spatial Attention in Vision Transformers","volume":"34","author":"Chu","year":"2021","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2022.3206108"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1007\/s41095-022-0274-8"},{"key":"ref22","article-title":"Conditional Positional Encodings for Vision Transformers","author":"Chu","year":"2021","journal-title":"arXiv preprint"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01170"},{"key":"ref25","first-page":"23818","article-title":"Efficient Training of Visual Transformers with Small Datasets","volume":"34","author":"Liu","year":"2021","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19778-9_18"},{"key":"ref27","article-title":"FasterViT: Fast Vision Transformers with Hierarchical Attention","author":"Hatamizadeh","year":"2023","journal-title":"arXiv preprint"},{"key":"ref28","first-page":"21665","article-title":"Fast Transformers with Clustered Attention","volume":"33","author":"Vyas","year":"2020","journal-title":"Advances in Neural Information Processing Systems"}],"event":{"name":"2025 International Conference on Electronics, Information, and Communication (ICEIC)","location":"Osaka, Japan","start":{"date-parts":[[2025,1,19]]},"end":{"date-parts":[[2025,1,22]]}},"container-title":["2025 International Conference on Electronics, Information, and Communication (ICEIC)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10879455\/10879601\/10879767.pdf?arnumber=10879767","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,2,20]],"date-time":"2025-02-20T20:00:14Z","timestamp":1740081614000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10879767\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,1,19]]},"references-count":28,"URL":"https:\/\/doi.org\/10.1109\/iceic64972.2025.10879767","relation":{},"subject":[],"published":{"date-parts":[[2025,1,19]]}}}