{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,22]],"date-time":"2025-11-22T05:43:58Z","timestamp":1763790238216,"version":"3.45.0"},"reference-count":49,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/legalcode"}],"funder":[{"name":"Tokyo City University Prioritized Studies"},{"DOI":"10.13039\/501100001691","name":"Japan Society for Science Promotion (JSPS) Grant-in-Aid for Scientific Research","doi-asserted-by":"publisher","award":["23K03743","23H01176","23K25872","23H04520"],"award-info":[{"award-number":["23K03743","23H01176","23K25872","23H04520"]}],"id":[{"id":"10.13039\/501100001691","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Grant-in-Aid for JSPS Fellows","award":["22KF0329"],"award-info":[{"award-number":["22KF0329"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2025]]},"DOI":"10.1109\/access.2025.3628273","type":"journal-article","created":{"date-parts":[[2025,11,3]],"date-time":"2025-11-03T18:45:59Z","timestamp":1762195559000},"page":"188839-188854","source":"Crossref","is-referenced-by-count":0,"title":["Driver Activity Recognition With Vision Transformer Using Time\u2013Frequency Representations Derived From Wrist-Worn Sensors"],"prefix":"10.1109","volume":"13","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8810-4813","authenticated-orcid":false,"given":"Yusuke","family":"Sakai","sequence":"first","affiliation":[{"name":"Research Center for Space Science, Advanced Research Laboratories, Tokyo City University, Setagaya-ku, Japan"}]},{"given":"Takuma","family":"Akiduki","sequence":"additional","affiliation":[{"name":"Department of Mechanical Engineering, Faculty of Engineering, University of Yamanashi, Kofu, Yamanashi, Japan"}]},{"given":"Marco","family":"Meyer-Conde","sequence":"additional","affiliation":[{"name":"Research Center for Space Science, Advanced Research Laboratories, Tokyo City University, Setagaya-ku, Japan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0596-4397","authenticated-orcid":false,"given":"Hirotaka","family":"Takahashi","sequence":"additional","affiliation":[{"name":"Research Center for Space Science, Advanced Research Laboratories, Tokyo City University, Setagaya-ku, Japan"}]}],"member":"263","reference":[{"key":"ref1","first-page":"649","article-title":"Nhtsa driver distraction research: Past, present, and future","volume-title":"Proc. SAE Tech. Paper","author":"Ranney"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1016\/j.trf.2010.12.001"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/TVT.2019.2908425"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2022.3218711"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.3390\/s23218844"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/TR.2017.2778754"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1016\/j.smhl.2018.07.022"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.3390\/s21103346"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1177\/154193120004401702"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.3390\/s22010352"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1145\/3161179"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1016\/j.apergo.2017.02.016"},{"issue":"4","key":"ref13","first-page":"1115","article-title":"A method for measuring and analyzing driving behavior using wireless accelerometers","volume":"91","author":"Tada","year":"2008","journal-title":"IEICE Trans. Inf. Syst."},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1145\/3729467"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1145\/2499621"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.3390\/su13031342"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1177\/0018720821995000"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.3156\/jsoft.34.2_544"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.23919\/FRUCT48121.2019.8981511"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/SMC42975.2020.9282860"},{"key":"ref21","article-title":"UMAP: Uniform manifold approximation and projection for dimension reduction","author":"McInnes","year":"2018","journal-title":"arXiv:1802.03426"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1706.03762"},{"key":"ref23","first-page":"4171","article-title":"BERT: Pre-training of deep bidirectional transformers for language understanding","volume-title":"Proc. Conf. North Amer. Chapter Assoc. Comput. Linguistics: Hum. Lang. Technol.","author":"Devlin"},{"key":"ref24","article-title":"Hierarchical text-conditional image generation with CLIP latents","author":"Ramesh","year":"2022","journal-title":"arXiv:2204.06125"},{"key":"ref25","first-page":"1877","article-title":"Language models are few-shot learners","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Brown"},{"key":"ref26","article-title":"GPT-4 technical report","volume-title":"arXiv:2303.08774","author":"Achiam","year":"2023"},{"key":"ref27","article-title":"An image is worth 16\u00d716 words: Transformers for image recognition at scale","author":"Dosovitskiy","year":"2020","journal-title":"arXiv:2010.11929"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01424-7_27"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1186\/s40537-017-0089-0"},{"key":"ref31","article-title":"A recent survey of heterogeneous transfer learning","author":"Bao","year":"2023","journal-title":"arXiv:2310.08459"},{"key":"ref32","article-title":"Deep transfer learning: A new deep learning glitch classification method for advanced LIGO","author":"George","year":"2017","journal-title":"arXiv:1706.07446"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1016\/b978-0-12-374370-1.x0001-8"},{"article-title":"Exploring frequency-based features for driver activity recognition with Wrist\u2013Worn accelerometers","volume-title":"Proc. 18th Int. Conf. Innov. Comput., Inf. Control","author":"Akiduki","key":"ref34"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.7551\/mitpress\/1120.003.0080"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1162\/089976603321780317"},{"issue":"4","key":"ref37","volume":"4","author":"Bishop","year":"2006","journal-title":"Pattern Recognition and Machine Learning"},{"key":"ref38","first-page":"10347","article-title":"Training data-efficient image transformers & distillation through attention","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Touvron"},{"key":"ref39","article-title":"Scaling laws for neural language models","author":"Kaplan","year":"2020","journal-title":"arXiv:2001.08361"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1083"},{"key":"ref41","article-title":"Sophia: A scalable stochastic second-order optimizer for language model pre-training","author":"Liu","year":"2023","journal-title":"arXiv:2305.14342"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref43","article-title":"Are convolutional neural networks or transformers more like human vision?","author":"Tuli","year":"2021","journal-title":"arXiv:2105.07197"},{"key":"ref44","article-title":"Channel vision transformers: An image is worth 1 \u00d7 16 \u00d7 16 words","author":"Bao","year":"2023","journal-title":"arXiv:2309.16108"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00502"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2021-698"},{"key":"ref47","article-title":"Scaling laws for transfer","author":"Hernandez","year":"2021","journal-title":"arXiv:2102.01293"},{"key":"ref48","article-title":"MobileViT: Light-weight, general-purpose, and mobile-friendly vision transformer","author":"Mehta","year":"2021","journal-title":"arXiv:2110.02178"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19803-8_5"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/6287639\/10820123\/11224562.pdf?arnumber=11224562","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,22]],"date-time":"2025-11-22T05:40:12Z","timestamp":1763790012000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11224562\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"references-count":49,"URL":"https:\/\/doi.org\/10.1109\/access.2025.3628273","relation":{},"ISSN":["2169-3536"],"issn-type":[{"type":"electronic","value":"2169-3536"}],"subject":[],"published":{"date-parts":[[2025]]}}}