{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T14:45:06Z","timestamp":1776091506953,"version":"3.50.1"},"reference-count":44,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/legalcode"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2021]]},"DOI":"10.1109\/access.2021.3070646","type":"journal-article","created":{"date-parts":[[2021,4,9]],"date-time":"2021-04-09T02:33:27Z","timestamp":1617935607000},"page":"53540-53547","source":"Crossref","is-referenced-by-count":105,"title":["Boosting Inertial-Based Human Activity Recognition With Transformers"],"prefix":"10.1109","volume":"9","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-3393-6070","authenticated-orcid":false,"given":"Yoli","family":"Shavit","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7846-0654","authenticated-orcid":false,"given":"Itzik","family":"Klein","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","article-title":"Layer normalization","author":"lei ba","year":"2016","journal-title":"arXiv 1607 06450"},{"key":"ref38","article-title":"Training data-efficient image transformers & distillation through attention","author":"touvron","year":"2020","journal-title":"arXiv 2012 12877"},{"key":"ref33","first-page":"4171","article-title":"BERT: Pre-training of deep bidirectional transformers for language understanding","volume":"1","author":"devlin","year":"2019","journal-title":"Proc Conf North Amer Chapter Assoc Comput Linguistics Hum Lang Technol"},{"key":"ref32","first-page":"2","article-title":"Neural machine translation by jointly learning to align and translate","author":"bahdanau","year":"2015","journal-title":"Proc 3rd Int Conf Learn Represent (ICLR)"},{"key":"ref31","first-page":"5998","article-title":"Attention is all you need","volume":"30","author":"vaswani","year":"2017","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.3390\/s20010214"},{"key":"ref37","article-title":"X-BERT: Extreme multi-label text classification with using bidirectional encoder representations from transformers","author":"chang","year":"2019","journal-title":"arXiv 1905 02331"},{"key":"ref36","article-title":"An image is worth 16x16 words: Transformers for image recognition at scale","author":"dosovitskiy","year":"2020","journal-title":"arXiv 2010 11929"},{"key":"ref35","first-page":"213","article-title":"End-to-end object detection with transformers","author":"carion","year":"2020","journal-title":"Vision Computer"},{"key":"ref34","first-page":"9","article-title":"Language models are unsupervised multitask learners","volume":"1","author":"radford","year":"2019","journal-title":"OpenAIRE blog"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.3390\/s20092653"},{"key":"ref40","first-page":"8026","article-title":"Pytorch: An imperative style, high-performance deep learning library","volume":"32","author":"paszke","year":"2019","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1088\/0957-0233\/20\/1\/015203"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/IGARSS.2017.8128396"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1016\/j.patrec.2018.02.010"},{"key":"ref14","first-page":"1","article-title":"Activity recognition using inertial sensing for healthcare, wellbeing and sports applications: A survey","author":"avci","year":"2010","journal-title":"Proc 23th Int Conf Archit Comput Syst"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/TSMCC.2012.2198883"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2020.107561"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.3390\/s150102059"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1002\/widm.1254"},{"key":"ref19","doi-asserted-by":"crossref","first-page":"110","DOI":"10.3390\/app7010110","article-title":"A comprehensive review on handcrafted and learning-based action representation approaches for human activity recognition","volume":"7","author":"sargano","year":"2017","journal-title":"Appl Sci"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.2993534"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.3390\/rs11091068"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.3390\/s19040840"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2019.2948102"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/TSMCA.2011.2116004"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/JSEN.2021.3053843"},{"key":"ref5","first-page":"2737","article-title":"Human activity recognition for video surveillance","author":"lin","year":"2008","journal-title":"Proc IEEE Int Symp Circuits Syst"},{"key":"ref8","article-title":"Deep, convolutional, and recurrent models for human activity recognition using wearables","author":"hammerla","year":"2016","journal-title":"arXiv 1604 08880"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/JSEN.2011.2166953"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/JSAC.2017.2679658"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/BSN.2015.7299406"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW.2011.6130379"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.3037715"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1002\/navi.120"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.3390\/info7040072"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1145\/3302505.3310068"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/JSEN.2017.2737825"},{"key":"ref41","first-page":"621","article-title":"RIDI: Robust IMU double integration","author":"yan","year":"2018","journal-title":"Proc Eur Conf Comput Vis (ECCV)"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/JSEN.2018.2861395"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1145\/3195258.3195260"},{"key":"ref26","first-page":"1","article-title":"IONet: Learning to cure the curse of drift in inertial odometry","volume":"32","author":"chen","year":"2018","journal-title":"Proc AAAI Conf Artif Intell"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.3390\/s140610146"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2020.2985082"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/6287639\/9312710\/09393889.pdf?arnumber=9393889","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,1,26]],"date-time":"2022-01-26T17:37:00Z","timestamp":1643218620000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9393889\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"references-count":44,"URL":"https:\/\/doi.org\/10.1109\/access.2021.3070646","relation":{},"ISSN":["2169-3536"],"issn-type":[{"value":"2169-3536","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021]]}}}