{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,6]],"date-time":"2026-05-06T16:43:46Z","timestamp":1778085826267,"version":"3.51.4"},"reference-count":75,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"4","license":[{"start":{"date-parts":[[2022,10,1]],"date-time":"2022-10-01T00:00:00Z","timestamp":1664582400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2022,10,1]],"date-time":"2022-10-01T00:00:00Z","timestamp":1664582400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2022,10,1]],"date-time":"2022-10-01T00:00:00Z","timestamp":1664582400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"name":"China Scholarship Council &#x2013; University of St Andrews Scholarships","award":["201908060250"],"award-info":[{"award-number":["201908060250"]}]},{"name":"National Key Research and Development Project of China","award":["2019YFB1312000"],"award-info":[{"award-number":["2019YFB1312000"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62076195"],"award-info":[{"award-number":["62076195"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012226","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"publisher","award":["AUGA5710011522"],"award-info":[{"award-number":["AUGA5710011522"]}],"id":[{"id":"10.13039\/501100012226","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Affective Comput."],"published-print":{"date-parts":[[2022,10,1]]},"DOI":"10.1109\/taffc.2022.3213509","type":"journal-article","created":{"date-parts":[[2022,10,10]],"date-time":"2022-10-10T20:18:08Z","timestamp":1665433088000},"page":"1973-1985","source":"Crossref","is-referenced-by-count":140,"title":["Short and Long Range Relation Based Spatio-Temporal Transformer for Micro-Expression Recognition"],"prefix":"10.1109","volume":"13","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-1860-5281","authenticated-orcid":false,"given":"Liangfei","family":"Zhang","sequence":"first","affiliation":[{"name":"School of Computer Science, University of St Andrews, St Andrews, U.K."}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0611-0636","authenticated-orcid":false,"given":"Xiaopeng","family":"Hong","sequence":"additional","affiliation":[{"name":"Harbin Institute of Technology, Harbin, Heilongjiang, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9314-194X","authenticated-orcid":false,"given":"Ognjen","family":"Arandjelovic","sequence":"additional","affiliation":[{"name":"School of Computer Science, University of St Andrews, St Andrews, U.K."}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3694-206X","authenticated-orcid":false,"given":"Guoying","family":"Zhao","sequence":"additional","affiliation":[{"name":"University of Oulu, Oulu, Finland"}]}],"member":"263","reference":[{"key":"ref73","doi-asserted-by":"publisher","DOI":"10.1109\/FG.2019.8756579"},{"key":"ref72","doi-asserted-by":"publisher","DOI":"10.1109\/FG.2019.8756544"},{"key":"ref71","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2020.3018222"},{"key":"ref70","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2020.10.082"},{"key":"ref74","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2022.3174895"},{"key":"ref39","first-page":"325","article-title":"Micro-expression recognition using robust principal component analysis and local spatiotemporal directional features","volume":"8925","author":"wang","year":"2014","journal-title":"Proc Eur Conf Comput Vis"},{"key":"ref75","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2022.3182342"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1145\/3476100.3484462"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2020.3035042"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1016\/j.image.2019.02.005"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1016\/j.image.2017.11.006"},{"key":"ref30","first-page":"5826","article-title":"Video-based facial micro-expression analysis: A survey of datasets, features and algorithms","volume":"44","author":"ben","year":"2022","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"ref37","author":"ekman","year":"2002","journal-title":"Facial Action Coding System Investigator s Guide"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2016.2573832"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0086041"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/FG.2013.6553717"},{"key":"ref60","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2018.05.107"},{"key":"ref62","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"ref61","article-title":"RIFE: Real-time intermediate flow estimation for video frame interpolation","author":"zhewei","year":"2020"},{"key":"ref63","doi-asserted-by":"publisher","DOI":"10.1109\/FG.2019.8756611"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"ref64","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2015.10.096"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00033"},{"key":"ref65","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP.2019.8802965"},{"key":"ref66","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW.2015.10"},{"key":"ref29","article-title":"Deformable DETR: Deformable transformers for end-to-end object detection","author":"zhu","year":"2020"},{"key":"ref67","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2018.2820321"},{"key":"ref68","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2020.06.005"},{"key":"ref69","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2020.2986962"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/EMBC44109.2020.9176632"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1037\/h0030377"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW53098.2021.00173"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.01075"},{"key":"ref21","article-title":"Transformers in vision: A survey","author":"khan","year":"2021"},{"key":"ref24","article-title":"An image is worth 16x16 words: Transformers for image recognition at scale","author":"dosovitskiy","year":"2020"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00583"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00756"},{"key":"ref25","article-title":"Training data-efficient image transformers & distillation through attention","author":"touvron","year":"2020"},{"key":"ref50","first-page":"7680","article-title":"Facial action unit detection with transformers","author":"jacob","year":"2021","journal-title":"Proc IEEE Conf Comput Vis and Pattern Recog"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2021.3122146"},{"key":"ref59","first-page":"1755","article-title":"Dlib-ML: A machine learning toolkit","volume":"10","author":"king","year":"2009","journal-title":"J Mach Learn Res"},{"key":"ref58","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.241"},{"key":"ref57","article-title":"SGDR: Stochastic gradient descent with warm restarts","author":"loshchilov","year":"2017","journal-title":"Proc Int Conf Learn Representations"},{"key":"ref56","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"ref55","article-title":"Adaptive input representations for neural language modeling","author":"baevski","year":"2019","journal-title":"Proc Int Conf Learn Representations"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1176"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2015.2424055"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2014.2378034"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2015.2485205"},{"key":"ref11","first-page":"2258","article-title":"Selective deep features for micro-expression recognition","author":"patel","year":"2016","journal-title":"Proc 23rd Int Conf Pattern Recognit"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1109\/ACII.2019.8925525"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/FG.2018.00105"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1007\/s10044-018-0757-5"},{"key":"ref14","first-page":"1","article-title":"Spontaneous facial micro-expression recognition via deep convolutional network","author":"xia","year":"2019","journal-title":"Proc IEEE 8th Int Conf Image Process Theory Tools Appl"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2019.2931351"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/MIPR49039.2020.00023"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.3390\/jimaging6120130"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1145\/3394171.3414012"},{"key":"ref19","first-page":"1511","article-title":"Micro-expression classification based on landmark relations with graph attention convolutional network","author":"kumar","year":"2021","journal-title":"Proc IEEE\/CVF Conf Comput Vis Pattern Recognit Workshops"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1080\/00332747.1969.11023575"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4684-6045-2"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2011.6126401"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.3390\/make3020021"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1049\/ic.2009.0244"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2007.1110"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1109\/BIBM52615.2021.9669556"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2017.2667642"},{"key":"ref46","article-title":"BERT: Pre-training of deep bidirectional transformers for language understanding","author":"devlin","year":"2018"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01212"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2020.3037496"},{"key":"ref47","first-page":"5998","article-title":"Attention is all you need","author":"vaswani","year":"2017","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref42","first-page":"1","article-title":"A neural micro-expression recognizer","author":"liu","year":"2019","journal-title":"Proc 14th IEEE Int Conf Autom Face Gesture Recognit"},{"key":"ref41","first-page":"1","article-title":"Shallow triple stream three-dimensional CNN (STSTNet) for micro-expression recognition","author":"liong","year":"2019","journal-title":"Proc IEEE Int Conf Autom Face Gesture Recognit"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1145\/2964284.2967247"},{"key":"ref43","first-page":"1","article-title":"Spontaneous facial micro-expression recognition using 3D spatiotemporal convolutional neural networks","author":"reddy","year":"2019","journal-title":"Proc Int Joint Conf Neural Netw"}],"container-title":["IEEE Transactions on Affective Computing"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/5165369\/9964459\/09915457.pdf?arnumber=9915457","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,12,19]],"date-time":"2022-12-19T19:40:20Z","timestamp":1671478820000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9915457\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,10,1]]},"references-count":75,"journal-issue":{"issue":"4"},"URL":"https:\/\/doi.org\/10.1109\/taffc.2022.3213509","relation":{},"ISSN":["1949-3045","2371-9850"],"issn-type":[{"value":"1949-3045","type":"electronic"},{"value":"2371-9850","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,10,1]]}}}