{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,19]],"date-time":"2026-03-19T23:57:35Z","timestamp":1773964655903,"version":"3.50.1"},"reference-count":30,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Knowledge-Based Systems"],"published-print":{"date-parts":[[2026,4]]},"DOI":"10.1016\/j.knosys.2026.115589","type":"journal-article","created":{"date-parts":[[2026,2,20]],"date-time":"2026-02-20T07:48:05Z","timestamp":1771573685000},"page":"115589","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["Lightweight adaptive AI for novel real-time facial expression recognition"],"prefix":"10.1016","volume":"339","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-7694-1806","authenticated-orcid":false,"given":"Sabina","family":"Umirzakova","sequence":"first","affiliation":[]},{"given":"Jushkin","family":"Baltayev","sequence":"additional","affiliation":[]},{"given":"Sevara","family":"Mardieva","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6223-4502","authenticated-orcid":false,"given":"Shakhnoza","family":"Muksimova","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.knosys.2026.115589_bib0002","doi-asserted-by":"crossref","first-page":"215","DOI":"10.1016\/j.neucom.2020.10.081","article-title":"Deep face recognition: a survey","volume":"429","author":"Wang","year":"2021","journal-title":"Neurocomputing"},{"issue":"6","key":"10.1016\/j.knosys.2026.115589_bib0003","doi-asserted-by":"crossref","first-page":"268","DOI":"10.3390\/info13060268","article-title":"Facial emotion recognition using conventional machine learning and deep learning methods: current achievements, analysis and remaining challenges","volume":"13","author":"Khan","year":"2022","journal-title":"Information"},{"key":"10.1016\/j.knosys.2026.115589_bib0004","series-title":"Soft Comput.","first-page":"1","article-title":"Real-time facial emotion recognition model based on kernel autoencoder and convolutional neural network for autism children","author":"Talaat","year":"2024"},{"key":"10.1016\/j.knosys.2026.115589_bib0005","series-title":"World Conference Intelligent System for Industrial Automation","first-page":"105","article-title":"A new type of architecture for neural networks with multi-connected weights in classification problems","author":"Rakhimovich","year":"2022"},{"issue":"3","key":"10.1016\/j.knosys.2026.115589_bib0008","doi-asserted-by":"crossref","first-page":"135","DOI":"10.3390\/info15030135","article-title":"Advances in facial expression recognition: a survey of methods, benchmarks, models, and datasets","volume":"15","author":"Kopalidis","year":"2024","journal-title":"Information"},{"issue":"2","key":"10.1016\/j.knosys.2026.115589_bib0012","first-page":"546","article-title":"Facial expression recognition with local binary patterns and linear programming","volume":"15","author":"Feng","year":"2005","journal-title":"Pattern Recognit. Image Anal. C\/C Raspoznavaniye Obraz. I Anal. Izobr."},{"key":"10.1016\/j.knosys.2026.115589_bib0013","series-title":"18th International Conference on Pattern Recognition (ICPR'06)","first-page":"536","article-title":"Facial expression recognition based on fusion of multiple Gabor features","volume":"3","author":"Liu","year":"2006"},{"issue":"24","key":"10.1016\/j.knosys.2026.115589_bib0014","doi-asserted-by":"crossref","first-page":"1","DOI":"10.17485\/ijst\/2019\/v12i24\/145093","article-title":"Facial expression recognition with histogram of oriented gradients using CNN","volume":"12","author":"Jumani","year":"2019","journal-title":"Indian J. Sci. Technol."},{"issue":"1","key":"10.1016\/j.knosys.2026.115589_bib0015","doi-asserted-by":"crossref","first-page":"35","DOI":"10.5829\/ije.2025.38.01a.04","article-title":"EmoSense: pioneering facial emotion recognition with precision through model optimization and face emotion constraints","volume":"38","author":"Kasar","year":"2025","journal-title":"Int. J. Eng."},{"key":"10.1016\/j.knosys.2026.115589_bib0016","article-title":"A hierarchical cross-modal spatial fusion network for multimodal emotion recognition","author":"Xu","year":"2025","journal-title":"IEEE Trans. Artif. Intell."},{"key":"10.1016\/j.knosys.2026.115589_bib0017","series-title":"World Conference Intelligent System for Industrial Automation","first-page":"105","article-title":"A new type of architecture for neural networks with multi-connected weights in classification problems","author":"Rakhimovich","year":"2022"},{"key":"10.1016\/j.knosys.2026.115589_bib0018","doi-asserted-by":"crossref","DOI":"10.1109\/TAFFC.2025.3535811","article-title":"CSE-GResNet: a simple and highly efficient network for facial expression recognition","author":"Jiang","year":"2025","journal-title":"IEEE Trans. Affect. Comput."},{"key":"10.1016\/j.knosys.2026.115589_bib0019","doi-asserted-by":"crossref","first-page":"26146","DOI":"10.1109\/ACCESS.2017.2777003","article-title":"Facial expression recognition using salient features and convolutional neural network","volume":"5","author":"Uddin","year":"2017","journal-title":"IEEE Access"},{"key":"10.1016\/j.knosys.2026.115589_bib0020","doi-asserted-by":"crossref","first-page":"6544","DOI":"10.1109\/TIP.2021.3093397","article-title":"Learning deep global multi-scale and local attention features for facial expression recognition in the wild","volume":"30","author":"Zhao","year":"2021","journal-title":"IEEE Trans. Image Process."},{"issue":"2","key":"10.1016\/j.knosys.2026.115589_bib0021","doi-asserted-by":"crossref","first-page":"199","DOI":"10.3390\/biomimetics8020199","article-title":"Distract your attention: multi-head cross attention network for facial expression recognition","volume":"8","author":"Wen","year":"2023","journal-title":"Biomimetics"},{"key":"10.1016\/j.knosys.2026.115589_bib0022","doi-asserted-by":"crossref","first-page":"781","DOI":"10.1016\/j.ins.2022.11.068","article-title":"Patch attention convolutional vision transformer for facial expression recognition with occlusion","volume":"619","author":"Liu","year":"2023","journal-title":"Inf. Sci."},{"key":"10.1016\/j.knosys.2026.115589_bib0009","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1109\/TIM.2020.3031835","article-title":"Facial expression recognition using local gravitational force descriptor-based deep convolution neural networks","volume":"70","author":"Mohan","year":"2020","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"15","key":"10.1016\/j.knosys.2026.115589_bib0011","doi-asserted-by":"crossref","first-page":"9125","DOI":"10.1007\/s00521-020-05676-y","article-title":"FER-net: facial expression recognition using deep neural net","volume":"33","author":"Mohan","year":"2021","journal-title":"Neural Comput. Appl."},{"key":"10.1016\/j.knosys.2026.115589_bib0025","article-title":"Poster++: a simpler and stronger facial expression recognition network","author":"Mao","year":"2024","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.knosys.2026.115589_bib0026","first-page":"1","article-title":"Hybrid attention-aware learning network for facial expression recognition in the wild","author":"Gong","year":"2024","journal-title":"Arab. J. Sci. Eng."},{"key":"10.1016\/j.knosys.2026.115589_bib0031","article-title":"MMATrans: muscle movement aware representation learning for facial expression recognition via transformers","author":"Liu","year":"2024","journal-title":"IEEE Trans. Ind. Inform."},{"key":"10.1016\/j.knosys.2026.115589_bib0032","article-title":"FER-former: multimodal transformer for facial expression recognition","author":"Li","year":"2024","journal-title":"IEEE Trans. Multimed."},{"key":"10.1016\/j.knosys.2026.115589_bib0033","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2024.123635","article-title":"Multi-geometry embedded transformer for facial expression recognition in videos","volume":"249","author":"Chen","year":"2024","journal-title":"Expert Syst. Appl."},{"issue":"3","key":"10.1016\/j.knosys.2026.115589_bib0023","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3587038","article-title":"Edge-AI-driven framework with efficient mobile network design for facial expression recognition","volume":"22","author":"Wu","year":"2023","journal-title":"ACM Trans. Embed. Comput. Syst."},{"issue":"8","key":"10.1016\/j.knosys.2026.115589_bib0024","doi-asserted-by":"crossref","first-page":"3848","DOI":"10.1109\/TCSVT.2023.3234312","article-title":"Multi-relations aware network for in-the-wild facial expression recognition","volume":"33","author":"Chen","year":"2023","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10.1016\/j.knosys.2026.115589_bib0027","series-title":"CF-DAN: Facial-Expression Recognition Based on Cross-Fusion Dual-Attention Network","first-page":"1","author":"Zhang","year":"2024"},{"key":"10.1016\/j.knosys.2026.115589_bib0028","doi-asserted-by":"crossref","DOI":"10.1016\/j.engappai.2024.108535","article-title":"Dynamic facial expression recognition based on spatial key-points optimized region feature fusion and temporal self-attention","volume":"133","author":"Huang","year":"2024","journal-title":"Eng. Appl. Artif. Intell."},{"issue":"9","key":"10.1016\/j.knosys.2026.115589_bib0029","doi-asserted-by":"crossref","first-page":"1606","DOI":"10.3390\/electronics13091606","article-title":"SlowR50-SA: a self-attention enhanced dynamic facial expression recognition model for tactile internet applications","volume":"13","author":"Neshov","year":"2024","journal-title":"Electronics"},{"issue":"4","key":"10.1016\/j.knosys.2026.115589_bib0030","doi-asserted-by":"crossref","first-page":"1535","DOI":"10.3390\/app14041535","article-title":"Context transformer and adaptive method with visual transformer for robust facial expression recognition","volume":"14","author":"Xiong","year":"2024","journal-title":"Appl. Sci."},{"key":"10.1016\/j.knosys.2026.115589_bib0007","doi-asserted-by":"crossref","first-page":"337","DOI":"10.1016\/j.neunet.2023.11.033","article-title":"Hierarchical attention network with progressive feature fusion for facial expression recognition","volume":"170","author":"Tao","year":"2024","journal-title":"Neural Netw."}],"container-title":["Knowledge-Based Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0950705126003291?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0950705126003291?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,3,19]],"date-time":"2026-03-19T21:31:50Z","timestamp":1773955910000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0950705126003291"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,4]]},"references-count":30,"alternative-id":["S0950705126003291"],"URL":"https:\/\/doi.org\/10.1016\/j.knosys.2026.115589","relation":{},"ISSN":["0950-7051"],"issn-type":[{"value":"0950-7051","type":"print"}],"subject":[],"published":{"date-parts":[[2026,4]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Lightweight adaptive AI for novel real-time facial expression recognition","name":"articletitle","label":"Article Title"},{"value":"Knowledge-Based Systems","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.knosys.2026.115589","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2026 Elsevier B.V. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"115589"}}