{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T17:57:44Z","timestamp":1775066264925,"version":"3.50.1"},"reference-count":30,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2024,10,1]],"date-time":"2024-10-01T00:00:00Z","timestamp":1727740800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,10,1]],"date-time":"2024-10-01T00:00:00Z","timestamp":1727740800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Wireless Pers Commun"],"published-print":{"date-parts":[[2024,10]]},"DOI":"10.1007\/s11277-024-11564-8","type":"journal-article","created":{"date-parts":[[2024,10,15]],"date-time":"2024-10-15T14:53:59Z","timestamp":1729004039000},"page":"2183-2205","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":9,"title":["Automatic Facial Expression Recognition Using Modified LPQ and HOG Features with Stacked Deep Convolutional Autoencoders"],"prefix":"10.1007","volume":"138","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-6922-8095","authenticated-orcid":false,"given":"H. N.","family":"Naveen Kumar","sequence":"first","affiliation":[]},{"given":"Chandrashekar M.","family":"Patil","sequence":"additional","affiliation":[]},{"given":"B. G.","family":"Nagaraja","sequence":"additional","affiliation":[]},{"given":"Amith K.","family":"Jain","sequence":"additional","affiliation":[]},{"given":"K. V.","family":"Sudheesh","sequence":"additional","affiliation":[]},{"given":"S","family":"Mahadevaswamy","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,10,15]]},"reference":[{"key":"11564_CR1","doi-asserted-by":"publisher","DOI":"10.3389\/frobt.2020.532279","volume":"7","author":"M Spezialetti","year":"2020","unstructured":"Spezialetti, M., Placidi, G., & Rossi, S. (2020). Emotion recognition for human- robot interaction: Recent advances and future perspectives. Frontiers in Robotics and AI, 7, 532279.","journal-title":"Frontiers in Robotics and AI"},{"issue":"1","key":"11564_CR2","doi-asserted-by":"publisher","first-page":"26","DOI":"10.1016\/j.cdtm.2021.07.002","volume":"8","author":"R Hooda","year":"2022","unstructured":"Hooda, R., Joshi, V., & Shah, M. A. (2022). comprehensive review of approaches to detect fatigue using machine learning techniques. Chronic Diseases and Translational Medicine, 8(1), 26\u201335.","journal-title":"Chronic Diseases and Translational Medicine"},{"issue":"1","key":"11564_CR3","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.vrih.2020.10.003","volume":"3","author":"H Pan","year":"2021","unstructured":"Pan, H., Xie, L., Wang, Z., Liu, B., Yang, M., & Tao, J. (2021). Review of micro- expression spotting and recognition in video sequences. Virtual Reality & Intelligent Hardware, 3(1), 1\u20137.","journal-title":"Virtual Reality & Intelligent Hardware"},{"key":"11564_CR4","doi-asserted-by":"publisher","first-page":"544","DOI":"10.1016\/j.procs.2019.11.154","volume":"161","author":"P Utami","year":"2019","unstructured":"Utami, P., Hartanto, R., & Soesanti, I. (2019). A study on facial expression recognition in assessing teaching skills: Datasets and methods. Procedia Computer Science, 161, 544\u2013552.","journal-title":"Procedia Computer Science"},{"issue":"3","key":"11564_CR5","doi-asserted-by":"publisher","first-page":"128","DOI":"10.3390\/info11030128","volume":"11","author":"M Leo","year":"2020","unstructured":"Leo, M., Carcagn, P., Mazzeo, P. L., Spagnolo, P., Cazzato, D., & Distante, C. (2020). Analysis of facial information for healthcare applications: A survey on computer vision-based approaches. Information, 11(3), 128.","journal-title":"Information"},{"key":"11564_CR6","doi-asserted-by":"publisher","first-page":"104887","DOI":"10.1109\/ACCESS.2022.3208587","volume":"10","author":"M Kohli","year":"2022","unstructured":"Kohli, M., Kar, A. K., & Sinha, S. (2022). The role of intelligent technologies in early detection of autism spectrum disorder (asd): A scoping review. IEEE Access, 10, 104887\u2013104913.","journal-title":"IEEE Access"},{"issue":"4","key":"11564_CR7","doi-asserted-by":"publisher","first-page":"1111","DOI":"10.1049\/ipr2.12700","volume":"17","author":"KHN Naveen","year":"2023","unstructured":"Naveen, K. H. N., Kumar, A. S., Prasad, M. S. G., & Shah, M. A. (2023). Automatic facial expression recognition combining texture and shape features from prominent facial regions. IET Image Processing, 17(4), 1111\u20131125.","journal-title":"IET Image Processing"},{"key":"11564_CR8","unstructured":"HN, N. K., Patil, C. M., Jain, A. K., & Sudheesh, K. V. (2022). A Comprehensive Study on Geometric, Appearance, and Deep Feature based Methods for Automatic Facial Expression Recognition. In 2022 Fourth International Conference on Cognitive Computing and Information Processing (CCIP) (pp. 1-6). IEEE."},{"issue":"3","key":"11564_CR9","doi-asserted-by":"publisher","first-page":"1195","DOI":"10.1109\/TAFFC.2020.2981446","volume":"13","author":"S Li","year":"2020","unstructured":"Li, S., & Deng, W. (2020). Deep facial expression recognition: A survey. IEEE Transactions on Affective Computing, 13(3), 1195\u20131215.","journal-title":"IEEE Transactions on Affective Computing"},{"key":"11564_CR10","doi-asserted-by":"crossref","unstructured":"Lyons, M., Akamatsu, S., Kamachi, M., & Gyoba, J. (1998). Coding facial expressions with gabor wavelets. In Proceedings Third IEEE international conference on automatic face and gesture recognition (pp. 200-205). IEEE.","DOI":"10.1109\/AFGR.1998.670949"},{"key":"11564_CR11","doi-asserted-by":"crossref","unstructured":"Lucey, P., Cohn, J. F., Kanade, T., Saragih, J., Ambadar, Z., & Matthews, I. (2010). The extended cohn-kanade dataset (ck+): A complete dataset for action unit and emotion-specified expression. In 2010 ieee computer society conference on computer vision and pattern recognition-workshops (pp. 94-101). IEEE.","DOI":"10.1109\/CVPRW.2010.5543262"},{"key":"11564_CR12","doi-asserted-by":"crossref","unstructured":"Lundqvist, D., & Litton, J. E. (1998). The averaged Karolinska directed emotional faces. Stockholm: Karolinska Institute, Department of Clinical Neuroscience, Section Psychology.","DOI":"10.1037\/t27732-000"},{"key":"11564_CR13","doi-asserted-by":"crossref","unstructured":"Liu M., Li S., Shan S., Wang R., Chen X.: Deeply learning deformable facial action parts model for dynamic expression analysis. In Asian conference on computer vision, pp. 143\u2013157, (2014).","DOI":"10.1007\/978-3-319-16817-3_10"},{"key":"11564_CR14","doi-asserted-by":"crossref","unstructured":"Mollahosseini, A., Chan, D., & Mahoor, M. H. (2016). Going deeper in facial expression recognition using deep neural networks. In 2016 IEEE Winter conference on applications of computer vision (WACV) (pp. 1-10). IEEE.","DOI":"10.1109\/WACV.2016.7477450"},{"issue":"2","key":"11564_CR15","first-page":"167","volume":"10","author":"K Bailly","year":"2017","unstructured":"Bailly, K., & Dubuisson, S. (2017). Dynamic pose-robust facial expression recognition by multi-view pairwise conditional random forests. IEEE Transactions on Affective Computing, 10(2), 167\u2013181.","journal-title":"IEEE Transactions on Affective Computing"},{"key":"11564_CR16","doi-asserted-by":"crossref","unstructured":"Haque, M. I. U., & Valles, D. (2019). Facial expression recognition using dcnn and development of an ios app for children with asd to enhance communication abilities. In 2019 IEEE 10th Annual Ubiquitous Computing, Electronics & Mobile Communication Conference (UEMCON) (pp. 0476-0482). IEEE.","DOI":"10.1109\/UEMCON47517.2019.8993051"},{"key":"11564_CR17","doi-asserted-by":"crossref","unstructured":"Zhou, Y., & Shi, B. E. (2017). Action unit selective feature maps in deep networks for facial expression recognition. In 2017 international joint conference on neural networks (IJCNN) (pp. 2031-2038). IEEE.","DOI":"10.1109\/IJCNN.2017.7966100"},{"issue":"08","key":"11564_CR18","doi-asserted-by":"publisher","first-page":"1850121","DOI":"10.1142\/S0218126618501219","volume":"27","author":"Z Sun","year":"2018","unstructured":"Sun, Z., Hu, Z. P., Chiong, R., Wang, M., & He, W. (2018). Combining the kernel collaboration representation and deep subspace learning for facial expression recognition. Journal of Circuits Systems and Computers, 27(08), 1850121.","journal-title":"Journal of Circuits Systems and Computers"},{"key":"11564_CR19","doi-asserted-by":"crossref","unstructured":"Liu, Y., Zeng, J., Shan, S., & Zheng, Z. (2018, May). Multi-channel pose-aware convolution neural networks for multi-view facial expression recognition. In 2018 13th IEEE international conference on automatic face & gesture recognition (FG 2018) (pp. 458-465). IEEE.","DOI":"10.1109\/FG.2018.00074"},{"key":"11564_CR20","doi-asserted-by":"publisher","first-page":"353","DOI":"10.1016\/j.neucom.2021.07.079","volume":"462","author":"X Jin","year":"2021","unstructured":"Jin, X., & Jin, Z. (2021). MiniExpNet: A small and effective facial expression recognition network based on facial local regions. Neurocomputing, 462, 353\u2013364.","journal-title":"Neurocomputing"},{"issue":"7","key":"11564_CR21","doi-asserted-by":"publisher","first-page":"2005","DOI":"10.1049\/ipr2.12743","volume":"17","author":"W Zhang","year":"2023","unstructured":"Zhang, W., Zhang, X., & Tang, Y. (2023). Facial expression recognition based on improved residual network. IET Image Processing, 17(7), 2005\u20132014.","journal-title":"IET Image Processing"},{"key":"11564_CR22","doi-asserted-by":"publisher","first-page":"195","DOI":"10.1016\/j.ins.2021.07.034","volume":"578","author":"Y Liu","year":"2021","unstructured":"Liu, Y., Dai, W., Fang, F., Chen, Y., Huang, R., Wang, R., & Wan, Bo. (2021). Dynamic multi-channel metric network for joint pose-aware and identity-invariant facial expression recognition. Information Sciences, 578, 195\u2013213.","journal-title":"Information Sciences"},{"key":"11564_CR23","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2020.106124","volume":"204","author":"Z Sun","year":"2020","unstructured":"Sun, Z., Chiong, R., & Zheng-ping, Hu. (2020). Self-adaptive feature learning based on a priori knowledge for facial expression recognition. Knowledge-Based Systems, 204, 106124.","journal-title":"Knowledge-Based Systems"},{"key":"11564_CR24","doi-asserted-by":"publisher","first-page":"1556","DOI":"10.1109\/TNSRE.2024.3385336","volume":"32","author":"S Hossain","year":"2024","unstructured":"Hossain, S., Umer, S., Rout, R. K., & Marzouqi, H. A. (2024). A Deep quantum convolutional neural network based facial expression recognition for mental health analysis. IEEE Transactions on Neural Systems and Rehabilitation Engineering, 32, 1556\u20131565.","journal-title":"IEEE Transactions on Neural Systems and Rehabilitation Engineering"},{"issue":"2","key":"11564_CR25","doi-asserted-by":"publisher","first-page":"137","DOI":"10.1023\/B:VISI.0000013087.49260.fb","volume":"57","author":"P Viola","year":"2004","unstructured":"Viola, P., & Jones, M. J. (2004). Robust real-time face detection. International journal of computer vision, 57(2), 137\u2013154.","journal-title":"International journal of computer vision"},{"key":"11564_CR26","doi-asserted-by":"crossref","unstructured":"HN, N. K., Jagadeesha, S., & Jain, A. K. (2016, July). Human Facial Expression Recognition from static images using shape and appearance feature. In 2016 2nd international conference on applied and theoretical computing and communication technology (iCATccT) (pp. 598-603). IEEE.","DOI":"10.1109\/ICATCCT.2016.7912070"},{"issue":"12","key":"11564_CR27","doi-asserted-by":"publisher","first-page":"2037","DOI":"10.1109\/TPAMI.2006.244","volume":"28","author":"T Ahonen","year":"2006","unstructured":"Ahonen, T., Hadid, A., & Pietikainen, M. (2006). Face description with local binary patterns: Application to face recognition. IEEE transactions on pattern analysis and machine intelligence, 28(12), 2037\u20132041.","journal-title":"IEEE transactions on pattern analysis and machine intelligence"},{"key":"11564_CR28","doi-asserted-by":"crossref","unstructured":"Ojansivu, V., & Heikkil\u00e4, J. (2008). Blur insensitive texture classification using local phase quantization. In Image and Signal Processing: 3rd International Conference, ICISP 2008. Cherbourg-Octeville, France, July 1-3, 2008. Proceedings 3 (pp. 236-243). Springer Berlin Heidelberg.","DOI":"10.1007\/978-3-540-69905-7_27"},{"key":"11564_CR29","doi-asserted-by":"publisher","first-page":"331","DOI":"10.1016\/j.jvcir.2018.05.024","volume":"55","author":"C Turan","year":"2018","unstructured":"Turan, C., & Lam, K. M. (2018). Histogram-based local descriptors for facial expression recognition (FER): A comprehensive study. Journal of visual communication and image representation, 55, 331\u2013341.","journal-title":"Journal of visual communication and image representation"},{"key":"11564_CR30","doi-asserted-by":"publisher","first-page":"4557","DOI":"10.1007\/s10489-020-01813-1","volume":"50","author":"X Li","year":"2020","unstructured":"Li, X., Zhang, T., Zhao, X., & Yi, Z. (2020). Guided autoencoder for dimensionality reduction of pedestrian features. Applied Intelligence, 50, 4557\u20134567.","journal-title":"Applied Intelligence"}],"container-title":["Wireless Personal Communications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11277-024-11564-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11277-024-11564-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11277-024-11564-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,28]],"date-time":"2024-10-28T14:13:23Z","timestamp":1730124803000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11277-024-11564-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10]]},"references-count":30,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2024,10]]}},"alternative-id":["11564"],"URL":"https:\/\/doi.org\/10.1007\/s11277-024-11564-8","relation":{},"ISSN":["0929-6212","1572-834X"],"issn-type":[{"value":"0929-6212","type":"print"},{"value":"1572-834X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,10]]},"assertion":[{"value":"2 September 2024","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"15 October 2024","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"On behalf of all authors, I consent to publish our manuscript.<b>Ethical Approval<\/b>Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to Publish"}}]}}