{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,4,1]],"date-time":"2025-04-01T09:45:24Z","timestamp":1743500724296,"version":"3.37.3"},"reference-count":43,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/legalcode"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61971382"],"award-info":[{"award-number":["61971382"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2021]]},"DOI":"10.1109\/access.2021.3082694","type":"journal-article","created":{"date-parts":[[2021,5,21]],"date-time":"2021-05-21T15:33:43Z","timestamp":1621611223000},"page":"76224-76234","source":"Crossref","is-referenced-by-count":10,"title":["Saliency Maps-Based Convolutional Neural Networks for Facial Expression Recognition"],"prefix":"10.1109","volume":"9","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2710-0410","authenticated-orcid":false,"given":"Qinglan","family":"Wei","sequence":"first","affiliation":[]}],"member":"263","reference":[{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2019.2916063"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00233"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2016.09.072"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/ICPR.2016.7900174"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.71"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1017\/S0952523813000370"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2016.12.007"},{"key":"ref36","first-page":"3823","article-title":"Variational laws of visual attention for dynamic scenes","author":"zanca","year":"2017","journal-title":"Proc NIPS"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.38"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1016\/S0893-6080(98)00116-6"},{"key":"ref10","article-title":"CAT2000: A large scale fifixation dataset for boosting saliency research","author":"borji","year":"2015","journal-title":"arXiv 1505 03581"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1016\/j.patrec.2014.10.004"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2010.5543262"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2014.06.002"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1007\/978-94-009-3833-5_5"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/34.730558"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP.2003.1246946"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1167\/8.7.32"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2017.2710620"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2018.2851672"},{"key":"ref19","article-title":"EML-NET: An expandable multi-layer network for saliency prediction","author":"jia","year":"2018","journal-title":"arXiv 1805 01047"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2017.2784096"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-019-01630-9"},{"key":"ref27","first-page":"657","article-title":"Facial expression recognition with deep age","author":"luo","year":"2017","journal-title":"Proc IEEE Int Conf Multimedia Expo Workshops (ICMEW)"},{"key":"ref3","first-page":"1097","article-title":"ImageNet classification with deep convolutional neural networks","volume":"25","author":"krizhevsky","year":"2012","journal-title":"Proc NIPS"},{"key":"ref6","first-page":"839","article-title":"VEGAC: Visual saliency-based age, gender, and facial expression classification using convolutional neural networks","author":"gurnani","year":"2019","journal-title":"Proc WACV"},{"key":"ref29","article-title":"Attended end-to-end architecture for age estimation from facial expression videos","author":"pei","year":"2017","journal-title":"arXiv 1711 08690"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW.2017.327"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2017.282"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/ICMLA.2016.0147"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1037\/0003-066X.48.4.384"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2019.2948288"},{"key":"ref1","first-page":"146","article-title":"The expression of the emotions in man and animals","volume":"123","author":"darwin","year":"2012","journal-title":"Portable Darwin"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2020.05.004"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2006.244"},{"key":"ref21","first-page":"3087","article-title":"Application of prior know ledge in deep learning","volume":"36","author":"xuan","year":"2015","journal-title":"Eng Comput"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1007\/s10044-017-0645-4"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2005.177"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1109\/FG.2017.108"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1023\/B:VISI.0000029664.99615.94"},{"key":"ref26","article-title":"Convolutional neural networks for facial expression recognition","author":"alizadeh","year":"2017","journal-title":"arXiv 1704 06756"},{"key":"ref43","first-page":"23031.1","article-title":"Spontaneous smile intensity estimation by fusing saliency maps and convolutional neural networks","volume":"28","author":"wei","year":"2019","journal-title":"J Electron Imag"},{"key":"ref25","first-page":"139","article-title":"Facial expression recognition using a hybrid CNN-SIFT aggregator","author":"al-shabi","year":"2016","journal-title":"Proc Int Workshop Multi-Disciplinary Trends Artif Intell"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/6287639\/9312710\/09438697.pdf?arnumber=9438697","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,12,17]],"date-time":"2021-12-17T19:56:15Z","timestamp":1639770975000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9438697\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"references-count":43,"URL":"https:\/\/doi.org\/10.1109\/access.2021.3082694","relation":{},"ISSN":["2169-3536"],"issn-type":[{"type":"electronic","value":"2169-3536"}],"subject":[],"published":{"date-parts":[[2021]]}}}