{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,27]],"date-time":"2025-10-27T20:52:45Z","timestamp":1761598365108,"version":"3.37.3"},"reference-count":24,"publisher":"Wiley","license":[{"start":{"date-parts":[[2021,8,19]],"date-time":"2021-08-19T00:00:00Z","timestamp":1629331200000},"content-version":"unspecified","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100004000","name":"Guangzhou Science and Technology Program key projects","doi-asserted-by":"publisher","award":["2019050001","JCYJ20190809145407809","TD2020E001","2019BT02C241","IRT_17R40","2017B030301007","201705030007"],"award-info":[{"award-number":["2019050001","JCYJ20190809145407809","TD2020E001","2019BT02C241","IRT_17R40","2017B030301007","201705030007"]}],"id":[{"id":"10.13039\/501100004000","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Shenzhen Science and Technology Innovation Committee","award":["2019050001","JCYJ20190809145407809","TD2020E001","2019BT02C241","IRT_17R40","2017B030301007","201705030007"],"award-info":[{"award-number":["2019050001","JCYJ20190809145407809","TD2020E001","2019BT02C241","IRT_17R40","2017B030301007","201705030007"]}]},{"name":"Shenzhen Institute of Information Technology School-level Innovative Scientific Research Team","award":["2019050001","JCYJ20190809145407809","TD2020E001","2019BT02C241","IRT_17R40","2017B030301007","201705030007"],"award-info":[{"award-number":["2019050001","JCYJ20190809145407809","TD2020E001","2019BT02C241","IRT_17R40","2017B030301007","201705030007"]}]},{"name":"Guangdong Innovative and Entrepreneurial Teams","award":["2019050001","JCYJ20190809145407809","TD2020E001","2019BT02C241","IRT_17R40","2017B030301007","201705030007"],"award-info":[{"award-number":["2019050001","JCYJ20190809145407809","TD2020E001","2019BT02C241","IRT_17R40","2017B030301007","201705030007"]}]},{"name":"Chang Jiang Scholars and Innovative Research Teams in Universities","award":["2019050001","JCYJ20190809145407809","TD2020E001","2019BT02C241","IRT_17R40","2017B030301007","201705030007"],"award-info":[{"award-number":["2019050001","JCYJ20190809145407809","TD2020E001","2019BT02C241","IRT_17R40","2017B030301007","201705030007"]}]},{"name":"Guangdong Provincial Key Laboratory of Optical Information Materials and Technology","award":["2019050001","JCYJ20190809145407809","TD2020E001","2019BT02C241","IRT_17R40","2017B030301007","201705030007"],"award-info":[{"award-number":["2019050001","JCYJ20190809145407809","TD2020E001","2019BT02C241","IRT_17R40","2017B030301007","201705030007"]}]},{"name":"Guangzhou Key Laboratory of Electronic Paper Displays Materials and Devices","award":["2019050001","JCYJ20190809145407809","TD2020E001","2019BT02C241","IRT_17R40","2017B030301007","201705030007"],"award-info":[{"award-number":["2019050001","JCYJ20190809145407809","TD2020E001","2019BT02C241","IRT_17R40","2017B030301007","201705030007"]}]},{"DOI":"10.13039\/501100013314","name":"111 Project","doi-asserted-by":"crossref","award":["2019050001","JCYJ20190809145407809","TD2020E001","2019BT02C241","IRT_17R40","2017B030301007","201705030007"],"award-info":[{"award-number":["2019050001","JCYJ20190809145407809","TD2020E001","2019BT02C241","IRT_17R40","2017B030301007","201705030007"]}],"id":[{"id":"10.13039\/501100013314","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Security and Communication Networks"],"published-print":{"date-parts":[[2021,8,19]]},"abstract":"<jats:p>Facial video big sensor data (BSD) is the core data of wireless sensor network industry application and technology research. It plays an important role in many industries, such as urban safety management, unmanned driving, senseless attendance, and venue management. The construction of video big sensor data security application and intelligent algorithm model has become a hot and difficult topic in related fields based on facial expression recognition. This paper focused on the experimental analysis of Cohn\u2013Kanade dataset plus (CK+) dataset with frontal pose and great clarity. Firstly, face alignment and the selection of peak image were utilized to preprocess the expression sequence. Then, the output vector from convolution network 1 and \u03b2-VAE were connected proportionally and input to support vector machine (SVM) classifier to complete facial expression recognition. The testing accuracy of the proposed model in CK + dataset can reach 99.615%. The number of expression sequences involved in training was 2417, and the number of expression sequences in testing was 519.<\/jats:p>","DOI":"10.1155\/2021\/9539022","type":"journal-article","created":{"date-parts":[[2021,8,20]],"date-time":"2021-08-20T22:30:38Z","timestamp":1629498638000},"page":"1-10","source":"Crossref","is-referenced-by-count":1,"title":["Real-Time Facial Expression Recognition System for Video Big Sensor Data Security Application"],"prefix":"10.1155","volume":"2021","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-4421-3804","authenticated-orcid":true,"given":"Zhi","family":"Yao","sequence":"first","affiliation":[{"name":"Guangdong Provincial Key Laboratory of Optical Information Materials and Technology & Institute of Electronic Paper Displays, South China Academy of Advanced Optoelectronics, South China Normal University, Guangzhou 510006, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7292-4611","authenticated-orcid":true,"given":"Hailing","family":"Sun","sequence":"additional","affiliation":[{"name":"Guangdong Provincial Key Laboratory of Optical Information Materials and Technology & Institute of Electronic Paper Displays, South China Academy of Advanced Optoelectronics, South China Normal University, Guangzhou 510006, China"}]},{"given":"Guofu","family":"Zhou","sequence":"additional","affiliation":[{"name":"Guangdong Provincial Key Laboratory of Optical Information Materials and Technology & Institute of Electronic Paper Displays, South China Academy of Advanced Optoelectronics, South China Normal University, Guangzhou 510006, China"},{"name":"Shenzhen Guohua Optoelectronics Technology Co., Ltd., Shenzhen 518110, China"}]}],"member":"311","reference":[{"key":"1","doi-asserted-by":"publisher","DOI":"10.1016\/j.procs.2017.06.071"},{"key":"2","doi-asserted-by":"publisher","DOI":"10.5626\/jok.2015.42.1.54"},{"key":"3","doi-asserted-by":"publisher","DOI":"10.3390\/technologies6010017"},{"key":"4","doi-asserted-by":"publisher","DOI":"10.3390\/s18020416"},{"key":"5","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2009.08.002"},{"key":"6","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2010.5543262"},{"key":"7","doi-asserted-by":"publisher","DOI":"10.1631\/fitee.1400209"},{"key":"8","doi-asserted-by":"publisher","DOI":"10.1109\/cvpr.2018.00231"},{"key":"9","doi-asserted-by":"publisher","DOI":"10.1109\/cvpr.2018.00354"},{"key":"10","doi-asserted-by":"publisher","DOI":"10.1109\/cvprw.2018.00286"},{"key":"11","doi-asserted-by":"publisher","DOI":"10.1109\/fg.2017.140"},{"key":"12","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijme.2019.01.001"},{"volume-title":"Affective Computing","year":"2019","author":"R. W. Picard","key":"13"},{"issue":"47","key":"14","first-page":"126","article-title":"Facial action coding system (FACS): a technique for the measurement of facial actions","volume":"12","author":"P. Ekman","year":"1978","journal-title":"Rivista di Psichiatria"},{"issue":"99","key":"15","first-page":"1","article-title":"Deep facial expression recognition: a survey","author":"L. Shan","year":"2018","journal-title":"IEEE Transactions on Affective Computing"},{"key":"16","doi-asserted-by":"publisher","DOI":"10.1016\/j.compedu.2019.103649"},{"key":"17","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2006.14"},{"key":"18","doi-asserted-by":"publisher","DOI":"10.1088\/1742-6596\/1413\/1\/012021"},{"key":"19","doi-asserted-by":"publisher","DOI":"10.1109\/lsp.2016.2603342"},{"article-title":"Acquisition of localization confidence for accurate object detection","year":"2018","author":"B. R. Jiang","key":"20"},{"key":"21","doi-asserted-by":"publisher","DOI":"10.1109\/tmm.2021.3068580"},{"key":"22","doi-asserted-by":"publisher","DOI":"10.1109\/tip.2003.819861"},{"key":"23","article-title":"Auto-encoding variational Bayes","volume":"1050","author":"D. P. Kingma","year":"2014","journal-title":"Clinical Orthopaedics and Related Research"},{"article-title":"Beta-VAE: learning basic visual concepts with a constrained variational framework","year":"2016","author":"I. Higgins","key":"24"}],"container-title":["Security and Communication Networks"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/downloads.hindawi.com\/journals\/scn\/2021\/9539022.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/downloads.hindawi.com\/journals\/scn\/2021\/9539022.xml","content-type":"application\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/downloads.hindawi.com\/journals\/scn\/2021\/9539022.pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,8,20]],"date-time":"2021-08-20T22:30:51Z","timestamp":1629498651000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.hindawi.com\/journals\/scn\/2021\/9539022\/"}},"subtitle":[],"editor":[{"given":"Youwen","family":"Zhu","sequence":"additional","affiliation":[]}],"short-title":[],"issued":{"date-parts":[[2021,8,19]]},"references-count":24,"alternative-id":["9539022","9539022"],"URL":"https:\/\/doi.org\/10.1155\/2021\/9539022","relation":{},"ISSN":["1939-0122","1939-0114"],"issn-type":[{"type":"electronic","value":"1939-0122"},{"type":"print","value":"1939-0114"}],"subject":[],"published":{"date-parts":[[2021,8,19]]}}}