{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,2,13]],"date-time":"2024-02-13T19:31:04Z","timestamp":1707852664729},"reference-count":15,"publisher":"IGI Global","issue":"2","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020,4,1]]},"abstract":"<p>Facial expression intensity has been proposed to digitize the degree of facial expressions in order to retrieve impressive scenes from lifelog videos. The intensity is calculated based on the correlation of facial features compared to each facial expression. However, the correlation is not determined objectively. It should be determined statistically based on the contribution score of the facial features necessary for expression recognition. Therefore, the proposed method recognizes facial expressions by using a neural network and calculates the contribution score of input toward the output. First, the authors improve some facial features. After that, they verify the score correctly by comparing the accuracy transitions depending on reducing useful and useless features and process the score statistically. As a result, they extract useful facial features from the neural network.<\/p>","DOI":"10.4018\/ijsi.2020040105","type":"journal-article","created":{"date-parts":[[2020,2,7]],"date-time":"2020-02-07T19:52:12Z","timestamp":1581105132000},"page":"68-84","source":"Crossref","is-referenced-by-count":1,"title":["Finding Useful Features for Facial Expression Recognition and Intensity Estimation by Neural Network"],"prefix":"10.4018","volume":"8","author":[{"given":"Naoki","family":"Imamura","sequence":"first","affiliation":[{"name":"Kyoto Institute of Technology, Kyoto, Japan"}]},{"given":"Hiroki","family":"Nomiya","sequence":"additional","affiliation":[{"name":"Graduate School of Information Science, Kyoto Institute of Technology, Kyoto, Japan"}]},{"given":"Teruhisa","family":"Hochin","sequence":"additional","affiliation":[{"name":"Graduate School of Information Science, Kyoto Institute of Technology, Kyoto, Japan"}]}],"member":"2432","reference":[{"key":"IJSI.2020040105-0","unstructured":"Aizawa, K. (2009). Life log: Food log and future prospect of life log. IPSJ magazine, 50(7), 592-597. (in Japanese)"},{"key":"IJSI.2020040105-1","author":"M.Cox","year":"2013","journal-title":"CSIRO Face Analysis SDK"},{"key":"IJSI.2020040105-2","doi-asserted-by":"publisher","DOI":"10.1109\/AFGR.2000.840611"},{"key":"IJSI.2020040105-3","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW.2015.12"},{"key":"IJSI.2020040105-4","doi-asserted-by":"crossref","unstructured":"Kobayashi, H., & Hara, F. (1995). Analysis of Neural Network Recognition Characteristics of 6 Basic Facial Expressions. Transactions of the Japan society of mechanical engineers series C,61(582), 678-685. (in Japanese)","DOI":"10.1299\/kikaic.61.678"},{"key":"IJSI.2020040105-5","doi-asserted-by":"publisher","DOI":"10.1109\/AFGR.1998.670949"},{"key":"IJSI.2020040105-6","unstructured":"Megvii. (2019). Face++ Cognitive Services. Retrieved from https:\/\/www.faceplusplus.com\/"},{"key":"IJSI.2020040105-7","author":"R.Memisevic","year":"2014","journal-title":"Zero-bias autoencoders and the benefits of co-adapting features"},{"key":"IJSI.2020040105-8","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2016.11.008"},{"issue":"1","key":"IJSI.2020040105-9","first-page":"32","article-title":"Expression Strength for the Emotional Scene Detection from Lifelog Videos.","volume":"16","author":"A.Morikuni","year":"2015","journal-title":"International Journal of Computer and Information Science"},{"key":"IJSI.2020040105-10","doi-asserted-by":"crossref","unstructured":"Nishime, T., Endo, S., Toma, N., Yamada, K., & Akamine, Y. (2017). Feature Acquisition and Analysis for Facial Expression Recognition Using Convolutional Neural Networks. Transactions of the Japanese Society for Artificial Intelligence, 32(5).","DOI":"10.1527\/tjsai.F-H34"},{"key":"IJSI.2020040105-11","unstructured":"Nomiya, H., Morikuni, A., & Hochin, T. (2013). Efficient expression scene detection based on the positional relationship of facial feature points for lifelog video retrieval. Journal of the Institute of Electronics, Information and Communication Engineers, 1279-1289 (in Japanese)"},{"key":"IJSI.2020040105-12","first-page":"317","article-title":"Web-based database for facial expression analysis.","author":"M.Pantic","year":"2005","journal-title":"Proc. of IEEE International Conference on Multimedia and Expo"},{"key":"IJSI.2020040105-13","first-page":"121","article-title":"Estimation of Emotional Scene from Lifelog Videos in Consideration of Intensity of Various Facial Expressions.","author":"S.Sakaue","year":"2017","journal-title":"Proceedings of the International Conference on Software Engineering, Artificial Intelligence, Networking and Parallel\/Distributed Computing"},{"key":"IJSI.2020040105-14","doi-asserted-by":"publisher","DOI":"10.1109\/CSII.2018.00030"}],"container-title":["International Journal of Software Innovation"],"original-title":[],"language":"ng","link":[{"URL":"https:\/\/www.igi-global.com\/viewtitle.aspx?TitleId=248531","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,5,6]],"date-time":"2022-05-06T21:10:32Z","timestamp":1651871432000},"score":1,"resource":{"primary":{"URL":"https:\/\/services.igi-global.com\/resolvedoi\/resolve.aspx?doi=10.4018\/IJSI.2020040105"}},"subtitle":[""],"short-title":[],"issued":{"date-parts":[[2020,4,1]]},"references-count":15,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2020,4]]}},"URL":"https:\/\/doi.org\/10.4018\/ijsi.2020040105","relation":{},"ISSN":["2166-7160","2166-7179"],"issn-type":[{"value":"2166-7160","type":"print"},{"value":"2166-7179","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020,4,1]]}}}