{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:18:30Z","timestamp":1750220310610,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":30,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,1,14]],"date-time":"2022-01-14T00:00:00Z","timestamp":1642118400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"Guangdong Provincial Natural Science Foundation","award":["Grant No. 2214050002868"],"award-info":[{"award-number":["Grant No. 2214050002868"]}]},{"name":"Zhuhai Science and Technology Planning Project","award":["Grant No. ZH22036201210161PWC"],"award-info":[{"award-number":["Grant No. ZH22036201210161PWC"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["Grant No. 62077009,62177006"],"award-info":[{"award-number":["Grant No. 62077009,62177006"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,1,14]]},"DOI":"10.1145\/3512353.3512380","type":"proceedings-article","created":{"date-parts":[[2022,3,14]],"date-time":"2022-03-14T17:25:06Z","timestamp":1647278706000},"page":"183-190","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["A Prior-knowledge guided Multi-scale Deep Network for Speech Emotion Recognition"],"prefix":"10.1145","author":[{"given":"Jun","family":"He","sequence":"first","affiliation":[{"name":"School of artificial intelligence, Beijing Normal University, China and College of education for the future, Beijing Normal University, China"}]},{"given":"Yangcai","family":"Zhong","sequence":"additional","affiliation":[{"name":"School of artificial intelligence, Beijing Normal University, China"}]},{"given":"Penghao","family":"Rao","sequence":"additional","affiliation":[{"name":"School of artificial intelligence, Beijing Normal University, China"}]},{"given":"Bo","family":"Sun","sequence":"additional","affiliation":[{"name":"School of artificial intelligence, Beijing Normal University, China and College of education for the future, Beijing Normal University, China"}]},{"given":"Yinghui","family":"Zhang","sequence":"additional","affiliation":[{"name":"College of education for the future, Beijing Normal University, China"}]}],"member":"320","published-online":{"date-parts":[[2022,3,14]]},"reference":[{"key":"e_1_3_2_1_1_1","article-title":"CHEAVD: a Chinese natural emotional audio\u2013visual database","author":"Li","year":"2016","unstructured":"Li , Ya; Tao, Jianhua ; Chao, Linlin ; Bao, Wei ; Liu, Yazhu ( 2016 ). CHEAVD: a Chinese natural emotional audio\u2013visual database . Journal of Ambient Intelligence and Humanized Computing. Li, Ya; Tao, Jianhua; Chao, Linlin; Bao, Wei; Liu, Yazhu (2016). CHEAVD: a Chinese natural emotional audio\u2013visual database. Journal of Ambient Intelligence and Humanized Computing.","journal-title":"Journal of Ambient Intelligence and Humanized Computing."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2015-3"},{"key":"e_1_3_2_1_3_1","volume-title":"Yiqin Zhao and Chao Li. Exploring Spatio-Temporal Representations by Integrating Attention-based Bidirectional-LSTM-RNNs and FCNs for Speech Emotion Recognition. Interspeech, 272-276","author":"Ziping Zhao","year":"2018","unstructured":"Ziping Zhao , Yu Zheng, Zixing Zhang , Haishuai Wang , Yiqin Zhao and Chao Li. Exploring Spatio-Temporal Representations by Integrating Attention-based Bidirectional-LSTM-RNNs and FCNs for Speech Emotion Recognition. Interspeech, 272-276 , 2018 . DOI: 10.21437\/Interspeech.2018-1477 10.21437\/Interspeech.2018-1477 Ziping Zhao, Yu Zheng, Zixing Zhang, Haishuai Wang, Yiqin Zhao and Chao Li. Exploring Spatio-Temporal Representations by Integrating Attention-based Bidirectional-LSTM-RNNs and FCNs for Speech Emotion Recognition. Interspeech, 272-276, 2018. DOI: 10.21437\/Interspeech.2018-1477"},{"key":"e_1_3_2_1_4_1","volume-title":"CNN+LSTM Architecture for Speech Emotion Recognition with Data Augmentation. 21-25. 10.21437\/SMM.2018-5","author":"Etienne","year":"2018","unstructured":"Etienne , Caroline & Fidanza, Guillaume & Petrovskii, Andrei & Devillers, Laurence & Schmauch, Benoit . ( 2018 ). CNN+LSTM Architecture for Speech Emotion Recognition with Data Augmentation. 21-25. 10.21437\/SMM.2018-5 . Etienne, Caroline & Fidanza, Guillaume & Petrovskii, Andrei & Devillers, Laurence & Schmauch, Benoit. (2018). CNN+LSTM Architecture for Speech Emotion Recognition with Data Augmentation. 21-25. 10.21437\/SMM.2018-5."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2016.7472669"},{"key":"e_1_3_2_1_6_1","volume-title":"An Argument for Basic Emotions., 169\u2013200","author":"Ekman P","year":"1992","unstructured":"Ekman , P ( 1992 ). An Argument for Basic Emotions., 169\u2013200 ( Lawrence Erlbaum) . Ekman,P (1992). An Argument for Basic Emotions., 169\u2013200 (Lawrence Erlbaum)."},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1016\/B978-0-12-558701-3.50007-7"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-99495-6_21"},{"key":"e_1_3_2_1_9_1","first-page":"227","volume-title":"INTERSPEECH","author":"Kun Han","year":"2014","unstructured":"Kun Han , Dong Yu, and Ivan Tashev . 2014 . Speech emotion recognition using deep neural network and extreme learning machine. In Haizhou Li, Helen M. Meng, Bin Ma, Engsiong Chng, and Lei Xie, editors , INTERSPEECH , pages 223\u2013 227 . ISCA. Kun Han, Dong Yu, and Ivan Tashev. 2014. Speech emotion recognition using deep neural network and extreme learning machine. In Haizhou Li, Helen M. Meng, Bin Ma, Engsiong Chng, and Lei Xie, editors, INTERSPEECH, pages 223\u2013227. ISCA."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/LSP.2018.2860246"},{"key":"e_1_3_2_1_11_1","volume-title":"CoRR abs\/1512.00567.","author":"Szegedy C.","year":"2015","unstructured":"Szegedy , C. , Vanhoucke , V. , Ioffe , S. , Shlens , J. , and Wojna , Z . ( 2015 ). \u201c Rethinking the inception architecture for computer vision .,\u201d CoRR abs\/1512.00567. Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., and Wojna, Z. (2015). \u201cRethinking the inception architecture for computer vision.,\u201d CoRR abs\/1512.00567."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.2990405"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2017.7952552"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1109\/SLT.2018.8639633"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D16-1061"},{"key":"e_1_3_2_1_16_1","first-page":"115","article-title":"Learning precise timing with LSTM recurrent networks","volume":"3","author":"Gers F. A.","year":"2002","unstructured":"Gers , F. A. , Schraudolph , N. , and Schmidhuber , J. ( 2002 ). \u201c Learning precise timing with LSTM recurrent networks ,\u201d Journal of Machine Learning Research 3 , 115 \u2013 143 . Gers, F. A., Schraudolph, N., and Schmidhuber, J. (2002). \u201cLearning precise timing with LSTM recurrent networks,\u201d Journal of Machine Learning Research 3, 115\u2013143.","journal-title":"Journal of Machine Learning Research"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2005.06.042"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10579-008-9076-6"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1007\/s12652-016-0406-z"},{"key":"e_1_3_2_1_20_1","first-page":"1","volume-title":"Beijing","author":"Li J.","year":"2018","unstructured":"Y. Li , J. Tao , B. Schuller , S. Shan , D. Jiang and J. Jia , \" MEC 2017: Multimodal Emotion Recognition Challenge,\" 2018 First Asian Conference on Affective Computing and Intelligent Interaction (ACII Asia) , Beijing , 2018 , pp. 1 - 5 . Y. Li, J. Tao, B. Schuller, S. Shan, D. Jiang and J. Jia, \"MEC 2017: Multimodal Emotion Recognition Challenge,\" 2018 First Asian Conference on Affective Computing and Intelligent Interaction (ACII Asia), Beijing, 2018, pp. 1-5."},{"key":"e_1_3_2_1_21_1","volume-title":"CNN+LSTM Architecture for Speech Emotion Recognition with Data Augmentation. 21-25. 10.21437\/SMM.2018-5","author":"Etienne","year":"2018","unstructured":"Etienne , Caroline & Fidanza, Guillaume & Petrovskii, Andrei & Devillers, Laurence & Schmauch, Benoit . ( 2018 ). CNN+LSTM Architecture for Speech Emotion Recognition with Data Augmentation. 21-25. 10.21437\/SMM.2018-5 . Etienne, Caroline & Fidanza, Guillaume & Petrovskii, Andrei & Devillers, Laurence & Schmauch, Benoit. (2018). CNN+LSTM Architecture for Speech Emotion Recognition with Data Augmentation. 21-25. 10.21437\/SMM.2018-5."},{"key":"e_1_3_2_1_22_1","first-page":"6685","author":"Yeh","year":"2019","unstructured":"Yeh , Sung Lin & Lin, Yun-Shao & Lee, Chi-Chun . ( 2019 ). An Interaction-aware Attention Network for Speech Emotion Recognition in Spoken Dialogs. 6685 - 6689 . 10.1109\/ICASSP.2019.8683293. Yeh, Sung Lin & Lin, Yun-Shao & Lee, Chi-Chun. (2019). An Interaction-aware Attention Network for Speech Emotion Recognition in Spoken Dialogs. 6685-6689. 10.1109\/ICASSP.2019.8683293.","journal-title":"An Interaction-aware Attention Network for Speech Emotion Recognition in Spoken Dialogs."},{"key":"e_1_3_2_1_23_1","volume-title":"Attention Based Fully Convolutional Network for Speech Emotion Recognition. 2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC)","author":"Zhang Y.","year":"2018","unstructured":"Zhang , Y. , , Attention Based Fully Convolutional Network for Speech Emotion Recognition. 2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC) , 2018 . Zhang, Y., , Attention Based Fully Convolutional Network for Speech Emotion Recognition. 2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), 2018."},{"key":"e_1_3_2_1_24_1","first-page":"1","article-title":"Exploring Deep Spectrum Representations via Attention-Based Recurrent and Convolutional Neural Networks for Speech Emotion Recognition","author":"Zhao","year":"2019","unstructured":"Zhao , Ziping & Bao, Zhongtian & Zhao, Yiqin & Zhang, Zixing & Cummins, Nicholas & Ren, Zhao & Schuller, Bj\u00f6rn . ( 2019 ). Exploring Deep Spectrum Representations via Attention-Based Recurrent and Convolutional Neural Networks for Speech Emotion Recognition . IEEE Access. PP. 1 - 1 . Zhao, Ziping & Bao, Zhongtian & Zhao, Yiqin & Zhang, Zixing & Cummins, Nicholas & Ren, Zhao & Schuller, Bj\u00f6rn. (2019). Exploring Deep Spectrum Representations via Attention-Based Recurrent and Convolutional Neural Networks for Speech Emotion Recognition. IEEE Access. PP. 1-1.","journal-title":"IEEE Access. PP."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2019.2938007"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.2990405"},{"key":"e_1_3_2_1_27_1","first-page":"6319","article-title":"-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"2021","author":"Xu M","year":"2021","unstructured":"Xu M , Zhang F , Cui X , Speech Emotion Recognition with Multiscale Area Attention and Data Augmentation. 2021 -2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) . IEEE , 2021 : 6319 - 6323 . Xu M, Zhang F, Cui X, Speech Emotion Recognition with Multiscale Area Attention and Data Augmentation. 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2021: 6319-6323.","journal-title":"IEEE"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2015.7301268"},{"key":"e_1_3_2_1_29_1","volume-title":"Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies","volume":"1","author":"Zhou D.","year":"2018","unstructured":"Zhou , D. , Yang , Y. , & He , Y. ( 2018 , June). Relevant emotion ranking from text constrained with emotion relationships . In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies , Volume 1 (Long Papers) (pp. 561-571). Zhou, D., Yang, Y., & He, Y. (2018, June). Relevant emotion ranking from text constrained with emotion relationships. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers) (pp. 561-571)."},{"key":"e_1_3_2_1_30_1","unstructured":"https:\/\/www.jianshu.com\/p\/e7795cfbf76b?utm_campaign=maleskine&utm_content=note&utm_medium=seo_notes&utm_source=recommendation  https:\/\/www.jianshu.com\/p\/e7795cfbf76b?utm_campaign=maleskine&utm_content=note&utm_medium=seo_notes&utm_source=recommendation"}],"event":{"name":"APIT 2022: 2022 4th Asia Pacific Information Technology Conference","acronym":"APIT 2022","location":"Virtual Event Thailand"},"container-title":["2022 4th Asia Pacific Information Technology Conference"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3512353.3512380","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3512353.3512380","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T20:11:42Z","timestamp":1750191102000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3512353.3512380"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,1,14]]},"references-count":30,"alternative-id":["10.1145\/3512353.3512380","10.1145\/3512353"],"URL":"https:\/\/doi.org\/10.1145\/3512353.3512380","relation":{},"subject":[],"published":{"date-parts":[[2022,1,14]]},"assertion":[{"value":"2022-03-14","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}