{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,26]],"date-time":"2026-03-26T04:09:08Z","timestamp":1774498148296,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":49,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,10,28]],"date-time":"2024-10-28T00:00:00Z","timestamp":1730073600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/https:\/\/doi.org\/10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["No. 62302145"],"award-info":[{"award-number":["No. 62302145"]}],"id":[{"id":"10.13039\/https:\/\/doi.org\/10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Anhui Province Science Foundation for Youths","award":["No. 2308085QF230"],"award-info":[{"award-number":["No. 2308085QF230"]}]},{"name":"the Major scientific and technological project of Anhui Provincial Science and Technology Innovation Platform","award":["No. 202305a12020012"],"award-info":[{"award-number":["No. 202305a12020012"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,10,28]]},"DOI":"10.1145\/3664647.3681546","type":"proceedings-article","created":{"date-parts":[[2024,10,26]],"date-time":"2024-10-26T06:59:27Z","timestamp":1729925967000},"page":"311-320","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":17,"title":["FacialPulse: An Efficient RNN-based Depression Detection via Temporal Facial Landmarks"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0006-8925-3419","authenticated-orcid":false,"given":"Ruiqi","family":"Wang","sequence":"first","affiliation":[{"name":"Hefei University of Technology, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5483-2812","authenticated-orcid":false,"given":"Jinyang","family":"Huang","sequence":"additional","affiliation":[{"name":"Hefei University of Technology, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1124-0854","authenticated-orcid":false,"given":"Jie","family":"Zhang","sequence":"additional","affiliation":[{"name":"CFAR and IHPC, A*STAR, Singapore, Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-0669-268X","authenticated-orcid":false,"given":"Xin","family":"Liu","sequence":"additional","affiliation":[{"name":"Hefei University of Technology, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0413-6135","authenticated-orcid":false,"given":"Xiang","family":"Zhang","sequence":"additional","affiliation":[{"name":"University of Science and Technology of China, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0537-4522","authenticated-orcid":false,"given":"Zhi","family":"Liu","sequence":"additional","affiliation":[{"name":"The University of Electro-Communications, Tokyo, Japan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6026-5743","authenticated-orcid":false,"given":"Peng","family":"Zhao","sequence":"additional","affiliation":[{"name":"Hefei University of Technology, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-5444-6814","authenticated-orcid":false,"given":"Sigui","family":"Chen","sequence":"additional","affiliation":[{"name":"Hefei University of Technology, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9750-7032","authenticated-orcid":false,"given":"Xiao","family":"Sun","sequence":"additional","affiliation":[{"name":"hefei university of technology, Hefei, China"}]}],"member":"320","published-online":{"date-parts":[[2024,10,28]]},"reference":[{"key":"e_1_3_2_2_1_1","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2018.2870884"},{"key":"e_1_3_2_2_2_1","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2021.3124142"},{"key":"e_1_3_2_2_3_1","doi-asserted-by":"publisher","DOI":"10.3390\/app14020635"},{"key":"e_1_3_2_2_4_1","doi-asserted-by":"publisher","DOI":"10.1109\/FG47880.2020.00110"},{"key":"e_1_3_2_2_5_1","volume-title":"Combining global and local convolutional 3d networks for detecting depression from facial expressions. In 2019 14th ieee international conference on automatic face & gesture recognition (fg","author":"de Melo Wheidima Carneiro","year":"2019","unstructured":"Wheidima Carneiro de Melo, Eric Granger, and Abdenour Hadid. 2019. Combining global and local convolutional 3d networks for detecting depression from facial expressions. In 2019 14th ieee international conference on automatic face & gesture recognition (fg 2019). IEEE, 1--8."},{"key":"e_1_3_2_2_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP.2019.8803467"},{"key":"e_1_3_2_2_7_1","volume-title":"A deep multiscale spatiotemporal network for assessing depression from facial dynamics","author":"de Melo Wheidima Carneiro","year":"2020","unstructured":"Wheidima Carneiro de Melo, Eric Granger, and Abdenour Hadid. 2020. A deep multiscale spatiotemporal network for assessing depression from facial dynamics. IEEE transactions on affective computing, Vol. 13, 3 (2020), 1581--1592."},{"key":"e_1_3_2_2_8_1","volume-title":"ICASSP 2020--2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 1080--1084","author":"De Melo Wheidima Carneiro","year":"2020","unstructured":"Wheidima Carneiro De Melo, Eric Granger, and Miguel Bordallo Lopez. 2020. Encoding temporal information for automatic depression recognition from facial analysis. In ICASSP 2020--2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 1080--1084."},{"key":"e_1_3_2_2_9_1","volume-title":"MDN: A deep maximization-differentiation network for spatio-temporal depression detection","author":"de Melo Wheidima Carneiro","year":"2021","unstructured":"Wheidima Carneiro de Melo, Eric Granger, and Miguel Bordallo Lopez. 2021. MDN: A deep maximization-differentiation network for spatio-temporal depression detection. IEEE transactions on affective computing (2021)."},{"key":"e_1_3_2_2_10_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.rcim.2023.102660"},{"key":"e_1_3_2_2_11_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00552"},{"key":"e_1_3_2_2_12_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.sigpro.2020.107755"},{"key":"e_1_3_2_2_13_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2020.10.015"},{"key":"e_1_3_2_2_14_1","doi-asserted-by":"publisher","DOI":"10.1002\/int.22426"},{"key":"e_1_3_2_2_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2018.2877129"},{"key":"e_1_3_2_2_16_1","volume-title":"KeystrokeSniffer: An Off-the-Shelf Smartphone Can Eavesdrop on Your Privacy from Anywhere","author":"Huang Jinyang","year":"2024","unstructured":"Jinyang Huang, Jia-Xuan Bai, Xiang Zhang, Zhi Liu, Yuanhao Feng, Jianchun Liu, Xiao Sun, Mianxiong Dong, and Meng Li. 2024. KeystrokeSniffer: An Off-the-Shelf Smartphone Can Eavesdrop on Your Privacy from Anywhere. IEEE Transactions on Information Forensics and Security (2024)."},{"key":"e_1_3_2_2_17_1","volume-title":"Phyfinatt: An undetectable attack framework against phy layer fingerprint-based wifi authentication","author":"Huang Jinyang","year":"2023","unstructured":"Jinyang Huang, Bin Liu, Chenglin Miao, Xiang Zhang, Jiancun Liu, Lu Su, Zhi Liu, and Yu Gu. 2023. Phyfinatt: An undetectable attack framework against phy layer fingerprint-based wifi authentication. IEEE Transactions on Mobile Computing (2023)."},{"key":"e_1_3_2_2_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/2661806.2661812"},{"key":"e_1_3_2_2_19_1","volume-title":"MMDA: A Multimodal Dataset for Depression and Anxiety Detection. In International Conference on Pattern Recognition. Springer, 691--702","author":"Jiang Yueqi","year":"2022","unstructured":"Yueqi Jiang, Ziyang Zhang, and Xiao Sun. 2022. MMDA: A Multimodal Dataset for Depression and Anxiety Detection. In International Conference on Pattern Recognition. Springer, 691--702."},{"key":"e_1_3_2_2_20_1","volume-title":"Classifying major depressive disorder and response to deep brain stimulation over time by analyzing facial expressions","author":"Jiang Zifan","year":"2020","unstructured":"Zifan Jiang, Sahar Harati, Andrea Crowell, Helen S Mayberg, Shamim Nemati, and Gari D Clifford. 2020. Classifying major depressive disorder and response to deep brain stimulation over time by analyzing facial expressions. IEEE transactions on biomedical engineering, Vol. 68, 2 (2020), 664--672."},{"key":"e_1_3_2_2_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/2661806.2661814"},{"key":"e_1_3_2_2_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/3581783.3612053"},{"key":"e_1_3_2_2_23_1","volume-title":"Deep facial expression recognition: A survey","author":"Li Shan","year":"2020","unstructured":"Shan Li and Weihong Deng. 2020. Deep facial expression recognition: A survey. IEEE transactions on affective computing, Vol. 13, 3 (2020), 1195--1215."},{"key":"e_1_3_2_2_24_1","doi-asserted-by":"publisher","DOI":"10.1109\/TNET.2024.3390416"},{"key":"e_1_3_2_2_25_1","volume-title":"Finch: Enhancing federated learning with hierarchical neural architecture search","author":"Liu Jianchun","year":"2023","unstructured":"Jianchun Liu, Jiaming Yan, Hongli Xu, Zhiyuan Wang, Jinyang Huang, and Yang Xu. 2023. Finch: Enhancing federated learning with hierarchical neural architecture search. IEEE Transactions on Mobile Computing (2023)."},{"key":"e_1_3_2_2_26_1","doi-asserted-by":"publisher","DOI":"10.3390\/s23031080"},{"key":"e_1_3_2_2_27_1","doi-asserted-by":"publisher","DOI":"10.3390\/electronics9050764"},{"key":"e_1_3_2_2_28_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2022.117512"},{"key":"e_1_3_2_2_29_1","doi-asserted-by":"publisher","DOI":"10.1109\/ACIIW.2019.8925158"},{"key":"e_1_3_2_2_30_1","volume-title":"Generalized Taylor's formula. Applied Mathematics and computation","author":"Odibat Zaid M","year":"2007","unstructured":"Zaid M Odibat and Nabil T Shawagfeh. 2007. Generalized Taylor's formula. Applied Mathematics and computation, Vol. 186, 1 (2007), 286--293."},{"key":"e_1_3_2_2_31_1","doi-asserted-by":"publisher","DOI":"10.1145\/3581783.3612555"},{"key":"e_1_3_2_2_32_1","volume-title":"Integrating deep facial priors into landmarks for privacy preserving multimodal depression recognition","author":"Pan Yuchen","year":"2023","unstructured":"Yuchen Pan, Yuanyuan Shang, Zhuhong Shao, Tie Liu, Guodong Guo, and Hui Ding. 2023. Integrating deep facial priors into landmarks for privacy preserving multimodal depression recognition. IEEE Transactions on Affective Computing (2023)."},{"key":"e_1_3_2_2_33_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICTC52510.2021.9620798"},{"key":"e_1_3_2_2_34_1","volume-title":"The Verbal and Non Verbal Signals of Depression--Combining Acoustics, Text and Visuals for Estimating Depression Level. arXiv preprint arXiv:1904.07656","author":"Qureshi Syed Arbaaz","year":"2019","unstructured":"Syed Arbaaz Qureshi, Mohammed Hasanuzzaman, Sriparna Saha, and Ga\u00ebl Dias. 2019. The Verbal and Non Verbal Signals of Depression--Combining Acoustics, Text and Visuals for Estimating Depression Level. arXiv preprint arXiv:1904.07656 (2019)."},{"key":"e_1_3_2_2_35_1","first-page":"171","article-title":"A multimodal emotion recognition system using facial landmark analysis. Iranian Journal of Science and Technology","volume":"43","author":"Rahdari Farhad","year":"2019","unstructured":"Farhad Rahdari, Esmat Rashedi, and Mahdi Eftekhari. 2019. A multimodal emotion recognition system using facial landmark analysis. Iranian Journal of Science and Technology, Transactions of Electrical Engineering, Vol. 43 (2019), 171--189.","journal-title":"Transactions of Electrical Engineering"},{"key":"e_1_3_2_2_36_1","doi-asserted-by":"publisher","DOI":"10.1111\/srt.12977"},{"key":"e_1_3_2_2_37_1","volume-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision. 20839--20849","author":"Song Luchuan","year":"2023","unstructured":"Luchuan Song, Guojun Yin, Zhenchao Jin, Xiaoyi Dong, and Chenliang Xu. 2023. Emotional listener portrait: Neural listener head generation with emotion. In Proceedings of the IEEE\/CVF International Conference on Computer Vision. 20839--20849."},{"key":"e_1_3_2_2_38_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICPR48806.2021.9412942"},{"key":"e_1_3_2_2_39_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICAIE53562.2021.00020"},{"key":"e_1_3_2_2_40_1","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2020.2970418"},{"key":"e_1_3_2_2_41_1","doi-asserted-by":"publisher","DOI":"10.1145\/2661806.2661807"},{"key":"e_1_3_2_2_42_1","volume-title":"Two-stage temporal modelling framework for video-based depression recognition using graph representation. arXiv preprint arXiv:2111.15266","author":"Xu Jiaqi","year":"2021","unstructured":"Jiaqi Xu, Siyang Song, Keerthy Kusumam, Hatice Gunes, and Michel Valstar. 2021. Two-stage temporal modelling framework for video-based depression recognition using graph representation. arXiv preprint arXiv:2111.15266 (2021)."},{"key":"e_1_3_2_2_43_1","volume-title":"Drivers' visual distraction detection using facial landmarks and head pose. Transportation research record","author":"Zhang Shile","year":"2022","unstructured":"Shile Zhang and Mohamed Abdel-Aty. 2022. Drivers' visual distraction detection using facial landmarks and head pose. Transportation research record, Vol. 2676, 9 (2022), 491--501."},{"key":"e_1_3_2_2_44_1","doi-asserted-by":"publisher","DOI":"10.1109\/THMS.2023.3264247"},{"key":"e_1_3_2_2_45_1","volume-title":"ReSup: Reliable Label Noise Suppression for Facial Expression Recognition. arXiv preprint arXiv:2305.17895","author":"Zhang Xiang","year":"2023","unstructured":"Xiang Zhang, Yan Lu, Huan Yan, Jingyang Huang, Yusheng Ji, and Yu Gu. 2023. ReSup: Reliable Label Noise Suppression for Facial Expression Recognition. arXiv preprint arXiv:2305.17895 (2023)."},{"key":"e_1_3_2_2_46_1","doi-asserted-by":"publisher","DOI":"10.1145\/3503161.3548330"},{"key":"e_1_3_2_2_47_1","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2020.3022732"},{"key":"e_1_3_2_2_48_1","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2017.2650899"},{"key":"e_1_3_2_2_49_1","doi-asserted-by":"publisher","DOI":"10.1109\/TGRS.2024.3367986"}],"event":{"name":"MM '24: The 32nd ACM International Conference on Multimedia","location":"Melbourne VIC Australia","acronym":"MM '24","sponsor":["SIGMM ACM Special Interest Group on Multimedia"]},"container-title":["Proceedings of the 32nd ACM International Conference on Multimedia"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3664647.3681546","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3664647.3681546","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T00:57:48Z","timestamp":1750294668000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3664647.3681546"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,28]]},"references-count":49,"alternative-id":["10.1145\/3664647.3681546","10.1145\/3664647"],"URL":"https:\/\/doi.org\/10.1145\/3664647.3681546","relation":{},"subject":[],"published":{"date-parts":[[2024,10,28]]},"assertion":[{"value":"2024-10-28","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}