{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,13]],"date-time":"2026-01-13T06:52:59Z","timestamp":1768287179258,"version":"3.49.0"},"reference-count":47,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61841203"],"award-info":[{"award-number":["61841203"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"Ph.D. Program of Yanyuan Science and Technology Innovation Fund","doi-asserted-by":"publisher","award":["2023BSZX04"],"award-info":[{"award-number":["2023BSZX04"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2025]]},"DOI":"10.1109\/access.2025.3551549","type":"journal-article","created":{"date-parts":[[2025,3,14]],"date-time":"2025-03-14T17:55:22Z","timestamp":1741974922000},"page":"51208-51219","source":"Crossref","is-referenced-by-count":3,"title":["Hierarchical Multi-Task Learning Based on Interactive Multi-Head Attention Feature Fusion for Speech Depression Recognition"],"prefix":"10.1109","volume":"13","author":[{"ORCID":"https:\/\/orcid.org\/0009-0001-3684-1464","authenticated-orcid":false,"given":"Yujuan","family":"Xing","sequence":"first","affiliation":[{"name":"School of Digital Media (Computer), Lanzhou University of Arts and Science, Lanzhou, China"}]},{"given":"Ruifang","family":"He","sequence":"additional","affiliation":[{"name":"Second Provincial People&#x2019;s Hospital of Gansu, Lanzhou, China"}]},{"given":"Chengwen","family":"Zhang","sequence":"additional","affiliation":[{"name":"School of Digital Media (Computer), Lanzhou University of Arts and Science, Lanzhou, China"}]},{"given":"Ping","family":"Tan","sequence":"additional","affiliation":[{"name":"School of Digital Media (Computer), Lanzhou University of Arts and Science, Lanzhou, China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.3390\/brainsci10120977"},{"key":"ref2","volume-title":"World Health Statistics","year":"2017"},{"key":"ref3","volume-title":"World Health Statistics","year":"2021"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/BSN.2013.6575522"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1016\/j.cpr.2007.10.001"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.3390\/s20185041"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.3390\/s22010067"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2021.02.019"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/BIBM52615.2021.9669546"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1016\/j.ypmed.2021.106847"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2023\/658"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2017-887"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.23919\/EUSIPCO54536.2021.9615933"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/IALP54817.2021.9675220"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1142\/S0218213023400043"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.4103\/0256-4602.64604.2010"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/JSTARS.2020.3041868"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/JSEN.2024.3393619"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/CCDC49329.2020.9164733"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.3389\/fnbot.2023.1181598"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1007\/978-981-99-8462-6_6"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2022.10.032"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/WACV48630.2021.00360"},{"key":"ref24","first-page":"1","article-title":"Learning multiple layers of features from tiny images","author":"Krizhevsky","year":"2009","journal-title":"Handbook Systemic Autoimmune Diseases"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-015-0816-y"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2024.106552"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.3390\/s23239402"},{"key":"ref28","first-page":"623","article-title":"Multi-modal depression estimation based on sub-attentional fusion","volume-title":"Proc. ECCV Workshops","author":"Wei"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2024.3350071"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2017.2724035"},{"key":"ref31","doi-asserted-by":"crossref","first-page":"41","DOI":"10.1023\/A:1007379606734","article-title":"Multitask learning","volume":"28","author":"Caruana","year":"1998","journal-title":"Mach. Learn."},{"key":"ref32","first-page":"2803","article-title":"Improved end-to-end speech emotion recognition using self attention mechanism and multitask learning","volume-title":"Proc. Interspeech","author":"Li"},{"issue":"16","key":"ref33","doi-asserted-by":"crossref","first-page":"22355","DOI":"10.1007\/s11042-021-11282-4","article-title":"Ensemble of multi-task deep convolutional neural networks using transfer learning for fruit freshness classification","volume":"81","author":"Kang","year":"2022","journal-title":"Multimedia Tools Appl."},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP43922.2022.9747466"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/JBHI.2024.3412103"},{"issue":"8","key":"ref36","first-page":"2395","article-title":"A multi-domain text classification method based on recurrent convolution multi-task learning","volume":"43","author":"Xie","year":"2021","journal-title":"J. Electron. Inf. Technol."},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1145\/2661806.2661807"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1145\/2512530.2512533"},{"key":"ref39","first-page":"1","article-title":"FEELTRACE: An instrument for recording perceived emotion in real time","volume-title":"Proc. ISCA Tutorial Res. Workshop (ITRW) Speech Emotion","author":"Cowie"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2014.6853741"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1142\/S1469026822500092"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1155\/2021\/9994340"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2005.159"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.3389\/fnbot.2021.684037"},{"key":"ref45","doi-asserted-by":"crossref","first-page":"299","DOI":"10.1016\/j.jad.2022.11.060","article-title":"Depression recognition using a proposed speech chain model fusing speech production and perception features","volume":"323","author":"Du","year":"2023","journal-title":"J. Affect. Disorders"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1109\/TCDS.2023.3273614"},{"key":"ref47","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2023.105898","article-title":"A deep learning model for depression detection based on MFCC and CNN generated spectrogram features","volume":"90","author":"Das","year":"2024","journal-title":"Biomed. Signal Process. Control"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/6287639\/10820123\/10926518.pdf?arnumber=10926518","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,3,31]],"date-time":"2025-03-31T23:44:38Z","timestamp":1743464678000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10926518\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"references-count":47,"URL":"https:\/\/doi.org\/10.1109\/access.2025.3551549","relation":{},"ISSN":["2169-3536"],"issn-type":[{"value":"2169-3536","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]}}}