{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,8]],"date-time":"2026-01-08T08:22:54Z","timestamp":1767860574432,"version":"3.49.0"},"reference-count":97,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"21","license":[{"start":{"date-parts":[[2021,11,1]],"date-time":"2021-11-01T00:00:00Z","timestamp":1635724800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2021,11,1]],"date-time":"2021-11-01T00:00:00Z","timestamp":1635724800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-009"},{"start":{"date-parts":[[2021,11,1]],"date-time":"2021-11-01T00:00:00Z","timestamp":1635724800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-001"}],"funder":[{"name":"Zhejiang Lab\u2019s International Talent Fund for Young Professionals under (Project HANAMI), China"},{"DOI":"10.13039\/501100001691","name":"JSPS Postdoctoral Fellowship for Research in Japan","doi-asserted-by":"publisher","award":["P19081 from the Japan Society for the Promotion of Science (JSPS), Japan"],"award-info":[{"award-number":["P19081 from the Japan Society for the Promotion of Science (JSPS), Japan"]}],"id":[{"id":"10.13039\/501100001691","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001691","name":"Japan Society for the Promotion of Science","doi-asserted-by":"publisher","award":["19F19081"],"award-info":[{"award-number":["19F19081"]}],"id":[{"id":"10.13039\/501100001691","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001691","name":"Japan Society for the Promotion of Science","doi-asserted-by":"publisher","award":["20H00569 from the Ministry of Education, Culture, Sports, Science and Technology (MEXT), Japan"],"award-info":[{"award-number":["20H00569 from the Ministry of Education, Culture, Sports, Science and Technology (MEXT), Japan"]}],"id":[{"id":"10.13039\/501100001691","id-type":"DOI","asserted-by":"publisher"}]},{"name":"European Union\u2019s Horizon 2020 Programme by the Smart Environments for Person-Centered Sustainable Work and Well-Being","award":["826506"],"award-info":[{"award-number":["826506"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Internet Things J."],"published-print":{"date-parts":[[2021,11,1]]},"DOI":"10.1109\/jiot.2021.3067605","type":"journal-article","created":{"date-parts":[[2021,3,22]],"date-time":"2021-03-22T20:52:35Z","timestamp":1616446355000},"page":"16035-16046","source":"Crossref","is-referenced-by-count":21,"title":["Computer Audition for Fighting the SARS-CoV-2 Corona Crisis\u2014Introducing the Multitask Speech Corpus for COVID-19"],"prefix":"10.1109","volume":"8","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-1918-6453","authenticated-orcid":false,"given":"Kun","family":"Qian","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7453-5612","authenticated-orcid":false,"given":"Maximilian","family":"Schmitt","sequence":"additional","affiliation":[]},{"given":"Huaiyuan","family":"Zheng","sequence":"additional","affiliation":[]},{"given":"Tomoya","family":"Koike","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5776-6849","authenticated-orcid":false,"given":"Jing","family":"Han","sequence":"additional","affiliation":[]},{"given":"Juan","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Wei","family":"Ji","sequence":"additional","affiliation":[]},{"given":"Junjun","family":"Duan","sequence":"additional","affiliation":[]},{"given":"Meishu","family":"Song","sequence":"additional","affiliation":[]},{"given":"Zijiang","family":"Yang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0707-5016","authenticated-orcid":false,"given":"Zhao","family":"Ren","sequence":"additional","affiliation":[]},{"given":"Shuo","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Zixing","family":"Zhang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1132-0355","authenticated-orcid":false,"given":"Yoshiharu","family":"Yamamoto","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6478-8699","authenticated-orcid":false,"given":"Bjorn W.","family":"Schuller","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1007\/s10772-017-9442-0"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2015.05.002"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2010-739"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2019-1122"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2011-801"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-30817-9_8"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1155\/2010\/926951"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2018-51"},{"key":"ref35","author":"li","year":"2020","journal-title":"A Mini Review on Current Clinical and Research Findings for Children Suffering From COVID-19"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2012-119"},{"key":"ref28","first-page":"56","article-title":"Evaluation of the pain level from speech: Introducing a novel pain database and benchmarks","author":"ren","year":"2018","journal-title":"Proc ITG Conf Speech Commun"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.21437\/SpeechProsody.2016-86"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1016\/j.cgh.2020.03.043"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/JBHI.2019.2955281"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1002\/9781118706664"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/JBHI.2020.3012666"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1183\/13993003.00547-2020"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1016\/j.csl.2018.02.004"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2020-0032"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2017-43"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1109\/EUSIPCO.2016.7760493"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.3813\/AAA.919357"},{"key":"ref59","doi-asserted-by":"publisher","DOI":"10.1007\/BF00994018"},{"key":"ref58","doi-asserted-by":"publisher","DOI":"10.1016\/j.compbiomed.2018.01.007"},{"key":"ref57","doi-asserted-by":"publisher","DOI":"10.1145\/2502081.2502224"},{"key":"ref56","first-page":"1459","article-title":"openSMILE&#x2014;The Munich versatile and fast open-source audio feature extractor","author":"eyben","year":"2010","journal-title":"Proc ACM MM"},{"key":"ref55","author":"eyben","year":"2015","journal-title":"Real-Time Speech and Music Classification by Large Audio Feature Space Extraction"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2020-2223"},{"key":"ref53","article-title":"Can appliances understand the behaviour of elderly via machine learning? A feasibility study","author":"qian","year":"2020","journal-title":"IEEE Internet of Things Journal"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1016\/j.imu.2020.100378"},{"key":"ref40","first-page":"43","article-title":"Crowd++ unsupervised speaker count with smartphones","author":"xu","year":"2013","journal-title":"Proc UbiComp"},{"key":"ref4","doi-asserted-by":"crossref","first-page":"436","DOI":"10.1038\/nature14539","article-title":"Deep learning","volume":"521","author":"lecun","year":"2015","journal-title":"Nature"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-020-0184-3"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1016\/j.patrec.2020.09.010"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1148\/radiol.2020200905"},{"key":"ref8","author":"farooq","year":"2020","journal-title":"COVID-ResNet A Deep Learning Framework for Screening of COVID19 from Radiographs"},{"key":"ref7","first-page":"1","article-title":"COVID-Net: A tailored deep convolutional neural network design for detection of COVID-19 cases from chest radiography images","volume":"10","author":"wang","year":"2020","journal-title":"Sci Rep"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1109\/ISPACS48206.2019.8986277"},{"key":"ref9","first-page":"396","article-title":"Handwritten digit recognition with a back-propagation network","author":"lecun","year":"1989","journal-title":"Proc NIPS"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN.2015.7280317"},{"key":"ref45","first-page":"1490","article-title":"Automated lung sound analysis in patients with pneumonia","volume":"49","author":"murphy","year":"2004","journal-title":"Respiratory Care"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1007\/s10439-019-02217-0"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/TBME.2016.2619675"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1109\/TBME.2006.873548"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2016-129"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2016.7472920"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/EMBC.2014.6943870"},{"key":"ref73","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"ref72","doi-asserted-by":"publisher","DOI":"10.1145\/3065386"},{"key":"ref71","first-page":"1","article-title":"Very deep convolutional networks for large-scale image recognition","author":"simonyan","year":"2015","journal-title":"Proc ICLR"},{"key":"ref70","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref76","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2014.6854950"},{"key":"ref77","first-page":"1764","article-title":"Towards end-to-end speech recognition with recurrent neural networks","author":"graves","year":"2014","journal-title":"Proc ICML"},{"key":"ref74","first-page":"6340","article-title":"auDeep: Unsupervised learning of representations from audio with deep recurrent neural networks","volume":"18","author":"freitag","year":"2017","journal-title":"J Mach Learn Res"},{"key":"ref75","first-page":"739","article-title":"Off-road obstacle avoidance through end-to-end learning","author":"lecun","year":"2006","journal-title":"Proc NeurIPS"},{"key":"ref78","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2016.7472669"},{"key":"ref79","doi-asserted-by":"publisher","DOI":"10.23919\/EUSIPCO.2019.8902712"},{"key":"ref60","doi-asserted-by":"publisher","DOI":"10.1080\/00437956.1954.11659520"},{"key":"ref62","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2008.111"},{"key":"ref61","doi-asserted-by":"publisher","DOI":"10.4018\/jdet.2013040106"},{"key":"ref63","first-page":"3097","article-title":"Efficient and effective visual codebook generation using additive kernels","volume":"12","author":"wu","year":"2011","journal-title":"J Mach Learn Res"},{"key":"ref64","first-page":"1027","article-title":"K-means++: The advantages of careful seeding","author":"arthur","year":"2007","journal-title":"Proc ACM-SIAM SODA"},{"key":"ref65","first-page":"1","article-title":"openXBOW-Introducing the Passau open-source crossmodal bag-of-words toolkit","volume":"18","author":"schmitt","year":"2017","journal-title":"J Mach Learn Res"},{"key":"ref66","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2009.191"},{"key":"ref67","doi-asserted-by":"crossref","first-page":"3512","DOI":"10.21437\/Interspeech.2017-434","article-title":"Snore sound classification using image-based deep spectrum features","author":"amiriparian","year":"2017","journal-title":"Proc INTERSPEECH"},{"key":"ref68","doi-asserted-by":"publisher","DOI":"10.1109\/EMBC44109.2020.9175450"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-020-0181-6"},{"key":"ref69","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2020-1552"},{"key":"ref1","year":"2020","journal-title":"COVID-19 Case Tracker Follow Global Cases and Trends (Updated Daily)"},{"key":"ref95","first-page":"2672","article-title":"Generative adversarial nets","author":"goodfellow","year":"2014","journal-title":"Proc NIPS"},{"key":"ref94","first-page":"115","article-title":"Cooperative learning and its application to emotion recognition from speech","volume":"23","author":"zhang","year":"2015","journal-title":"IEEE Transactions on Audio Speech and Language Processing"},{"key":"ref93","doi-asserted-by":"publisher","DOI":"10.1121\/1.5004570"},{"key":"ref92","doi-asserted-by":"publisher","DOI":"10.3813\/AAA.919064"},{"key":"ref91","doi-asserted-by":"publisher","DOI":"10.2200\/S00429ED1V01Y201207AIM018"},{"key":"ref90","doi-asserted-by":"publisher","DOI":"10.7551\/mitpress\/9780262033589.001.0001"},{"key":"ref96","doi-asserted-by":"publisher","DOI":"10.1109\/JBHI.2019.2907286"},{"key":"ref97","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2018.2870052"},{"key":"ref10","author":"ge","year":"2020","journal-title":"A Data-Driven Drug Repositioning Framework Discovered a Potential Therapeutic Agent Targeting COVID-19"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.3389\/fimmu.2020.01581"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.3390\/jcm9030674"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-020-0180-7"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1017\/ice.2020.61"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/TAI.2020.3020521"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/MSP.2021.3057298"},{"key":"ref82","first-page":"237","article-title":"Gradient flow in recurrent nets: The difficulty of learning long-term dependencies","author":"hochreiter","year":"2001","journal-title":"A Field Guide to Dynamical Recurrent Neural Networks"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1007\/s10489-020-01862-6"},{"key":"ref81","article-title":"deepSELF: An open source deep self end-to-end learning framework","author":"koike","year":"2020"},{"key":"ref18","article-title":"COVID-19 and computer audition: An overview on what speech & sound analysis could contribute in the SARS-CoV-2 Corona crisis","author":"schuller","year":"0","journal-title":"Frontiers Digit Health"},{"key":"ref84","first-page":"1","article-title":"Empirical evaluation of gated recurrent neural networks on sequence modeling","author":"chung","year":"2014","journal-title":"Proc NIPS DLRL Workshop"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.3389\/fdgth.2020.00005"},{"key":"ref83","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"ref80","doi-asserted-by":"publisher","DOI":"10.1016\/0364-0213(90)90002-E"},{"key":"ref89","doi-asserted-by":"publisher","DOI":"10.2200\/S00196ED1V01Y200906AIM006"},{"key":"ref85","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2009-103"},{"key":"ref86","doi-asserted-by":"publisher","DOI":"10.1162\/089976698300017197"},{"key":"ref87","doi-asserted-by":"publisher","DOI":"10.1145\/1961189.1961199"},{"key":"ref88","first-page":"113","article-title":"Deep sequential image features on acoustic scene classification","author":"ren","year":"2017","journal-title":"Proc DCAS Workshop"}],"container-title":["IEEE Internet of Things Journal"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/6488907\/9585129\/09382380.pdf?arnumber=9382380","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,5,10]],"date-time":"2022-05-10T14:53:33Z","timestamp":1652194413000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9382380\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,11,1]]},"references-count":97,"journal-issue":{"issue":"21"},"URL":"https:\/\/doi.org\/10.1109\/jiot.2021.3067605","relation":{},"ISSN":["2327-4662","2372-2541"],"issn-type":[{"value":"2327-4662","type":"electronic"},{"value":"2372-2541","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021,11,1]]}}}