{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,29]],"date-time":"2025-08-29T10:11:52Z","timestamp":1756462312272,"version":"3.40.3"},"publisher-location":"Cham","reference-count":25,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031775703"},{"type":"electronic","value":"9783031775710"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-77571-0_17","type":"book-chapter","created":{"date-parts":[[2024,12,20]],"date-time":"2024-12-20T02:02:32Z","timestamp":1734660152000},"page":"163-175","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Hierarchical Zero-Shot Approach for\u00a0Human Activity Recognition in\u00a0Smart Homes"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0006-0038-1183","authenticated-orcid":false,"given":"Stefan Gerd","family":"Fritsch","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1870-0203","authenticated-orcid":false,"given":"Federico","family":"Cruciani","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8371-2921","authenticated-orcid":false,"given":"Vitor","family":"Fortes Rey","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2368-7354","authenticated-orcid":false,"given":"Ian","family":"Cleland","sequence":"additional","affiliation":[]},{"given":"Luke","family":"Nugent","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0320-6656","authenticated-orcid":false,"given":"Paul","family":"Lukowicz","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0882-7902","authenticated-orcid":false,"given":"Chris","family":"Nugent","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,12,21]]},"reference":[{"key":"17_CR1","doi-asserted-by":"publisher","first-page":"44949","DOI":"10.1109\/ACCESS.2024.3380432","volume":"12","author":"S Abbas","year":"2024","unstructured":"Abbas, S., et al.: Advancing healthcare and elderly activity recognition: active machine and deep learning for fine-grained heterogeneity activity recognition. IEEE Access 12, 44949\u201344959 (2024)","journal-title":"IEEE Access"},{"key":"17_CR2","doi-asserted-by":"crossref","unstructured":"Chen, L., Hoey, J., Nugent, C.D., Cook, D.J., Yu, Z.: Sensor-based activity recognition. IEEE Trans. Syst. Man. Cybern. Part C (Applications and Reviews) 42(6), 790\u2013808 (2012)","DOI":"10.1109\/TSMCC.2012.2198883"},{"issue":"7","key":"17_CR3","doi-asserted-by":"publisher","first-page":"2203","DOI":"10.3390\/s18072203","volume":"18","author":"F Cruciani","year":"2018","unstructured":"Cruciani, F., Cleland, I., Nugent, C., McCullagh, P., Synnes, K., Hallberg, J.: Automatic annotation for human activity recognition in free living using a smartphone. Sensors 18(7), 2203 (2018)","journal-title":"Sensors"},{"key":"17_CR4","unstructured":"Wang, W.: Sensor-based human activity recognition via zero-shot learning. Ph. D. thesis, Nanyang Technological University, Singapore (2019)"},{"key":"17_CR5","doi-asserted-by":"crossref","unstructured":"Cleland, I.,\u00a0Nugent, L.,\u00a0Cruciani, F.,\u00a0Nugent, C.: Leveraging large language models for activity recognition in smart environments. In: 6th IEEE International Conference on Activity and Behavior Computing (2024)","DOI":"10.1109\/ABC61795.2024.10651718"},{"key":"17_CR6","first-page":"4534","volume":"2024","author":"TH Nguyen","year":"2024","unstructured":"Nguyen, T.H., Rudra, K.: Human vs ChatGPT: effect of data annotation in interpretable crisis-related microblog classification. Proc. ACM Web Conf. 2024, 4534\u20134543 (2024)","journal-title":"Proc. ACM Web Conf."},{"key":"17_CR7","doi-asserted-by":"crossref","unstructured":"Van\u00a0Kasteren, T.,\u00a0Noulas, A.,\u00a0Englebienne, G.,\u00a0Kr\u00f6se, B.: Accurate activity recognition in a home setting. In: Proceedings of the 10th International Conference on Ubiquitous Computing, pp.\u00a01\u20139. ACM (2008)","DOI":"10.1145\/1409635.1409637"},{"key":"17_CR8","doi-asserted-by":"crossref","unstructured":"van Kasteren, T.L.,\u00a0Englebienne, G., Kr\u00f6se, B.J.: Human activity recognition from wireless sensor network data: benchmark and software. In: Activity Recognition in Pervasive Intelligent Environments, pp.\u00a0165\u2013186. Springer (2011)","DOI":"10.2991\/978-94-91216-05-3_8"},{"key":"17_CR9","doi-asserted-by":"publisher","first-page":"107561","DOI":"10.1016\/j.patcog.2020.107561","volume":"108","author":"LM Dang","year":"2020","unstructured":"Dang, L.M., Min, K., Wang, H., Piran, M.J., Lee, C.H., Moon, H.: Sensor-based and vision-based human activity recognition: a comprehensive survey. Pattern Recogn. 108, 107561 (2020)","journal-title":"Pattern Recogn."},{"issue":"18","key":"17_CR10","doi-asserted-by":"publisher","first-page":"6037","DOI":"10.3390\/s21186037","volume":"21","author":"D Bouchabou","year":"2021","unstructured":"Bouchabou, D., Nguyen, S.M., Lohr, C., LeDuc, B., Kanellos, I.: A survey of human activity recognition in smart homes based on IoT sensors algorithms: taxonomies, challenges, and opportunities with deep learning. Sensors 21(18), 6037 (2021)","journal-title":"Sensors"},{"issue":"11","key":"17_CR11","doi-asserted-by":"publisher","first-page":"13521","DOI":"10.1007\/s10462-023-10466-8","volume":"56","author":"SF Ahmed","year":"2023","unstructured":"Ahmed, S.F., et al.: Deep learning modelling techniques: current progress, applications, advantages, and challenges. Artif. Intell. Rev. 56(11), 13521\u201313617 (2023)","journal-title":"Artif. Intell. Rev."},{"issue":"7553","key":"17_CR12","doi-asserted-by":"publisher","first-page":"436","DOI":"10.1038\/nature14539","volume":"521","author":"Y LeCun","year":"2015","unstructured":"LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521(7553), 436\u2013444 (2015)","journal-title":"Nature"},{"key":"17_CR13","doi-asserted-by":"crossref","unstructured":"Bouchabou, D., Nguyen, S.M.,\u00a0Lohr, C.,\u00a0LeDuc, B.,\u00a0Kanellos, I.: Fully convolutional network bootstrapped by word encoding and embedding for activity recognition in smart homes. In: Deep Learning for Human Activity Recognition, pp.\u00a0111\u2013125. Springer (2021)","DOI":"10.1007\/978-981-16-0575-8_9"},{"key":"17_CR14","unstructured":"Hegselmann, S.,\u00a0Buendia, A.,\u00a0Lang, H.,\u00a0Agrawal, M.,\u00a0Jiang, X.,\u00a0Sontag, D.: TabLLM: few-shot classification of tabular data with large language models. In: International Conference on Artificial Intelligence and Statistics, pp.\u00a05549\u20135581. PMLR (2023)"},{"key":"17_CR15","unstructured":"Liu, X., et al.: Large language models are few-shot health learners. arXiv preprint arXiv:2305.15525 (2023)"},{"key":"17_CR16","doi-asserted-by":"crossref","unstructured":"Gao, J.,\u00a0Zhang, Y.,\u00a0Chen, Y.,\u00a0Zhang, T.,\u00a0Tang, B.,\u00a0Wang, X.: Unsupervised human activity recognition via large language models and iterative evolution. In: ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp.\u00a091\u201395. IEEE (2024)","DOI":"10.1109\/ICASSP48485.2024.10446819"},{"key":"17_CR17","unstructured":"Civitarese, G.,\u00a0Fiori, M.,\u00a0Choudhary, P.,\u00a0Bettini, C.: Large language models are zero-shot recognizers for activities of daily living. arXiv preprint arXiv:2407.01238 (2024)"},{"key":"17_CR18","doi-asserted-by":"crossref","unstructured":"Elhoseiny, M.,\u00a0Zhu, Y.,\u00a0Zhang, H.,\u00a0Elgammal, A.: Link the head to the \u201cbeak\u201d: zero shot learning from noisy text description at part precision. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.\u00a06288\u20136297 (2017)","DOI":"10.1109\/CVPR.2017.666"},{"key":"17_CR19","doi-asserted-by":"crossref","unstructured":"Wang, Q.,\u00a0Chen, K.: Alternative semantic representations for zero-shot human action recognition. In: Machine Learning and Knowledge Discovery in Databases, pp.\u00a087\u2013102. Springer (2017)","DOI":"10.1007\/978-3-319-71249-9_6"},{"key":"17_CR20","doi-asserted-by":"crossref","unstructured":"Verma, V.K.,\u00a0Rai, P.: A simple exponential family framework for zero-shot learning. In: Machine Learning and Knowledge Discovery in Databases, pp.\u00a0792\u2013808. Springer (2017)","DOI":"10.1007\/978-3-319-71246-8_48"},{"key":"17_CR21","doi-asserted-by":"publisher","first-page":"309","DOI":"10.1007\/s11263-016-0983-5","volume":"123","author":"X Xu","year":"2017","unstructured":"Xu, X., Hospedales, T., Gong, S.: Transductive zero-shot action recognition by word-vector embedding. Int. J. Comput. Vis. 123, 309\u2013333 (2017)","journal-title":"Int. J. Comput. Vis."},{"issue":"9","key":"17_CR22","doi-asserted-by":"publisher","first-page":"28147","DOI":"10.1007\/s11042-023-16566-5","volume":"83","author":"V Estevam","year":"2024","unstructured":"Estevam, V., Laroca, R., Pedrini, H., Menotti, D.: Tell me what you see: a zero-shot action recognition method based on natural language descriptions. Multimedia Tools Appl. 83(9), 28147\u201328173 (2024)","journal-title":"Multimedia Tools Appl."},{"key":"17_CR23","doi-asserted-by":"publisher","first-page":"5188","DOI":"10.1109\/ACCESS.2024.3349510","volume":"12","author":"K Huang","year":"2024","unstructured":"Huang, K., Mckeever, S., Miralles-Pechu\u00e1n, L.: Generalised zero-shot learning for action recognition fusing text and image GANs. IEEE Access 12, 5188\u20135202 (2024)","journal-title":"IEEE Access"},{"issue":"22","key":"17_CR24","doi-asserted-by":"publisher","first-page":"5043","DOI":"10.3390\/s19225043","volume":"19","author":"M Matsuki","year":"2019","unstructured":"Matsuki, M., Lago, P., Inoue, S.: Characterizing word embeddings for zero-shot sensor-based human activity recognition. Sensors 19(22), 5043 (2019)","journal-title":"Sensors"},{"key":"17_CR25","doi-asserted-by":"crossref","unstructured":"Reimers, N.,\u00a0Gurevych, I.: Sentence-BERT: sentence embeddings using siamese BERT-networks. arXiv preprint arXiv:1908.10084 (2019)","DOI":"10.18653\/v1\/D19-1410"}],"container-title":["Lecture Notes in Networks and Systems","Proceedings of the International Conference on Ubiquitous Computing and Ambient Intelligence (UCAmI 2024)"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-77571-0_17","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,20]],"date-time":"2024-12-20T02:06:22Z","timestamp":1734660382000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-77571-0_17"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9783031775703","9783031775710"],"references-count":25,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-77571-0_17","relation":{},"ISSN":["2367-3370","2367-3389"],"issn-type":[{"type":"print","value":"2367-3370"},{"type":"electronic","value":"2367-3389"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"21 December 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"UCAmI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Ubiquitous Computing and Ambient Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Belfast","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"United Kingdom","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 November 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 November 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ucami2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/ucami.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}