{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,30]],"date-time":"2025-08-30T16:56:22Z","timestamp":1756572982265,"version":"3.40.3"},"publisher-location":"Cham","reference-count":21,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031775703"},{"type":"electronic","value":"9783031775710"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-77571-0_38","type":"book-chapter","created":{"date-parts":[[2024,12,20]],"date-time":"2024-12-20T02:02:07Z","timestamp":1734660127000},"page":"393-405","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["MeciFace: Mechanomyography and\u00a0Inertial Fusion-Based Glasses for\u00a0Edge Real-Time Recognition of\u00a0Facial and\u00a0Eating Activities"],"prefix":"10.1007","author":[{"given":"Hymalai","family":"Bello","sequence":"first","affiliation":[]},{"given":"Sungho","family":"Suh","sequence":"additional","affiliation":[]},{"given":"Bo","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Paul","family":"Lukowicz","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,12,21]]},"reference":[{"key":"38_CR1","doi-asserted-by":"crossref","unstructured":"Aoki, H., Ohnishi, A., Isoyama, N., Terada, T., Tsukamoto, M.: FaceRecGlasses: a wearable system for recognizing self facial expressions using compact wearable cameras. In: Augmented Humans Conference 2021, pp. 55\u201365 (2021)","DOI":"10.1145\/3458709.3458983"},{"key":"38_CR2","doi-asserted-by":"crossref","unstructured":"Bedri, A., Li, D., Khurana, R., Bhuwalka, K., Goel, M.: FitByte: automatic diet monitoring in unconstrained situations using multimodal sensing on eyeglasses. In: Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, pp. 1\u201312 (2020)","DOI":"10.1145\/3313831.3376869"},{"key":"38_CR3","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2023.101886","volume":"99","author":"H Bello","year":"2023","unstructured":"Bello, H., Marin, L.A.S., Suh, S., Zhou, B., Lukowicz, P.: InMyFace: inertial and mechanomyography-based sensor fusion for wearable facial activity recognition. Inf. Fusion 99, 101886 (2023)","journal-title":"Inf. Fusion"},{"issue":"17","key":"38_CR4","doi-asserted-by":"publisher","first-page":"4904","DOI":"10.3390\/s20174904","volume":"20","author":"H Bello","year":"2020","unstructured":"Bello, H., Zhou, B., Lukowicz, P.: Facial muscle activity recognition with reconfigurable differential stethoscope-microphones. Sensors 20(17), 4904 (2020)","journal-title":"Sensors"},{"key":"38_CR5","doi-asserted-by":"crossref","unstructured":"Chen, T., Steeper, B., Alsheikh, K., Tao, S., Guimbreti\u00e8re, F., Zhang, C.: C-Face: continuously reconstructing facial expressions by deep learning contours of the face with ear-mounted miniature cameras. In: Proceedings of the 33rd Annual ACM Symposium on User Interface Software and Technology, pp. 112\u2013125 (2020)","DOI":"10.1145\/3379337.3415879"},{"key":"38_CR6","doi-asserted-by":"crossref","unstructured":"Gjoreski, H., et\u00a0al.: OCOsense glasses\u2013monitoring facial gestures and expressions for augmented human-computer interaction: OCOsense glasses for monitoring facial gestures and expressions. In: Extended Abstracts of the 2023 CHI Conference on Human Factors in Computing Systems, pp.\u00a01\u20134 (2023)","DOI":"10.1145\/3544549.3583918"},{"issue":"1","key":"38_CR7","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s12903-022-02301-3","volume":"22","author":"IJ Kwon","year":"2022","unstructured":"Kwon, I.J., Jung, T.Y., Son, Y., Kim, B., Kim, S.M., Lee, J.H.: Detection of volatile sulfur compounds (VSCS) in exhaled breath as a potential diagnostic method for oral squamous cell carcinoma. BMC Oral Health 22(1), 1\u20138 (2022)","journal-title":"BMC Oral Health"},{"key":"38_CR8","doi-asserted-by":"publisher","first-page":"146392","DOI":"10.1109\/ACCESS.2021.3121543","volume":"9","author":"J Kwon","year":"2021","unstructured":"Kwon, J., Ha, J., Kim, D.H., Choi, J.W., Kim, L.: Emotion recognition using a glasses-type wearable device via multi-channel facial responses. IEEE Access 9, 146392\u2013146403 (2021)","journal-title":"IEEE Access"},{"key":"38_CR9","doi-asserted-by":"publisher","DOI":"10.3389\/frai.2020.609673","volume":"3","author":"JR Lee","year":"2021","unstructured":"Lee, J.R., Wang, L., Wong, A.: EmotionNet nano: an efficient deep convolutional neural network design for real-time facial expression recognition. Front. Artif. Intell. 3, 609673 (2021)","journal-title":"Front. Artif. Intell."},{"key":"38_CR10","doi-asserted-by":"crossref","unstructured":"Masai, K., Kunze, K., Sakamoto, D., Sugiura, Y., Sugimoto, M.: Face commands-user-defined facial gestures for smart glasses. In: 2020 IEEE International Symposium on Mixed and Augmented Reality (ISMAR), pp. 374\u2013386. IEEE (2020)","DOI":"10.1109\/ISMAR50242.2020.00064"},{"issue":"4","key":"38_CR11","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3012941","volume":"7","author":"K Masai","year":"2017","unstructured":"Masai, K., Kunze, K., Sugiura, Y., Ogata, M., Inami, M., Sugimoto, M.: Evaluation of facial expression recognition by a smart eyewear for facial direction changes, repeatability, and positional drift. ACM Trans. Interact. Intell. Syst. (TiiS) 7(4), 1\u201323 (2017)","journal-title":"ACM Trans. Interact. Intell. Syst. (TiiS)"},{"key":"38_CR12","doi-asserted-by":"crossref","unstructured":"Meyer, J., Frank, A., Schlebusch, T., Kasneci, E.: U-HAR: a convolutional approach to human activity recognition combining head and eye movements for context-aware smart glasses. Proc. ACM Hum.-Comput. Interact. 6(ETRA), 1\u201319 (2022)","DOI":"10.1145\/3530884"},{"issue":"4","key":"38_CR13","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3533390","volume":"3","author":"MB Morshed","year":"2022","unstructured":"Morshed, M.B., et al.: Food, mood, context: examining college students\u2019 eating context and mental well-being. ACM Trans. Comput. Healthc. 3(4), 1\u201326 (2022)","journal-title":"ACM Trans. Comput. Healthc."},{"key":"38_CR14","doi-asserted-by":"publisher","DOI":"10.1016\/j.pmcj.2021.101424","volume":"75","author":"J Nie","year":"2021","unstructured":"Nie, J., et al.: SPIDERS+: a light-weight, wireless, and low-cost glasses-based wearable platform for emotion sensing and bio-signal acquisition. Pervasive Mob. Comput. 75, 101424 (2021)","journal-title":"Pervasive Mob. Comput."},{"key":"38_CR15","doi-asserted-by":"publisher","first-page":"1516","DOI":"10.3389\/fpsyg.2014.01516","volume":"5","author":"M Olszanowski","year":"2015","unstructured":"Olszanowski, M., Pochwatko, G., Kuklinski, K., Scibor-Rylski, M., Lewinski, P., Ohme, R.K.: Warsaw set of emotional facial expression pictures: a validation study of facial display photographs. Front. Psychol. 5, 1516 (2015)","journal-title":"Front. Psychol."},{"key":"38_CR16","doi-asserted-by":"crossref","unstructured":"Rachakonda, L., Kothari, A., Mohanty, S.P., Kougianos, E., Ganapathiraju, M.: Stress-log: an IoT-based smart system to monitor stress-eating. In: 2019 IEEE International Conference on Consumer Electronics (ICCE), pp.\u00a01\u20136. IEEE (2019)","DOI":"10.1109\/ICCE.2019.8661959"},{"key":"38_CR17","doi-asserted-by":"crossref","unstructured":"Shin, J., et al.: MyDJ: sensing food intakes with an attachable on your eyeglass frame. In: Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems, pp. 1\u201317 (2022)","DOI":"10.1145\/3491102.3502041"},{"key":"38_CR18","doi-asserted-by":"crossref","unstructured":"Song, X., Huang, K., Gao, W.: FaceListener: recognizing human facial expressions via acoustic sensing on commodity headphones. In: 2022 21st ACM\/IEEE International Conference on Information Processing in Sensor Networks (IPSN), pp. 145\u2013157. IEEE (2022)","DOI":"10.1109\/IPSN54338.2022.00019"},{"key":"38_CR19","doi-asserted-by":"publisher","unstructured":"Verma, D.: ExpressEar: sensing fine-grained facial expressions with earables. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 5, 129:1\u2013129:28 (2021). https:\/\/doi.org\/10.1145\/3478085","DOI":"10.1145\/3478085"},{"issue":"2","key":"38_CR20","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3448105","volume":"5","author":"W Xie","year":"2021","unstructured":"Xie, W., Zhang, Q., Zhang, J.: Acoustic-based upper facial action recognition for smart eyewear. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 5(2), 1\u201328 (2021)","journal-title":"Proc. ACM Interact. Mob. Wearable Ubiquitous Technol."},{"key":"38_CR21","doi-asserted-by":"crossref","unstructured":"Zhang, R., et al.: EchoSpeech: continuous silent speech recognition on minimally-obtrusive eyewear powered by acoustic sensing. In: Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, pp. 1\u201318 (2023)","DOI":"10.1145\/3544548.3580801"}],"container-title":["Lecture Notes in Networks and Systems","Proceedings of the International Conference on Ubiquitous Computing and Ambient Intelligence (UCAmI 2024)"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-77571-0_38","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,20]],"date-time":"2024-12-20T02:08:49Z","timestamp":1734660529000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-77571-0_38"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9783031775703","9783031775710"],"references-count":21,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-77571-0_38","relation":{},"ISSN":["2367-3370","2367-3389"],"issn-type":[{"type":"print","value":"2367-3370"},{"type":"electronic","value":"2367-3389"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"21 December 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"UCAmI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Ubiquitous Computing and Ambient Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Belfast","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"United Kingdom","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 November 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 November 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ucami2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/ucami.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}