{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T03:06:07Z","timestamp":1775012767940,"version":"3.50.1"},"reference-count":47,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2024,12,5]],"date-time":"2024-12-05T00:00:00Z","timestamp":1733356800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2024,12,5]],"date-time":"2024-12-05T00:00:00Z","timestamp":1733356800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/100000865","name":"Bill and Melinda Gates Foundation","doi-asserted-by":"publisher","award":["OPP1171395"],"award-info":[{"award-number":["OPP1171395"]}],"id":[{"id":"10.13039\/100000865","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["npj Digit. Med."],"abstract":"<jats:title>Abstract<\/jats:title><jats:p>We have developed a population-level method for dietary assessment using low-cost wearable cameras. Our approach, EgoDiet, employs an egocentric vision-based pipeline to learn portion sizes, addressing the shortcomings of traditional self-reported dietary methods. To evaluate the functionality of this method, field studies were conducted in London (Study A) and Ghana (Study B) among populations of Ghanaian and Kenyan origin. In Study A, EgoDiet\u2019s estimations were contrasted with dietitians\u2019 assessments, revealing a performance with a Mean Absolute Percentage Error (MAPE) of 31.9% for portion size estimation, compared to 40.1% for estimates made by dietitians. We further evaluated our approach in Study B, comparing its performance to the traditional 24-Hour Dietary Recall (24HR). Our approach demonstrated a MAPE of 28.0%, showing a reduction in error when contrasted with the 24HR, which exhibited a MAPE of 32.5%. This improvement highlights the potential of using passive camera technology to serve as an alternative to the traditional dietary assessment methods.<\/jats:p>","DOI":"10.1038\/s41746-024-01346-8","type":"journal-article","created":{"date-parts":[[2024,12,5]],"date-time":"2024-12-05T17:15:14Z","timestamp":1733418914000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":15,"title":["AI-enabled wearable cameras for assisting dietary assessment in African populations"],"prefix":"10.1038","volume":"7","author":[{"given":"Frank P.-W.","family":"Lo","sequence":"first","affiliation":[]},{"given":"Jianing","family":"Qiu","sequence":"additional","affiliation":[]},{"given":"Modou L.","family":"Jobarteh","sequence":"additional","affiliation":[]},{"given":"Yingnan","family":"Sun","sequence":"additional","affiliation":[]},{"given":"Zeyu","family":"Wang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3645-6301","authenticated-orcid":false,"given":"Shuo","family":"Jiang","sequence":"additional","affiliation":[]},{"given":"Tom","family":"Baranowski","sequence":"additional","affiliation":[]},{"given":"Alex K.","family":"Anderson","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5024-1465","authenticated-orcid":false,"given":"Megan A.","family":"McCrory","sequence":"additional","affiliation":[]},{"given":"Edward","family":"Sazonov","sequence":"additional","affiliation":[]},{"given":"Wenyan","family":"Jia","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7948-9205","authenticated-orcid":false,"given":"Mingui","family":"Sun","sequence":"additional","affiliation":[]},{"given":"Matilda","family":"Steiner-Asiedu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0529-6325","authenticated-orcid":false,"given":"Gary","family":"Frost","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5080-108X","authenticated-orcid":false,"given":"Benny","family":"Lo","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,12,5]]},"reference":[{"key":"1346_CR1","doi-asserted-by":"crossref","unstructured":"Jobarteh, M. L., McCrory, M. A., Lo, B. et al. Development and validation of an objective, passive dietary assessment method for estimating food and nutrient intake in households in low-and middle-income countries: a study protocol. Curr. Dev. Nutr. 4, nzaa020 (2020).","DOI":"10.1093\/cdn\/nzaa020"},{"key":"1346_CR2","doi-asserted-by":"crossref","unstructured":"He, K., Gkioxari, G., Doll\u00e1r, P. & Girshick, R. Mask r-cnn. 2961\u20132969 (IEEE ICCV, 2017).","DOI":"10.1109\/ICCV.2017.322"},{"key":"1346_CR3","doi-asserted-by":"crossref","unstructured":"Qiu, J. et al. Egocentric image captioning for privacy-preserved passive dietary intake monitoring. IEEE Transactions on Cybernetics (2023).","DOI":"10.1109\/TCYB.2023.3243999"},{"key":"1346_CR4","doi-asserted-by":"publisher","DOI":"10.1038\/s41598-023-47885-0","volume":"13","author":"FS Konstantakopoulos","year":"2023","unstructured":"Konstantakopoulos, F. S., Georga, E. I. & Fotiadis, D. I. A novel approach to estimate the weight of food items based on features extracted from an image using boosting algorithms. Sci. Rep. 13, 21040 (2023).","journal-title":"Sci. Rep."},{"key":"1346_CR5","first-page":"577","volume":"16","author":"FP-W Lo","year":"2019","unstructured":"Lo, F. P.-W., Sun, Y., Qiu, J. & Lo, B. P. Point2volume: a vision-based dietary assessment approach using view synthesis. IEEE TII 16, 577\u2013586 (2019).","journal-title":"IEEE TII"},{"key":"1346_CR6","first-page":"1090","volume":"19","author":"J Dehais","year":"2017","unstructured":"Dehais, J., Anthimopoulos, M., Shevchik, S. & Mougiakakou, S. Two-view 3d reconstruction for food volume estimation. IEEE TMM 19, 1090\u20131099 (2017).","journal-title":"IEEE TMM"},{"key":"1346_CR7","doi-asserted-by":"crossref","unstructured":"Rahman, M. H. et al. Food volume estimation in a mobile phone based dietary assessment system. In IEEE SITIS, 988\u2013995 (IEEE, 2012).","DOI":"10.1109\/SITIS.2012.146"},{"key":"1346_CR8","doi-asserted-by":"crossref","unstructured":"Fang, S., Liu, C., Zhu, F., Delp, E. J. & Boushey, C. J. Single-view food portion estimation based on geometric models. 385\u2013390 (IEEE ISM, 2015).","DOI":"10.1109\/ISM.2015.67"},{"key":"1346_CR9","doi-asserted-by":"publisher","first-page":"1671","DOI":"10.1017\/S1368980013003236","volume":"17","author":"W Jia","year":"2014","unstructured":"Jia, W., Chen, H.-C. & Yue, Y. et al. Accuracy of food portion size estimation from digital pictures acquired by a chest-worn camera. Public Health Nutr. 17, 1671\u20131681 (2014).","journal-title":"Public Health Nutr."},{"key":"1346_CR10","doi-asserted-by":"crossref","unstructured":"Xu, C., He, Y., Khanna, N., Boushey, C. J. & Delp, E. J. Model-based food volume estimation using 3d pose. 2534\u20132538 (IEEE ICIP, 2013).","DOI":"10.1109\/ICIP.2013.6738522"},{"key":"1346_CR11","first-page":"1180","volume":"22","author":"Y Yang","year":"2019","unstructured":"Yang, Y., Jia, W., Bucher, T., Zhang, H. & Sun, M. Image-based food portion size estimation using a smartphone without a fiducial marker. Public Health Nutr. 22, 1180\u20131192 (2019).","journal-title":"Public Health Nutr."},{"key":"1346_CR12","doi-asserted-by":"publisher","first-page":"76 \u2013 86","DOI":"10.1016\/j.jfoodeng.2011.09.031","volume":"109","author":"W Jia","year":"2012","unstructured":"Jia, W. et al. Image based estimation of food volume using circular referents in dietary assessment. J. Food Eng. 109, 76 \u2013 86 (2012).","journal-title":"J. Food Eng."},{"key":"1346_CR13","doi-asserted-by":"crossref","unstructured":"Lo, F. P. W., Sun, Y., Qiu, J. & Lo, B. Image-based food classification and volume estimation for dietary assessment: a review. (IEEE JBHI, 2020).","DOI":"10.1109\/JBHI.2020.2987943"},{"key":"1346_CR14","first-page":"377","volume":"19","author":"F Zhu","year":"2014","unstructured":"Zhu, F., Bosch, M., Khanna, N., Boushey, C. J. & Delp, E. J. Multiple hypotheses image segmentation and classification with application to dietary assessment. IEEE JBHI 19, 377\u2013388 (2014).","journal-title":"IEEE JBHI"},{"key":"1346_CR15","doi-asserted-by":"crossref","unstructured":"Zhu, F., Bosch, M., Boushey, C. J. & Delp, E. J. An image analysis system for dietary assessment and evaluation. In 2010 ICIP, 1853\u20131856 (IEEE, 2010).","DOI":"10.1109\/ICIP.2010.5650848"},{"key":"1346_CR16","first-page":"756","volume":"4","author":"F Zhu","year":"2010","unstructured":"Zhu, F. et al. The use of mobile devices in aiding dietary assessment and evaluation. IEEE J.-STSP 4, 756\u2013766 (2010).","journal-title":"IEEE J.-STSP"},{"key":"1346_CR17","doi-asserted-by":"crossref","unstructured":"Meyers, A. et al. Im2calories: towards an automated mobile vision food diary. 1233\u20131241 (ICCV, 2015).","DOI":"10.1109\/ICCV.2015.146"},{"key":"1346_CR18","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1260\/2040-2295.6.1.1","volume":"6","author":"M Sun","year":"2015","unstructured":"Sun, M., Burke, L. E. & Baranowski, T. et al. An exploratory study on a chest-worn computer for evaluation of diet, physical activity and lifestyle. J. Health. Eng. 6, 1\u201322 (2015).","journal-title":"J. Health. Eng."},{"key":"1346_CR19","doi-asserted-by":"crossref","unstructured":"Lo, F. P.-W., Sun, Y. & Lo, B. Depth estimation based on a single close-up image with volumetric annotations in the wild: A pilot study. In IEEE AIM, 513\u2013518 (IEEE, 2019).","DOI":"10.1109\/AIM.2019.8868629"},{"key":"1346_CR20","doi-asserted-by":"crossref","unstructured":"Doulah, A., Ghosh, T., Hossain, D., Imtiaz, M. H. & Sazonov, E. Automatic ingestion monitor version 2\u201d-a novel wearable device for automatic food intake detection and passive capture of food images. (IEEE JBHI, 2020).","DOI":"10.1109\/JBHI.2020.2995473"},{"key":"1346_CR21","doi-asserted-by":"crossref","unstructured":"Sun, M. et al. ebutton: a wearable computer for health monitoring and personal assistance. In Proceedings of the 51st Annual Design Automation Conference, 1\u20136 (2014).","DOI":"10.1145\/2593069.2596678"},{"key":"1346_CR22","doi-asserted-by":"publisher","first-page":"4075","DOI":"10.3390\/nu15184075","volume":"15","author":"ML Jobarteh","year":"2023","unstructured":"Jobarteh, M. L. et al. Evaluation of acceptability, functionality, and validity of a passive image-based dietary intake assessment method in adults and children of Ghanaian and Kenyan origin living in London, UK. Nutrients 15, 4075 (2023).","journal-title":"Nutrients"},{"key":"1346_CR23","doi-asserted-by":"crossref","unstructured":"Ferdinand Christ, P. et al. Diabetes60 - inferring bread units from food images using fully convolutional neural networks. (ICCV, 2017).","DOI":"10.1109\/ICCVW.2017.180"},{"key":"1346_CR24","doi-asserted-by":"crossref","unstructured":"Graikos, A., Charisis, V., Iakovakis, D., Hadjidimitriou, S. & Hadjileontiadis, L. Single image-based food volume estimation using monocular depth-prediction networks. In Universal Access in Human-Computer Interaction. Applications and Practice: 14th International Conference, UAHCI 2020, Held as Part of the 22nd HCI International Conference, HCII 2020, Copenhagen, Denmark, July 19\u201324, 2020, Proceedings, Part II 22, 532\u2013543 (Springer, 2020).","DOI":"10.1007\/978-3-030-49108-6_38"},{"key":"1346_CR25","doi-asserted-by":"crossref","unstructured":"Lu, Y., Stathopoulou, T. & Mougiakakou, S. Partially supervised multi-task network for single-view dietary assessment. In 2020 25th International Conference on Pattern Recognition (ICPR), 8156\u20138163 (IEEE, 2021).","DOI":"10.1109\/ICPR48806.2021.9412339"},{"key":"1346_CR26","doi-asserted-by":"publisher","first-page":"160","DOI":"10.1017\/S0007114515004262","volume":"115","author":"C Pettitt","year":"2016","unstructured":"Pettitt, C. et al. A pilot study to determine whether using a lightweight, wearable micro-camera improves dietary assessment accuracy and offers information on macronutrients and eating rate. Br. J. Nutr. 115, 160\u2013167 (2016).","journal-title":"Br. J. Nutr."},{"key":"1346_CR27","doi-asserted-by":"publisher","first-page":"919","DOI":"10.1079\/PHN2001124","volume":"4","author":"G Johansson","year":"2001","unstructured":"Johansson, G., Wikman, \u212b, \u212bhr\u00e9n, A.-M., Hallmans, G. & Johansson, I. Underreporting of energy intake in repeated 24-hour recalls related to gender, age, weight status, day of interview, educational level, reported food intake, smoking habits and area of living. Public Health Nutr. 4, 919\u2013927 (2001).","journal-title":"Public Health Nutr."},{"key":"1346_CR28","unstructured":"Liu, H., Li, C., Wu, Q. & Lee, Y. J. Visual instruction tuning (2023)."},{"key":"1346_CR29","doi-asserted-by":"crossref","unstructured":"Kirillov, A. et al. Segment anything. In Proceedings of the IEEE\/CVF International Conference on Computer Vision, 4015\u20134026 (2023).","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"1346_CR30","doi-asserted-by":"crossref","unstructured":"Yang, L. et al. Depth anything: Unleashing the power of large-scale unlabeled data. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 10371\u201310381 (2024).","DOI":"10.1109\/CVPR52733.2024.00987"},{"key":"1346_CR31","doi-asserted-by":"publisher","first-page":"741","DOI":"10.3390\/nu10060741","volume":"10","author":"MF Vasiloglou","year":"2018","unstructured":"Vasiloglou, M. F. et al. A comparative study on carbohydrate estimation: Gocarb vs. dietitians. Nutrients 10, 741 (2018).","journal-title":"Nutrients"},{"key":"1346_CR32","doi-asserted-by":"publisher","first-page":"732449","DOI":"10.3389\/fnut.2021.732449","volume":"8","author":"P Chotwanvirat","year":"2021","unstructured":"Chotwanvirat, P., Hnoohom, N., Rojroongwasinkul, N. & Kriengsinyos, W. Feasibility study of an automated carbohydrate estimation system using Thai food images in comparison with estimation by dietitians. Front. Nutr. 8, 732449 (2021).","journal-title":"Front. Nutr."},{"key":"1346_CR33","doi-asserted-by":"crossref","unstructured":"Qiu, J., Lo, F. P.-W. & Lo, B. Assessing individual dietary intake in food sharing scenarios with a 360 camera and deep learning. In IEEE BSN, 1\u20134 (IEEE, 2019).","DOI":"10.1109\/BSN.2019.8771095"},{"key":"1346_CR34","doi-asserted-by":"crossref","unstructured":"Nie, J. et al. Automatic detection of dining plates for image-based dietary evaluation. 4312\u20134315 (IEEE EMBC, 2010).","DOI":"10.1109\/IEMBS.2010.5626204"},{"key":"1346_CR35","first-page":"1168","volume":"22","author":"W Jia","year":"2019","unstructured":"Jia, W. et al. Automatic food detection in egocentric images using artificial intelligence technology. Public health Nutr. 22, 1168\u20131179 (2019).","journal-title":"Public health Nutr."},{"key":"1346_CR36","doi-asserted-by":"crossref","unstructured":"Lo, F. P. W., Guo, Y., Sun, Y., Qiu, J. & Lo, B. An intelligent vision-based nutritional assessment method for handheld food items. (IEEE Transactions on Multimedia, 2022).","DOI":"10.1109\/TMM.2022.3199911"},{"key":"1346_CR37","doi-asserted-by":"crossref","unstructured":"Lin, T.-Y. et al. Microsoft coco: Common objects in context. In ECCV, 740\u2013755 (Springer, 2014).","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"1346_CR38","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S. & Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, 770\u2013778 (2016).","DOI":"10.1109\/CVPR.2016.90"},{"key":"1346_CR39","doi-asserted-by":"crossref","unstructured":"Lin, T.-Y. et al. Feature pyramid networks for object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition, 2117\u20132125 (2017).","DOI":"10.1109\/CVPR.2017.106"},{"key":"1346_CR40","unstructured":"Abdulla, W. Mask r-cnn for object detection and instance segmentation on keras and tensorflow. https:\/\/github.com\/matterport\/Mask_RCNN (2017)."},{"key":"1346_CR41","doi-asserted-by":"crossref","unstructured":"Lei, J., Qiu, J., Lo, F. P.-W. & Lo, B. Assessing individual dietary intake in food sharing scenarios with food and human pose detection. In Pattern Recognition. ICPR International Workshops and Challenges: Virtual Event, January 10\u201315, 2021, Proceedings, Part V, 549\u2013557 (Springer, 2021).","DOI":"10.1007\/978-3-030-68821-9_45"},{"key":"1346_CR42","doi-asserted-by":"crossref","unstructured":"Gao, A., Lo, F. P. W. & Lo, B. Food volume estimation for quantifying dietary intake with a wearable camera. 110\u2013113 (IEEE BSN, 2018).","DOI":"10.1109\/BSN.2018.8329671"},{"key":"1346_CR43","doi-asserted-by":"crossref","unstructured":"Fang, S. et al. A comparison of food portion size estimation using geometric models and depth images. 26\u201330 (IEEE ICIP, 2016).","DOI":"10.1109\/ICIP.2016.7532312"},{"key":"1346_CR44","first-page":"1947","volume":"63","author":"P Pouladzadeh","year":"2014","unstructured":"Pouladzadeh, P., Shirmohammadi, S. & Al-Maghrabi, R. Measuring calorie and nutrition from food image. IEEE TIM 63, 1947\u20131956 (2014).","journal-title":"IEEE TIM"},{"key":"1346_CR45","doi-asserted-by":"crossref","unstructured":"Chokr, M. & Elbassuoni, S. Calories prediction from food images. In Proceedings of the AAAI Conference on Artificial Intelligence, vol. 31, 4664\u20134669 (2017).","DOI":"10.1609\/aaai.v31i2.19092"},{"key":"1346_CR46","doi-asserted-by":"crossref","unstructured":"Fang, S. et al. Single-view food portion estimation: Learning image-to-energy mappings using generative adversarial networks. In IEEE ICIP, 251\u2013255 (IEEE, 2018).","DOI":"10.1109\/ICIP.2018.8451461"},{"key":"1346_CR47","doi-asserted-by":"publisher","DOI":"10.1038\/s41598-021-03972-8","volume":"12","author":"KJ Pfisterer","year":"2022","unstructured":"Pfisterer, K. J. et al. Automated food intake tracking requires depth-refined semantic segmentation to rectify visual-volume discordance in long-term care homes. Sci. Rep. 12, 83 (2022).","journal-title":"Sci. Rep."}],"container-title":["npj Digital Medicine"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.nature.com\/articles\/s41746-024-01346-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.nature.com\/articles\/s41746-024-01346-8","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.nature.com\/articles\/s41746-024-01346-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,5]],"date-time":"2024-12-05T18:09:56Z","timestamp":1733422196000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.nature.com\/articles\/s41746-024-01346-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,5]]},"references-count":47,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2024,12]]}},"alternative-id":["1346"],"URL":"https:\/\/doi.org\/10.1038\/s41746-024-01346-8","relation":{},"ISSN":["2398-6352"],"issn-type":[{"value":"2398-6352","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,12,5]]},"assertion":[{"value":"30 May 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 November 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"5 December 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"The authors declare no competing interests.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"356"}}