{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:19:58Z","timestamp":1750220398402,"version":"3.41.0"},"reference-count":11,"publisher":"Association for Computing Machinery (ACM)","issue":"2","license":[{"start":{"date-parts":[[2021,9,14]],"date-time":"2021-09-14T00:00:00Z","timestamp":1631577600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["GetMobile: Mobile Comp. and Comm."],"published-print":{"date-parts":[[2021,9,14]]},"abstract":"<jats:p>Today's smartphones and wearable devices come equipped with an array of inertial sensors, along with IMU-based Human Activity Recognition models to monitor everyday activities. However, such models rely on large amounts of annotated training data, which require considerable time and effort for collection. One has to recruit human subjects, define clear protocols for the subjects to follow, and manually annotate the collected data, along with the administrative work that goes into organizing such a recording.<\/jats:p>","DOI":"10.1145\/3486880.3486891","type":"journal-article","created":{"date-parts":[[2021,9,16]],"date-time":"2021-09-16T04:07:16Z","timestamp":1631765236000},"page":"38-42","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Can You See It?"],"prefix":"10.1145","volume":"25","author":[{"given":"Hyeokhyen","family":"Kwon","sequence":"first","affiliation":[{"name":"Georgia Institute of Technology, Atlanta, GA, USA"}]},{"given":"Catherine","family":"Tong","sequence":"additional","affiliation":[{"name":"University of Cambridge, United Kingdom"}]},{"given":"Harish","family":"Haresamudram","sequence":"additional","affiliation":[{"name":"Georgia Institute of Technology, Atlanta, GA, USA"}]},{"given":"Yan","family":"Gao","sequence":"additional","affiliation":[{"name":"University of Cambridge, United Kingdom"}]},{"given":"Gregory D.","family":"Abowd","sequence":"additional","affiliation":[{"name":"Northeastern University, Boston, MA, USA"}]},{"given":"Nicholas D.","family":"Lane","sequence":"additional","affiliation":[{"name":"University of Cambridge, United Kingdom"}]},{"given":"Thomas","family":"Ploetz","sequence":"additional","affiliation":[{"name":"Georgia Institute of Technology, Atlanta, GA, USA"}]}],"member":"320","published-online":{"date-parts":[[2021,9,15]]},"reference":[{"key":"e_1_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1145\/3325424.3329662"},{"key":"e_1_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1145\/3267305.3267509"},{"volume-title":"The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 7291--7299","author":"Cao Z.","key":"e_1_2_1_3_1","unstructured":"Z. Cao , T. Simon , S. Wei , and Y. Sheikh . 2017. Realtime multi-person 2d pose estimation using part affinity fields . In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 7291--7299 . Z. Cao, T. Simon, S. Wei, and Y. Sheikh. 2017. Realtime multi-person 2d pose estimation using part affinity fields. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 7291--7299."},{"volume-title":"The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 7753--7762","author":"Pavllo D.","key":"e_1_2_1_4_1","unstructured":"D. Pavllo , C. Feichtenhofer , D. Grangier , and M. Auli . 2019. 3D human pose estimation in video with temporal convolutions and semisupervised training . In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 7753--7762 . D. Pavllo, C. Feichtenhofer, D. Grangier, and M. Auli. 2019. 3D human pose estimation in video with temporal convolutions and semisupervised training. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 7753--7762."},{"volume-title":"Proceedings of the International Conference on Information Processing in Sensor Networks (IPSN). IEEE, 199--210","author":"Young A.","key":"e_1_2_1_5_1","unstructured":"A. Young , M. Ling , and D. Arvind . 2011. IMUSim: A simulation environment for inertial sensing algorithm design and evaluation . In Proceedings of the International Conference on Information Processing in Sensor Networks (IPSN). IEEE, 199--210 . A. Young, M. Ling, and D. Arvind. 2011. IMUSim: A simulation environment for inertial sensing algorithm design and evaluation. In Proceedings of the International Conference on Information Processing in Sensor Networks (IPSN). IEEE, 199--210."},{"key":"e_1_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1145\/3411841"},{"key":"e_1_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.5555\/2969033.2969125"},{"key":"e_1_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1109\/ISWC.2012.13"},{"volume-title":"IEEE International Conference on Pervasive Computing and Communications (PerCom). IEEE, 1--9.","author":"Sztyler T.","key":"e_1_2_1_9_1","unstructured":"T. Sztyler and H. Stuckenschmidt . 2016. On-body localization of wearable devices: An investigation of position-aware activity recognition . In IEEE International Conference on Pervasive Computing and Communications (PerCom). IEEE, 1--9. T. Sztyler and H. Stuckenschmidt. 2016. On-body localization of wearable devices: An investigation of position-aware activity recognition. In IEEE International Conference on Pervasive Computing and Communications (PerCom). IEEE, 1--9."},{"key":"e_1_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.patrec.2012.12.014"},{"key":"e_1_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/2493988.2494353"}],"container-title":["GetMobile: Mobile Computing and Communications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3486880.3486891","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3486880.3486891","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T20:18:47Z","timestamp":1750191527000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3486880.3486891"}},"subtitle":["Good, So We Can Sense It!"],"short-title":[],"issued":{"date-parts":[[2021,9,14]]},"references-count":11,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2021,9,14]]}},"alternative-id":["10.1145\/3486880.3486891"],"URL":"https:\/\/doi.org\/10.1145\/3486880.3486891","relation":{},"ISSN":["2375-0529","2375-0537"],"issn-type":[{"type":"print","value":"2375-0529"},{"type":"electronic","value":"2375-0537"}],"subject":[],"published":{"date-parts":[[2021,9,14]]},"assertion":[{"value":"2021-09-15","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}