{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,3]],"date-time":"2026-05-03T05:45:11Z","timestamp":1777787111845,"version":"3.51.4"},"reference-count":82,"publisher":"Association for Computing Machinery (ACM)","issue":"1","license":[{"start":{"date-parts":[[2023,3,27]],"date-time":"2023-03-27T00:00:00Z","timestamp":1679875200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/501100013314","name":"111 Project","doi-asserted-by":"crossref","award":["No. B18008"],"award-info":[{"award-number":["No. B18008"]}],"id":[{"id":"10.13039\/501100013314","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["No. 62222202, No. 62232004, No. 61972044"],"award-info":[{"award-number":["No. 62222202, No. 62232004, No. 61972044"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["Proc. ACM Interact. Mob. Wearable Ubiquitous Technol."],"published-print":{"date-parts":[[2023,3,27]]},"abstract":"<jats:p>Millimeter wave radar is a promising sensing modality for enabling pervasive and privacy-preserving human sensing. However, the lack of large-scale radar datasets limits the potential of training deep learning models to achieve generalization and robustness. To close this gap, we resort to designing a software pipeline that leverages wealthy video repositories to generate synthetic radar data, but it confronts key challenges including i) multipath reflection and attenuation of radar signals among multiple humans, ii) unconvertible generated data leading to poor generality for various applications, and iii) the class-imbalance issue of videos leading to low model stability. To this end, we design Midas to generate realistic, convertible radar data from videos via two components: (i) a data generation network (DG-Net) combines several key modules, depth prediction, human mesh fitting and multi-human reflection model, to simulate the multipath reflection and attenuation of radar signals to output convertible coarse radar data, followed by a Transformer model to generate realistic radar data; (ii) a variant Siamese network (VS-Net) selects key video clips to eliminate data redundancy for addressing the class-imbalance issue. We implement and evaluate Midas with video data from various external data sources and real-world radar data, demonstrating its great advantages over the state-of-the-art approach for both activity recognition and object detection tasks.<\/jats:p>","DOI":"10.1145\/3580872","type":"journal-article","created":{"date-parts":[[2023,3,28]],"date-time":"2023-03-28T14:57:51Z","timestamp":1680015471000},"page":"1-26","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":32,"title":["Midas"],"prefix":"10.1145","volume":"7","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1123-6978","authenticated-orcid":false,"given":"Kaikai","family":"Deng","sequence":"first","affiliation":[{"name":"State Key Laboratory of Network and Switching Technology, Beijing University of Posts and Telecommunications, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7337-9168","authenticated-orcid":false,"given":"Dong","family":"Zhao","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Network and Switching Technology, Beijing University of Posts and Telecommunications, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6389-6818","authenticated-orcid":false,"given":"Qiaoyue","family":"Han","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Network and Switching Technology, Beijing University of Posts and Telecommunications, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3247-0504","authenticated-orcid":false,"given":"Zihan","family":"Zhang","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Network and Switching Technology, Beijing University of Posts and Telecommunications, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9008-9659","authenticated-orcid":false,"given":"Shuyue","family":"Wang","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Network and Switching Technology, Beijing University of Posts and Telecommunications, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8785-3350","authenticated-orcid":false,"given":"Anfu","family":"Zhou","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Network and Switching Technology, Beijing University of Posts and Telecommunications, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7199-5047","authenticated-orcid":false,"given":"Huadong","family":"Ma","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Network and Switching Technology, Beijing University of Posts and Telecommunications, Beijing, China"}]}],"member":"320","published-online":{"date-parts":[[2023,3,28]]},"reference":[{"key":"e_1_2_1_1_1","volume-title":"Youtube-8m: A large-scale video classification benchmark. arXiv preprint arXiv:1609.08675","author":"Abu-El-Haija Sami","year":"2016","unstructured":"Sami Abu-El-Haija, Nisarg Kothari, Joonseok Lee, Paul Natsev, George Toderici, Balakrishnan Varadarajan, and Sudheendra Vijaya-narasimhan. 2016. Youtube-8m: A large-scale video classification benchmark. arXiv preprint arXiv:1609.08675 (2016)."},{"key":"e_1_2_1_2_1","first-page":"1","article-title":"Ok Google, What Am I Doing? Acoustic Activity Recognition Bounded by Conversational Assistant Interactions","volume":"5","author":"Adaimi Rebecca","year":"2021","unstructured":"Rebecca Adaimi, Howard Yong, and Edison Thomaz. 2021. Ok Google, What Am I Doing? Acoustic Activity Recognition Bounded by Conversational Assistant Interactions. Proc. of ACM IMWUT 5, 1 (2021), 1--24.","journal-title":"Proc. of ACM IMWUT"},{"key":"e_1_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1145\/3411764.3445138"},{"key":"e_1_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1017\/S0305004100034095"},{"key":"e_1_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-48881-3_56"},{"key":"e_1_2_1_6_1","first-page":"1","article-title":"IMU2Doppler: Cross-Modal Domain Adaptation for Doppler-based Activity Recognition Using IMU Data","volume":"5","author":"Bhalla Sejal","year":"2021","unstructured":"Sejal Bhalla, Mayank Goel, and Rushil Khurana. 2021. IMU2Doppler: Cross-Modal Domain Adaptation for Doppler-based Activity Recognition Using IMU Data. Proc. of ACM IMWUT 5, 4 (2021), 1--20.","journal-title":"Proc. of ACM IMWUT"},{"key":"e_1_2_1_7_1","volume-title":"Proc. of IEEE CVPR. 4009--4018","author":"Bhat Shariq Farooq","year":"2021","unstructured":"Shariq Farooq Bhat, Ibraheem Alhashim, and Peter Wonka. 2021. Adabins: Depth estimation using adaptive bins. In Proc. of IEEE CVPR. 4009--4018."},{"key":"e_1_2_1_8_1","volume-title":"A systematic study of the class imbalance problem in convolutional neural networks. Neural networks 106","author":"Buda Mateusz","year":"2018","unstructured":"Mateusz Buda, Atsuto Maki, and Maciej A Mazurowski. 2018. A systematic study of the class imbalance problem in convolutional neural networks. Neural networks 106 (2018), 249--259."},{"key":"e_1_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298698"},{"key":"e_1_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.143"},{"key":"e_1_2_1_11_1","first-page":"3995","article-title":"Exploring rich and efficient spatial temporal interactions for real-time video salient object detection","volume":"30","author":"Chen Chenglizhao","year":"2021","unstructured":"Chenglizhao Chen, Guotao Wang, Chong Peng, Yuming Fang, Dingwen Zhang, and Hong Qin. 2021. Exploring rich and efficient spatial temporal interactions for real-time video salient object detection. IEEE TIP 30 (2021), 3995--4007.","journal-title":"IEEE TIP"},{"key":"e_1_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1109\/HealthCom.2016.7749457"},{"key":"e_1_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/3478119"},{"key":"e_1_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00200"},{"key":"e_1_2_1_15_1","first-page":"539","article-title":"Learning a similarity metric discriminatively, with application to face verification","volume":"1","author":"Chopra Sumit","year":"2005","unstructured":"Sumit Chopra, Raia Hadsell, and Yann LeCun. 2005. Learning a similarity metric discriminatively, with application to face verification. In Proc. of IEEE CVPR, Vol. 1. 539--546.","journal-title":"Proc. of IEEE CVPR"},{"key":"e_1_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP43922.2022.9746764"},{"key":"e_1_2_1_17_1","volume-title":"Karla Conn Welch, and Weihua Sheng","author":"Do Ha Manh","year":"2021","unstructured":"Ha Manh Do, Karla Conn Welch, and Weihua Sheng. 2021. Soham: A sound-based human activity monitoring framework for home service robots. IEEE TASAE (2021)."},{"key":"e_1_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.402"},{"key":"e_1_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/MAES.2015.7119820"},{"key":"e_1_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1109\/RADAR.2019.8835589"},{"key":"e_1_2_1_21_1","first-page":"226","article-title":"A density-based algorithm for discovering clusters in large spatial databases with noise","volume":"96","author":"Ester Martin","year":"1996","unstructured":"Martin Ester, Hans-Peter Kriegel, J\u00f6rg Sander, Xiaowei Xu, et al. 1996. A density-based algorithm for discovering clusters in large spatial databases with noise. In Proc. of ACM SIGKDD, Vol. 96. 226--231.","journal-title":"Proc. of ACM SIGKDD"},{"key":"e_1_2_1_22_1","volume-title":"Jirka Borovec","author":"Glenn Jocher","year":"2022","unstructured":"Jocher Glenn, Stoken Alex, and et al. Jirka Borovec. 2022. ultralytics\/yolov5. Retrieved 2022 from https:\/\/github.com\/ultralytics\/yolov5"},{"key":"e_1_2_1_23_1","unstructured":"Google. 2020. Google Project Soli. Retrieved 2020 from https:\/\/atap.google.com\/soli\/"},{"key":"e_1_2_1_24_1","volume-title":"Out-of-Domain Human Mesh Reconstruction via Dynamic Bilevel Online Adaptation","author":"Guan Shanyan","year":"2021","unstructured":"Shanyan Guan, Jingwei Xu, Michelle Z He, Yunbo Wang, Bingbing Ni, and Xiaokang Yang. 2021. Out-of-Domain Human Mesh Reconstruction via Dynamic Bilevel Online Adaptation. IEEE TPAMI (2021)."},{"key":"e_1_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1145\/3474085.3475285"},{"key":"e_1_2_1_27_1","volume-title":"Multimedia Sensor Networks","author":"Huadong Ma","unstructured":"Ma Huadong, Liu Liang, and Luo Hong. 2021. Multimedia Sensor Networks. Springer."},{"key":"e_1_2_1_28_1","unstructured":"Huawei. 2022. Huawei's whole-house smart strategy upgrade allows users to enjoy a highly intelligent interactive experience. Retrieved 2022 from https:\/\/inf.news\/en\/home\/677b28ebe5f083c7b2bdd779566c5fe0.html"},{"key":"e_1_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58574-7_35"},{"key":"e_1_2_1_30_1","unstructured":"Texas Instruments. 2019. TI IWR1443 single-chip 76-GHz to 81-GHz mmWave sensor evaluation module. Retrieved 2019 from https:\/\/www.ti.com\/tool\/IWR1443BOOST"},{"key":"e_1_2_1_31_1","unstructured":"Texas Instruments. 2020. DCA1000EVM: Real-time Data-Capture Adapter for Radar Sensing Evaluation Module. Retrieved 2020 from https:\/\/www.ti.com\/tool\/DCA1000EVM"},{"key":"e_1_2_1_32_1","first-page":"113","article-title":"Metaphoric hand gestures for orientation-aware VR object manipulation with an egocentric viewpoint","volume":"47","author":"Jang Youngkyoon","year":"2016","unstructured":"Youngkyoon Jang, Ikbeom Jeon, Tae-Kyun Kim, and Woontack Woo. 2016. Metaphoric hand gestures for orientation-aware VR object manipulation with an egocentric viewpoint. IEEE THMS 47, 1 (2016), 113--127.","journal-title":"IEEE THMS"},{"key":"e_1_2_1_33_1","volume-title":"New generation deep learning for video object detection: A survey","author":"Jiao Licheng","year":"2021","unstructured":"Licheng Jiao, Ruohan Zhang, Fang Liu, Shuyuan Yang, Biao Hou, Lingling Li, and Xu Tang. 2021. New generation deep learning for video object detection: A survey. IEEE TNNLS (2021)."},{"key":"e_1_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P16-1089"},{"key":"e_1_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00943"},{"key":"e_1_2_1_36_1","volume-title":"Proc. of ICLR","author":"Kingma Diederik P","year":"2014","unstructured":"Diederik P Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. Proc. of ICLR (2014)."},{"key":"e_1_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00530"},{"key":"e_1_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2011.6126543"},{"key":"e_1_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01282"},{"key":"e_1_2_1_40_1","first-page":"1","article-title":"IMUTube: Automatic extraction of virtual on-body accelerometry from video for human activity recognition","volume":"4","author":"Kwon Hyeokhyen","year":"2020","unstructured":"Hyeokhyen Kwon, Catherine Tong, Harish Haresamudram, Yan Gao, Gregory D Abowd, Nicholas D Lane, and Thomas Ploetz. 2020. IMUTube: Automatic extraction of virtual on-body accelerometry from video for human activity recognition. Proc. of ACM IMWUT 4, 3 (2020), 1--29.","journal-title":"Proc. of ACM IMWUT"},{"key":"e_1_2_1_41_1","first-page":"1","article-title":"Approaching the real-world: Supporting activity recognition training with virtual imu data","volume":"5","author":"Kwon Hyeokhyen","year":"2021","unstructured":"Hyeokhyen Kwon, Bingyao Wang, Gregory D Abowd, and Thomas Pl\u00f6tz. 2021. Approaching the real-world: Supporting activity recognition training with virtual imu data. Proc. of ACM IMWUT 5, 3 (2021), 1--32.","journal-title":"Proc. of ACM IMWUT"},{"key":"e_1_2_1_42_1","first-page":"52","article-title":"Velocity adaptation of space-time interest points","volume":"1","author":"Laptev Ivan","year":"2004","unstructured":"Ivan Laptev and Tony Lindeberg. 2004. Velocity adaptation of space-time interest points. In Proc. of IEEE ICPR, Vol. 1. 52--56.","journal-title":"Proc. of IEEE ICPR"},{"key":"e_1_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1145\/3242587.3242609"},{"key":"e_1_2_1_44_1","volume-title":"Dong Wook Ko, and Il Hong Suh","author":"Lee Jin Han","year":"2019","unstructured":"Jin Han Lee, Myung-Kyu Han, Dong Wook Ko, and Il Hong Suh. 2019. From big to small: Multi-scale local planar guidance for monocular depth estimation. arXiv preprint arXiv:1907.10326 (2019)."},{"key":"e_1_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1049\/joe.2019.0557"},{"key":"e_1_2_1_46_1","first-page":"1","article-title":"Audio-based activities of daily living (adl) recognition with large-scale acoustic embeddings from online videos","volume":"3","author":"Liang Dawei","year":"2019","unstructured":"Dawei Liang and Edison Thomaz. 2019. Audio-based activities of daily living (adl) recognition with large-scale acoustic embeddings from online videos. Proc. of ACM IMWUT 3, 1 (2019), 1--18.","journal-title":"Proc. of ACM IMWUT"},{"key":"e_1_2_1_47_1","volume-title":"VRT: A Video Restoration Transformer. arXiv preprint arXiv:2201.12288","author":"Liang Jingyun","year":"2022","unstructured":"Jingyun Liang, Jiezhang Cao, Yuchen Fan, Kai Zhang, Rakesh Ranjan, Yawei Li, Radu Timofte, and Luc Van Gool. 2022. VRT: A Video Restoration Transformer. arXiv preprint arXiv:2201.12288 (2022)."},{"key":"e_1_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1109\/IROS45743.2020.9340998"},{"key":"e_1_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"e_1_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1109\/CIS.2017.00111"},{"key":"e_1_2_1_51_1","first-page":"1","article-title":"mTransSee: Enabling Environment-Independent mmWave Sensing Based Gesture Recognition via Transfer Learning","volume":"6","author":"Liu Haipeng","year":"2022","unstructured":"Haipeng Liu, Kening Cui, Kaiyuan Hu, Yuheng Wang, Anfu Zhou, Liang Liu, and Huadong Ma. 2022. mTransSee: Enabling Environment-Independent mmWave Sensing Based Gesture Recognition via Transfer Learning. Proc. of ACM IMWUT 6, 1 (2022), 1--28.","journal-title":"Proc. of ACM IMWUT"},{"key":"e_1_2_1_52_1","first-page":"1","article-title":"Real-time arm gesture recognition in smart home scenarios via millimeter wave sensing","volume":"4","author":"Liu Haipeng","year":"2020","unstructured":"Haipeng Liu, Yuheng Wang, Anfu Zhou, Hanyue He, Wei Wang, Kunpeng Wang, Peilin Pan, Yixuan Lu, Liang Liu, and Huadong Ma. 2020. Real-time arm gesture recognition in smart home scenarios via millimeter wave sensing. Proc. of ACM IMWUT 4, 4 (2020), 1--28.","journal-title":"Proc. of ACM IMWUT"},{"key":"e_1_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1109\/COMST.2019.2934489"},{"key":"e_1_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46487-9_50"},{"key":"e_1_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1145\/3300061.3300116"},{"key":"e_1_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDCS47774.2020.00085"},{"key":"e_1_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.257"},{"key":"e_1_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-020-01359-2"},{"key":"e_1_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.1145\/3439726"},{"key":"e_1_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW53098.2021.00372"},{"key":"e_1_2_1_61_1","first-page":"5938","article-title":"DeepReS: A deep learning-based video summarization strategy for resource-constrained industrial surveillance scenarios","volume":"16","author":"Muhammad Khan","year":"2019","unstructured":"Khan Muhammad, Tanveer Hussain, Javier Del Ser, Vasile Palade, and Victor Hugo C De Albuquerque. 2019. DeepReS: A deep learning-based video summarization strategy for resource-constrained industrial surveillance scenarios. IEEE TII 16, 9 (2019), 5938--5947.","journal-title":"IEEE TII"},{"key":"e_1_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP.2019.8803392"},{"key":"e_1_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.1109\/COMST.2015.2476474"},{"key":"e_1_2_1_64_1","doi-asserted-by":"publisher","DOI":"10.1109\/RadarConf2147009.2021.9455194"},{"key":"e_1_2_1_65_1","volume-title":"Proc. of NIPS. 91--99","author":"Ren Shaoqing","year":"2015","unstructured":"Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun. 2015. Faster r-cnn: Towards real-time object detection with region proposal networks. In Proc. of NIPS. 91--99."},{"key":"e_1_2_1_66_1","volume-title":"Fundamentals of radar signal processing","author":"Richards Mark A","unstructured":"Mark A Richards. 2014. Fundamentals of radar signal processing. McGraw-Hill Education."},{"key":"e_1_2_1_67_1","doi-asserted-by":"publisher","DOI":"10.1109\/RADAR.2018.8378629"},{"key":"e_1_2_1_68_1","volume-title":"Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556","author":"Simonyan Karen","year":"2014","unstructured":"Karen Simonyan and Andrew Zisserman. 2014. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)."},{"key":"e_1_2_1_69_1","doi-asserted-by":"publisher","DOI":"10.1145\/3394171.3413611"},{"key":"e_1_2_1_70_1","volume-title":"Amir Roshan Zamir, and Mubarak Shah","author":"Soomro Khurram","year":"2012","unstructured":"Khurram Soomro, Amir Roshan Zamir, and Mubarak Shah. 2012. UCF101: A dataset of 101 human actions classes from videos in the wild. arXiv preprint arXiv:1212.0402 (2012)."},{"key":"e_1_2_1_71_1","unstructured":"Statista. 2020. Hours of video uploaded to YouTube every minute as of February. Retrieved 2020 from https:\/\/www.statista.com\/statistics\/259477\/hours-of-video-uploaded-to-youtube-every-minute\/"},{"key":"e_1_2_1_72_1","volume-title":"Proc. of NIPS 30","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, \u0141ukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. Proc. of NIPS 30 (2017)."},{"key":"e_1_2_1_73_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2019.8682917"},{"key":"e_1_2_1_74_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP39728.2021.9414686"},{"key":"e_1_2_1_75_1","doi-asserted-by":"crossref","unstructured":"Turner Whitted. 2005. An improved illumination model for shaded display. In ACM Siggraph 2005 Courses. 4-es.","DOI":"10.1145\/1198555.1198743"},{"key":"e_1_2_1_76_1","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM41043.2020.9155293"},{"key":"e_1_2_1_77_1","first-page":"4637","article-title":"Simultaneous video stabilization and rolling shutter removal","volume":"30","author":"Wu Huicong","year":"2021","unstructured":"Huicong Wu, Liang Xiao, and Zhihui Wei. 2021. Simultaneous video stabilization and rolling shutter removal. IEEE TIP 30 (2021), 4637--4652.","journal-title":"IEEE TIP"},{"key":"e_1_2_1_78_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376875"},{"key":"e_1_2_1_79_1","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM41043.2020.9155471"},{"key":"e_1_2_1_80_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7299101"},{"key":"e_1_2_1_81_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7299064"},{"key":"e_1_2_1_82_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.316"}],"container-title":["Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3580872","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3580872","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,7,14]],"date-time":"2025-07-14T04:45:46Z","timestamp":1752468346000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3580872"}},"subtitle":["Generating mmWave Radar Data from Videos for Training Pervasive and Privacy-preserving Human Sensing Tasks"],"short-title":[],"issued":{"date-parts":[[2023,3,27]]},"references-count":82,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2023,3,27]]}},"alternative-id":["10.1145\/3580872"],"URL":"https:\/\/doi.org\/10.1145\/3580872","relation":{},"ISSN":["2474-9567"],"issn-type":[{"value":"2474-9567","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,3,27]]},"assertion":[{"value":"2023-03-28","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}