{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,30]],"date-time":"2025-12-30T15:38:41Z","timestamp":1767109121910,"version":"3.40.3"},"publisher-location":"Cham","reference-count":27,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030992026"},{"type":"electronic","value":"9783030992033"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-030-99203-3_9","type":"book-chapter","created":{"date-parts":[[2022,3,23]],"date-time":"2022-03-23T16:06:49Z","timestamp":1648051609000},"page":"129-143","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["ResNet-Like CNN Architecture and Saliency Map for Human Activity Recognition"],"prefix":"10.1007","author":[{"given":"Zixuan","family":"Yan","sequence":"first","affiliation":[]},{"given":"Rabih","family":"Younes","sequence":"additional","affiliation":[]},{"given":"Jason","family":"Forsyth","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,3,24]]},"reference":[{"key":"9_CR1","unstructured":"Bordes, A., Glorot, X., Weston, J., Bengio, Y.: Joint learning of words and meaning representations for open-text semantic parsing. In: Artificial Intelligence and Statistics, pp. 127\u2013135. PMLR (2012)"},{"issue":"3","key":"9_CR2","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2499621","volume":"46","author":"A Bulling","year":"2014","unstructured":"Bulling, A., Blanke, U., Schiele, B.: A tutorial on human activity recognition using body-worn inertial sensors. ACM Comput. Surv. (CSUR) 46(3), 1\u201333 (2014)","journal-title":"ACM Comput. Surv. (CSUR)"},{"issue":"15","key":"9_CR3","doi-asserted-by":"publisher","first-page":"2033","DOI":"10.1016\/j.patrec.2012.12.014","volume":"34","author":"R Chavarriaga","year":"2013","unstructured":"Chavarriaga, R., et al.: The opportunity challenge: a benchmark database for on-body sensor-based activity recognition. Pattern Recogn. Lett. 34(15), 2033\u20132042 (2013)","journal-title":"Pattern Recogn. Lett."},{"key":"9_CR4","doi-asserted-by":"crossref","unstructured":"Chen, Y., Xue, Y.: A deep learning approach to human activity recognition based on single accelerometer. In: 2015 IEEE International Conference on Systems, Man, and Cybernetics, pp. 1488\u20131492. IEEE (2015)","DOI":"10.1109\/SMC.2015.263"},{"key":"9_CR5","doi-asserted-by":"publisher","unstructured":"Cleland, I., et al.: Optimal placement of accelerometers for the detection of everyday activities. Sensors (Basel) 13(7), 9183\u2013200 (2013). https:\/\/doi.org\/10.3390\/s130709183, https:\/\/www.ncbi.nlm.nih.gov\/pubmed\/23867744","DOI":"10.3390\/s130709183"},{"key":"9_CR6","unstructured":"Hammerla, N.Y., Halloran, S., Pl\u00f6tz, T.: Deep, convolutional, and recurrent models for human activity recognition using wearables. arXiv preprint arXiv:1604.08880 (2016)"},{"key":"9_CR7","doi-asserted-by":"crossref","unstructured":"Harrison, C., Tan, D., Morris, D.: Skinput: appropriating the body as an input surface. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 453\u2013462 (2010)","DOI":"10.1145\/1753326.1753394"},{"key":"9_CR8","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"issue":"6","key":"9_CR9","doi-asserted-by":"publisher","first-page":"82","DOI":"10.1109\/MSP.2012.2205597","volume":"29","author":"G Hinton","year":"2012","unstructured":"Hinton, G., et al.: Deep neural networks for acoustic modeling in speech recognition: the shared views of four research groups. IEEE Signal Process. Mag. 29(6), 82\u201397 (2012)","journal-title":"IEEE Signal Process. Mag."},{"key":"9_CR10","unstructured":"Howard, A.G., et al.: MobileNets: efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861 (2017)"},{"key":"9_CR11","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., Van Der Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4700\u20134708 (2017)","DOI":"10.1109\/CVPR.2017.243"},{"key":"9_CR12","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. In: Advances in Neural Information Processing Systems, pp. 1097\u20131105"},{"issue":"7553","key":"9_CR13","doi-asserted-by":"publisher","first-page":"436","DOI":"10.1038\/nature14539","volume":"521","author":"Y LeCun","year":"2015","unstructured":"LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521(7553), 436\u2013444 (2015)","journal-title":"Nature"},{"key":"9_CR14","unstructured":"Lin, M., Chen, Q., Yan, S.: Network in network. arXiv preprint arXiv:1312.4400 (2013)"},{"issue":"1","key":"9_CR15","doi-asserted-by":"publisher","first-page":"115","DOI":"10.3390\/s16010115","volume":"16","author":"FJ Ord\u00f3\u00f1ez","year":"2016","unstructured":"Ord\u00f3\u00f1ez, F.J., Roggen, D.: Deep convolutional and LSTM recurrent neural networks for multimodal wearable activity recognition. Sensors 16(1), 115 (2016)","journal-title":"Sensors"},{"key":"9_CR16","doi-asserted-by":"publisher","unstructured":"Pannurat, N., Thiemjarus, S., Nantajeewarawat, E., Anantavrasilp, I.: Analysis of optimal sensor positions for activity classification and application on a different data collection scenario. Sensors (Basel) 17(4) (2017). https:\/\/doi.org\/10.3390\/s17040774, https:\/\/www.ncbi.nlm.nih.gov\/pubmed\/28379208","DOI":"10.3390\/s17040774"},{"issue":"12","key":"9_CR17","doi-asserted-by":"publisher","first-page":"2095","DOI":"10.1109\/TSMC.2017.2705582","volume":"48","author":"B Pourbabaee","year":"2018","unstructured":"Pourbabaee, B., Roshtkhari, M.J., Khorasani, K.: Deep convolutional neural networks and learning ECG features for screening paroxysmal atrial fibrillation patients. IEEE Trans. Syst. Man Cybern. Syst. 48(12), 2095\u20132104 (2018)","journal-title":"IEEE Trans. Syst. Man Cybern. Syst."},{"issue":"2","key":"9_CR18","doi-asserted-by":"publisher","first-page":"756","DOI":"10.1109\/TIP.2015.2508600","volume":"25","author":"J Qin","year":"2015","unstructured":"Qin, J., Liu, L., Zhang, Z., Wang, Y., Shao, L.: Compressive sequential learning for action similarity labeling. IEEE Trans. Image Process. 25(2), 756\u2013769 (2015)","journal-title":"IEEE Trans. Image Process."},{"key":"9_CR19","unstructured":"Simonyan, K., Vedaldi, A., Zisserman, A.: Deep inside convolutional networks: visualising image classification models and saliency maps. arXiv preprint arXiv:1312.6034 (2013)"},{"key":"9_CR20","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. In: ICLR (2014)"},{"key":"9_CR21","doi-asserted-by":"crossref","unstructured":"Szegedy, C., et al.: Going deeper with convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1\u20139 (2015)","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"9_CR22","doi-asserted-by":"crossref","unstructured":"Tang, Y., Teng, Q., Zhang, L., Min, F., He, J.: Efficient convolutional neural networks with smaller filters for human activity recognition using wearable sensors. arXiv preprint arXiv:2005.03948 (2020)","DOI":"10.1109\/JSEN.2020.3015521"},{"key":"9_CR23","doi-asserted-by":"crossref","unstructured":"Vepakomma, P., De, D., Das, S.K., Bhansali, S.: A-wristocracy: Deep learning on wrist-worn sensing for recognition of user complex activities. In: 2015 IEEE 12th International Conference on Wearable and Implantable Body Sensor Networks (BSN), pp. 1\u20136. IEEE (2015)","DOI":"10.1109\/BSN.2015.7299406"},{"key":"9_CR24","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1016\/j.patrec.2018.02.010","volume":"119","author":"J Wang","year":"2019","unstructured":"Wang, J., Chen, Y., Hao, S., Peng, X., Hu, L.: Deep learning for sensor-based activity recognition: a survey. Pattern Recogn. Lett. 119, 3\u201311 (2019)","journal-title":"Pattern Recogn. Lett."},{"key":"9_CR25","doi-asserted-by":"publisher","first-page":"56855","DOI":"10.1109\/ACCESS.2020.2982225","volume":"8","author":"K Xia","year":"2020","unstructured":"Xia, K., Huang, J., Wang, H.: LSTM-CNN architecture for human activity recognition. IEEE Access 8, 56855\u201356866 (2020)","journal-title":"IEEE Access"},{"key":"9_CR26","unstructured":"Yang, J., Nguyen, M.N., San, P.P., Li, X.L., Krishnaswamy, S.: Deep convolutional neural networks on multichannel time series for human activity recognition. In: Twenty-Fourth International Joint Conference on Artificial Intelligence (2015)"},{"key":"9_CR27","doi-asserted-by":"crossref","unstructured":"Zeng, M., et al.: Convolutional neural networks for human activity recognition using mobile sensors. In: 6th International Conference on Mobile Computing, Applications and Services, pp. 197\u2013205. IEEE (2014)","DOI":"10.4108\/icst.mobicase.2014.257786"}],"container-title":["Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering","Mobile Computing, Applications, and Services"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-99203-3_9","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,3,23]],"date-time":"2022-03-23T16:08:29Z","timestamp":1648051709000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-99203-3_9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783030992026","9783030992033"],"references-count":27,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-99203-3_9","relation":{},"ISSN":["1867-8211","1867-822X"],"issn-type":[{"type":"print","value":"1867-8211"},{"type":"electronic","value":"1867-822X"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"24 March 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MobiCASE","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Mobile Computing, Applications, and Services","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 November 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"24 November 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"12","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"mobicase2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Confy+","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"21","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"9","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"43% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}