{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T11:55:17Z","timestamp":1743076517263,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":18,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819717101"},{"type":"electronic","value":"9789819717118"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-97-1711-8_22","type":"book-chapter","created":{"date-parts":[[2024,3,27]],"date-time":"2024-03-27T19:03:26Z","timestamp":1711566206000},"page":"284-299","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Multi-action Prediction Using an Iterative Masking Approach with Class Activation Mapping"],"prefix":"10.1007","author":[{"given":"Chia-Ying","family":"Wu","sequence":"first","affiliation":[]},{"given":"Yu-Wei","family":"Tsay","sequence":"additional","affiliation":[]},{"given":"Arthur Chun-Chieh","family":"Shih","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,3,28]]},"reference":[{"key":"22_CR1","unstructured":"Simonyan, K., Zisserman, A.: Two-stream convolutional networks for action recognition in videos. In: Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N., Weinberger, K.Q. (eds.) NIPS 2014, vol. 27 (2014)"},{"key":"22_CR2","doi-asserted-by":"publisher","unstructured":"Feichtenhofer, C., Pinz, A., Zisserman, A.: Convolutional two-stream network fusion for video action recognition. In: Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1933\u20131941 (2016). https:\/\/doi.org\/10.1109\/CVPR.2016.213","DOI":"10.1109\/CVPR.2016.213"},{"key":"22_CR3","unstructured":"Lin, T.Y. et al.: Microsoft COCO: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) Computer Vision \u2013 ECCV 2014. ECCV 2014. LNCS, vol. 8693, pp. 1\u201313. Springer, Cham (2014)"},{"key":"22_CR4","unstructured":"Bojarski, M., et al.: End to End Learning for Self-Driving Cars. arXiv e-prints (2016)"},{"key":"22_CR5","doi-asserted-by":"publisher","unstructured":"Taigman, Y., et al.: DeepFace: closing the gap to human-level performance in face verification. In: Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1701\u20131708 (2014). https:\/\/doi.org\/10.1109\/CVPR.2014.220","DOI":"10.1109\/CVPR.2014.220"},{"issue":"12","key":"22_CR6","doi-asserted-by":"publisher","first-page":"2481","DOI":"10.1109\/TPAMI.2016.2644615","volume":"39","author":"V Badrinarayanan","year":"2017","unstructured":"Badrinarayanan, V., et al.: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 39(12), 2481\u20132495 (2017). https:\/\/doi.org\/10.1109\/TPAMI.2016.2644615","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"22_CR7","doi-asserted-by":"crossref","unstructured":"Kalchbrenner, N., Grefenstette, E., Blunsom, P.: A convolutional neural network for modelling sentences. In: Proceedings of the Annual Meeting of the Association for Computational Linguistics (2014)","DOI":"10.3115\/v1\/P14-1062"},{"key":"22_CR8","doi-asserted-by":"publisher","unstructured":"Soomro, K., Roshan Zamir, A., Shah, M.: UCF101: A Dataset of 101 Human Actions Classes From Videos in The Wild. arXiv e-prints, arXiv:1212.0402 (2012). https:\/\/doi.org\/10.48550\/arXiv.1212.0402","DOI":"10.48550\/arXiv.1212.0402"},{"key":"22_CR9","doi-asserted-by":"publisher","unstructured":"Wu, C.-Y., et al.: Refined prediction of mouse and human actions based on a data-selective multiple-stage approach of 3D convolutional neural networks. In: Proceedings of the 2020 International Conference on Technologies and Applications of Artificial Intelligence (TAAI), pp. 242\u2013247 (2020). https:\/\/doi.org\/10.1109\/TAAI51410.2020.00052","DOI":"10.1109\/TAAI51410.2020.00052"},{"key":"22_CR10","doi-asserted-by":"publisher","first-page":"618","DOI":"10.1109\/ICCV.2017.74","volume":"2017","author":"RR Selvaraju","year":"2017","unstructured":"Selvaraju, R.R., et al.: Grad-CAM: visual explanations from deep networks via gradient-based localization. ICCV 2017, 618\u2013626 (2017). https:\/\/doi.org\/10.1109\/ICCV.2017.74","journal-title":"ICCV"},{"key":"22_CR11","doi-asserted-by":"publisher","first-page":"839","DOI":"10.1109\/WACV.2018.00097","volume":"2018","author":"A Chattopadhay","year":"2018","unstructured":"Chattopadhay, A., et al.: Grad-CAM++: generalized gradient-based visual explanations for deep convolutional networks. WACV 2018, 839\u2013847 (2018). https:\/\/doi.org\/10.1109\/WACV.2018.00097","journal-title":"WACV"},{"key":"22_CR12","doi-asserted-by":"publisher","unstructured":"Wang, H., et al.: Score-CAM: score-weighted visual explanations for convolutional neural networks. In: CVPRW 2020. IEEE\/CVF (2020). https:\/\/doi.org\/10.1109\/CVPRW50498.2020.00020","DOI":"10.1109\/CVPRW50498.2020.00020"},{"key":"22_CR13","doi-asserted-by":"publisher","unstructured":"Vinogradova, K., Dibrov, A., Myers, G.: Towards Interpretable Semantic Segmentation via Gradient-weighted Class Activation Mapping. arXiv e-prints (2020). https:\/\/doi.org\/10.48550\/arXiv.2002.11434","DOI":"10.48550\/arXiv.2002.11434"},{"key":"22_CR14","doi-asserted-by":"publisher","unstructured":"Wu, C.-Y., Tsay, Y.-W., Shih, A. C.-C.: Open action recognition by a 3d convolutional neural network combining with an open fuzzy min-max neural network. In: Proceedings of the 2022 International Conference on Advanced Robotics and Intelligent Systems (ARIS), pp. 1\u20136 (2022). https:\/\/doi.org\/10.1109\/ARIS56205.2022.9910444","DOI":"10.1109\/ARIS56205.2022.9910444"},{"key":"22_CR15","doi-asserted-by":"publisher","unstructured":"Zhou, B., et al.: Learning deep features for discriminative localization. In: Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2921\u20132929 (2016). https:\/\/doi.org\/10.1109\/CVPR.2016.319","DOI":"10.1109\/CVPR.2016.319"},{"issue":"1","key":"22_CR16","doi-asserted-by":"publisher","first-page":"68","DOI":"10.1038\/ncomms1064","volume":"1","author":"H Jhuang","year":"2010","unstructured":"Jhuang, H., et al.: Automated home-cage behavioural phenotyping of mice. Nat. Commun. 1(1), 68 (2010). https:\/\/doi.org\/10.1038\/ncomms1064","journal-title":"Nat. Commun."},{"issue":"1","key":"22_CR17","doi-asserted-by":"publisher","first-page":"221","DOI":"10.1109\/TPAMI.2012.59","volume":"35","author":"S Ji","year":"2013","unstructured":"Ji, S., Xu, W., Yang, M., Yu, K.: 3D Convolutional neural networks for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 35(1), 221\u2013231 (2013). https:\/\/doi.org\/10.1109\/TPAMI.2012.59","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"22_CR18","doi-asserted-by":"publisher","unstructured":"Abu-El-Haija, S., et al.: YouTube-8M: A Large-Scale Video Classification Benchmark. arXiv e-prints (2016). https:\/\/doi.org\/10.48550\/arXiv.1609.08675","DOI":"10.48550\/arXiv.1609.08675"}],"container-title":["Communications in Computer and Information Science","Technologies and Applications of Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-97-1711-8_22","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,27]],"date-time":"2024-03-27T19:21:23Z","timestamp":1711567283000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-97-1711-8_22"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9789819717101","9789819717118"],"references-count":18,"URL":"https:\/\/doi.org\/10.1007\/978-981-97-1711-8_22","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"type":"print","value":"1865-0929"},{"type":"electronic","value":"1865-0937"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"28 March 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"TAAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Technologies and Applications of Artificial Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Yunlin","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Taiwan","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1 December 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2 December 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"taai2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/taai2023.org.tw","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Our build submission system","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"193","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"35","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"12","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"18% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}