{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T16:21:43Z","timestamp":1743092503008,"version":"3.40.3"},"publisher-location":"Cham","reference-count":23,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783031113451"},{"type":"electronic","value":"9783031113468"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-11346-8_9","type":"book-chapter","created":{"date-parts":[[2022,7,23]],"date-time":"2022-07-23T03:33:48Z","timestamp":1658547228000},"page":"95-106","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Dyadic Interaction Recognition Using Dynamic Representation and Convolutional Neural Network"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-0835-5848","authenticated-orcid":false,"given":"R. Newlin","family":"Shebiah","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2579-501X","authenticated-orcid":false,"given":"S.","family":"Arivazhagan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,7,24]]},"reference":[{"key":"9_CR1","doi-asserted-by":"publisher","first-page":"265","DOI":"10.1007\/s11280-015-0348-y","volume":"19","author":"C Gao","year":"2016","unstructured":"Gao, C., Yang, L., Du, Y., Feng, Z., Liu, J.: From constrained to unconstrained datasets: an evaluation of local action descriptors and fusion strategies for interaction recognition. World Wide Web 19, 265\u2013276 (2016)","journal-title":"World Wide Web"},{"key":"9_CR2","doi-asserted-by":"crossref","unstructured":"Tian, Y., Sukthankar, R., Shah, M.: Spatiotemporal deformable part models for action detection. In: Computer Vision and Pattern Recognition, (CVPR), pp. 2642\u20132649 (2013)","DOI":"10.1109\/CVPR.2013.341"},{"key":"9_CR3","doi-asserted-by":"publisher","unstructured":"Bibi, S., Anjum, N., Sher, M.: Automated multi-feature human interaction recognition in complex environment. Comput. Ind. 99, 282\u2013293 (2018). ISSN 0166-3615, https:\/\/doi.org\/10.1016\/j.compind.2018.03.015","DOI":"10.1016\/j.compind.2018.03.015"},{"issue":"1","key":"9_CR4","doi-asserted-by":"publisher","first-page":"221","DOI":"10.1109\/TPAMI.2012.59","volume":"35","author":"S Ji","year":"2013","unstructured":"Ji, S., Xu, W., Yang, M., Yu, K.: 3d convolutional neural networks for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 35(1), 221\u2013231 (2013)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"9_CR5","unstructured":"Simonyan, A.Z.: Two-stream convolutional networks for action recognition in videos. In: Advances in Neural Information Processing Systems, pp. 568\u2013576 (2014)"},{"key":"9_CR6","doi-asserted-by":"publisher","first-page":"32","DOI":"10.1016\/j.patcog.2018.01.020","volume":"79","author":"Z Tu","year":"2018","unstructured":"Tu, Z., et al.: Multistream CNN: learning representations based on human-related regions for action recognition. Pattern Recogn. 79, 32\u201343 (2018)","journal-title":"Pattern Recogn."},{"key":"9_CR7","doi-asserted-by":"publisher","unstructured":"Ye, Q., Zhong, H., Qu, C., Zhang, Y.: Human interaction recognition based on whole-individual detection. Sensors 20(8), 2346 (2020). https:\/\/doi.org\/10.3390\/s20082346","DOI":"10.3390\/s20082346"},{"key":"9_CR8","doi-asserted-by":"crossref","unstructured":"Ibrahim, M.S., Muralidharan, S., Deng, Z., Vahdat, A., Mori, G.: A hierarchical deep temporal model for group activity recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1971\u20131980 (2016)","DOI":"10.1109\/CVPR.2016.217"},{"key":"9_CR9","unstructured":"Shu, X., Tang, J., Qi, G., Liu, W., Yang, J.: Hierarchical long short-term concurrent memory for human interaction recognition. IEEE Trans. Pattern Anal. Mach. Intell. (2019)"},{"key":"9_CR10","unstructured":"Tang, J., Shu, X., Yan, R., Zhang, L.: Coherence constrained graph lstm for group activity recognition. IEEE Trans. Pattern Anal. Mach. Intell. (2019)"},{"key":"9_CR11","doi-asserted-by":"crossref","unstructured":"Shu, X., Tang, J., Qi, G.-J., Song, Y., Li, Z., Zhang, L.: Concurrence-aware long short-term sub-memories for person-person action recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 1\u20138 (2017)","DOI":"10.1109\/CVPRW.2017.270"},{"key":"9_CR12","doi-asserted-by":"publisher","first-page":"198","DOI":"10.1016\/j.patcog.2018.08.006","volume":"85","author":"D-G Lee","year":"2019","unstructured":"Lee, D.-G., Lee, S.-W.: Prediction of partially observed human activity based on pre-trained deep representation. Pattern Recogn. 85, 198\u2013206 (2019)","journal-title":"Pattern Recogn."},{"key":"9_CR13","doi-asserted-by":"crossref","unstructured":"Mahmood, M., Jalal, A., Sidduqi, M.: Robust spatio-temporal features for human interaction recognition via artificial neural network. In: International Conference on Frontiers of Information Technology, pp. 218\u2013223. IEEE (2018)","DOI":"10.1109\/FIT.2018.00045"},{"key":"9_CR14","doi-asserted-by":"crossref","unstructured":"Deng, Z., Vahdat, A., Hu, H., Mori, G.: Structure inference machines: recurrent neural networks for analyzing relations in group activity recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4772\u20134781 (2016)","DOI":"10.1109\/CVPR.2016.516"},{"key":"9_CR15","unstructured":"Lee, D.-G., & Lee, S.-W.: Human Interaction Recognition Framework based on Interacting Body Part Attention (2021). http:\/\/arxiv.org\/abs\/2101.08967"},{"issue":"4","key":"9_CR16","doi-asserted-by":"publisher","first-page":"773","DOI":"10.1109\/TPAMI.2016.2558148","volume":"39","author":"B Fernando","year":"2017","unstructured":"Fernando, B., Gavves, E., JoseOramas, M., Ghodrati, A., Tuytelaars, T.: Rank pooling for action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 39(4), 773\u2013787 (2017). https:\/\/doi.org\/10.1109\/TPAMI.2016.2558148","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"9_CR17","unstructured":"Yun, K., Honorio, J., Chattopadhyay, D., Berg, T.L., Samaras, D.: The 2nd International Workshop on Human Activity Understanding from 3D Data at Conference on Computer Vision and Pattern Recognition, CVPR 2012 (2012)"},{"key":"9_CR18","doi-asserted-by":"crossref","unstructured":"Weinland, D., Ronfard, R., Boyer, E.: Free viewpoint action recognition using motion history volumes. Computer Vision and Image Understanding (CVIU), vol. 104, no. 2\u20133 (2006)","DOI":"10.1016\/j.cviu.2006.07.013"},{"key":"9_CR19","doi-asserted-by":"publisher","first-page":"2441","DOI":"10.1109\/TPAMI.2012.24","volume":"34","author":"A Patron-Perez","year":"2012","unstructured":"Patron-Perez, A., Marszalek, M., Reid, I., Zisserman, A.: Struc-tured learning of human interactions in TV shows. Trans. Pattern Anal. Mach. Intell. 34, 2441\u20132453 (2012)","journal-title":"Trans. Pattern Anal. Mach. Intell."},{"key":"9_CR20","doi-asserted-by":"crossref","unstructured":"Patron-Perez, A., Marszalek, M., Zisserman, A., Reid, I.D.: Highfive: Recognising human interactions in TV shows, in: British MachineVision Conference (BMVC) (2010)","DOI":"10.5244\/C.24.50"},{"key":"9_CR21","doi-asserted-by":"crossref","unstructured":"Song, S.; Lan, C.; Xing, J.; Zeng,W., Liu, J.: An End-to-End Spatio-Temporal Attention Model for Human Action Recognition from Skeleton Data. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), San Francisco, CA, USA, 4\u20139 February 2017","DOI":"10.1609\/aaai.v31i1.11212"},{"key":"9_CR22","doi-asserted-by":"publisher","first-page":"1586","DOI":"10.1109\/TIP.2017.2785279","volume":"27","author":"J Liu","year":"2018","unstructured":"Liu, J., Wang, G., Duan, L., Abdiyeva, K., Kot, A.C.: Skeleton-based human action recognition with global context-aware attention LSTM networks. IEEE Trans. Image Process. (TIP) 27, 1586\u20131599 (2018)","journal-title":"IEEE Trans. Image Process. (TIP)"},{"key":"9_CR23","doi-asserted-by":"publisher","unstructured":"Pham, H.H., Salmane, H., Khoudour, L., Crouzil, A., Velastin, S.A., Zegers, P.: A unified deep framework for joint 3D pose estimation and action recognition from a single RGB camera. Sensors (Switzerland), 20(7) (2020). https:\/\/doi.org\/10.3390\/s20071825","DOI":"10.3390\/s20071825"}],"container-title":["Communications in Computer and Information Science","Computer Vision and Image Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-11346-8_9","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,9,27]],"date-time":"2022-09-27T20:26:36Z","timestamp":1664310396000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-11346-8_9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031113451","9783031113468"],"references-count":23,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-11346-8_9","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"type":"print","value":"1865-0929"},{"type":"electronic","value":"1865-0937"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"24 July 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"CVIP","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Computer Vision and Image Processing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Rupnagar","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"India","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"3 December 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5 December 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"cvip2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/iitrpr.cvip2021.com\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"260","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"77","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"20","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"30% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}