{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,10]],"date-time":"2025-12-10T09:00:29Z","timestamp":1765357229171,"version":"3.40.3"},"publisher-location":"Cham","reference-count":67,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031198298"},{"type":"electronic","value":"9783031198304"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19830-4_38","type":"book-chapter","created":{"date-parts":[[2022,10,21]],"date-time":"2022-10-21T16:21:10Z","timestamp":1666369270000},"page":"670-688","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":16,"title":["ERA: Expert Retrieval and\u00a0Assembly for\u00a0Early Action Prediction"],"prefix":"10.1007","author":[{"given":"Lin Geng","family":"Foo","sequence":"first","affiliation":[]},{"given":"Tianjiao","family":"Li","sequence":"additional","affiliation":[]},{"given":"Hossein","family":"Rahmani","sequence":"additional","affiliation":[]},{"given":"Qiuhong","family":"Ke","sequence":"additional","affiliation":[]},{"given":"Jun","family":"Liu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,22]]},"reference":[{"key":"38_CR1","doi-asserted-by":"crossref","unstructured":"Chaabane, M., Trabelsi, A., Blanchard, N., Beveridge, R.: Looking ahead: Anticipating pedestrians crossing with future frames prediction. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 2297\u20132306 (2020)","DOI":"10.1109\/WACV45572.2020.9093426"},{"issue":"1","key":"38_CR2","doi-asserted-by":"publisher","first-page":"231","DOI":"10.1109\/TCSVT.2020.2975065","volume":"31","author":"L Chen","year":"2020","unstructured":"Chen, L., Lu, J., Song, Z., Zhou, J.: Recurrent semantic preserving generation for action prediction. IEEE Trans. Circuits Syst. Video Technol. 31(1), 231\u2013245 (2020)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"38_CR3","doi-asserted-by":"crossref","unstructured":"Chen, Y., Dai, X., Liu, M., Chen, D., Yuan, L., Liu, Z.: Dynamic convolution: Attention over convolution kernels. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11030\u201311039 (2020)","DOI":"10.1109\/CVPR42600.2020.01104"},{"key":"38_CR4","doi-asserted-by":"crossref","unstructured":"Chen, Y., Zhang, Z., Yuan, C., Li, B., Deng, Y., Hu, W.: Channel-wise topology refinement graph convolution for skeleton-based action recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 13359\u201313368 (2021)","DOI":"10.1109\/ICCV48922.2021.01311"},{"key":"38_CR5","doi-asserted-by":"crossref","unstructured":"Cheng, K., Zhang, Y., He, X., Chen, W., Cheng, J., Lu, H.: Skeleton-based action recognition with shift graph convolutional network. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 183\u2013192 (2020)","DOI":"10.1109\/CVPR42600.2020.00026"},{"key":"38_CR6","doi-asserted-by":"crossref","unstructured":"Emad, M., Ishack, M., Ahmed, M., Osama, M., Salah, M., Khoriba, G.: Early-anomaly prediction in surveillance cameras for security applications. In: 2021 International Mobile, Intelligent, and Ubiquitous Computing Conference (MIUCC), pp. 124\u2013128. IEEE (2021)","DOI":"10.1109\/MIUCC52538.2021.9447668"},{"issue":"2","key":"38_CR7","doi-asserted-by":"publisher","first-page":"2682","DOI":"10.3390\/s130202682","volume":"13","author":"I Fatima","year":"2013","unstructured":"Fatima, I., Fahim, M., Lee, Y.K., Lee, S.: A unified framework for activity recognition-based behavior analysis and action prediction in smart homes. Sensors 13(2), 2682\u20132699 (2013)","journal-title":"Sensors"},{"key":"38_CR8","doi-asserted-by":"crossref","unstructured":"Feichtenhofer, C., Fan, H., Malik, J., He, K.: SlowFast networks for video recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 6202\u20136211 (2019)","DOI":"10.1109\/ICCV.2019.00630"},{"key":"38_CR9","unstructured":"Finn, C., Abbeel, P., Levine, S.: Model-agnostic meta-learning for fast adaptation of deep networks. In: International Conference on Machine Learning, pp. 1126\u20131135. PMLR (2017)"},{"key":"38_CR10","doi-asserted-by":"crossref","unstructured":"Gammulle, H., Denman, S., Sridharan, S., Fookes, C.: Predicting the future: a jointly learnt model for action anticipation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5562\u20135571 (2019)","DOI":"10.1109\/ICCV.2019.00566"},{"key":"38_CR11","doi-asserted-by":"crossref","unstructured":"Gujjar, P., Vaughan, R.: Classifying pedestrian actions in advance using predicted video of urban driving scenes. In: 2019 International Conference on Robotics and Automation (ICRA), pp. 2097\u20132103. IEEE (2019)","DOI":"10.1109\/ICRA.2019.8794278"},{"key":"38_CR12","unstructured":"Han, Y., Huang, G., Song, S., Yang, L., Wang, H., Wang, Y.: Dynamic neural networks: A survey. arXiv preprint arXiv:2102.04906 (2021)"},{"key":"38_CR13","doi-asserted-by":"crossref","unstructured":"Hara, K., Kataoka, H., Satoh, Y.: Can spatiotemporal 3d CNNs retrace the history of 2d cnns and imagenet? In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6546\u20136555 (2018)","DOI":"10.1109\/CVPR.2018.00685"},{"key":"38_CR14","doi-asserted-by":"crossref","unstructured":"Hu, J.F., Zheng, W.S., Lai, J., Zhang, J.: Jointly learning heterogeneous features for RGB-d activity recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5344\u20135352 (2015)","DOI":"10.1109\/CVPR.2015.7299172"},{"key":"38_CR15","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"280","DOI":"10.1007\/978-3-319-46448-0_17","volume-title":"Computer Vision \u2013 ECCV 2016","author":"J-F Hu","year":"2016","unstructured":"Hu, J.-F., Zheng, W.-S., Ma, L., Wang, G., Lai, J.: Real-time RGB-D activity prediction by soft regression. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9905, pp. 280\u2013296. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46448-0_17"},{"issue":"11","key":"38_CR16","doi-asserted-by":"publisher","first-page":"2568","DOI":"10.1109\/TPAMI.2018.2863279","volume":"41","author":"JF Hu","year":"2018","unstructured":"Hu, J.F., Zheng, W.S., Ma, L., Wang, G., Lai, J., Zhang, J.: Early action prediction by soft regression. IEEE Trans. Pattern Anal. Mach. Intell. 41(11), 2568\u20132583 (2018)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"38_CR17","doi-asserted-by":"crossref","unstructured":"Huang, C.M., Mutlu, B.: Anticipatory robot control for efficient human-robot collaboration. In: 2016 11th ACM\/IEEE International Conference on Human-Robot Interaction (HRI), pp. 83\u201390. IEEE (2016)","DOI":"10.1109\/HRI.2016.7451737"},{"key":"38_CR18","doi-asserted-by":"crossref","unstructured":"Jain, A., Singh, A., Koppula, H.S., Soh, S., Saxena, A.: Recurrent neural networks for driver activity anticipation via sensory-fusion architecture. In: 2016 IEEE International Conference on Robotics and Automation (ICRA), pp. 3118\u20133125. IEEE (2016)","DOI":"10.1109\/ICRA.2016.7487478"},{"key":"38_CR19","unstructured":"Jang, E., Gu, S., Poole, B.: Categorical reparameterization with Gumbel-Softmax. arXiv preprint arXiv:1611.01144 (2016)"},{"issue":"1","key":"38_CR20","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s40537-019-0192-5","volume":"6","author":"JM Johnson","year":"2019","unstructured":"Johnson, J.M., Khoshgoftaar, T.M.: Survey on deep learning with class imbalance. J. Big Data 6(1), 1\u201354 (2019)","journal-title":"J. Big Data"},{"key":"38_CR21","doi-asserted-by":"crossref","unstructured":"Ke, Q., Bennamoun, M., An, S., Sohel, F., Boussaid, F.: A new representation of skeleton sequences for 3D action recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3288\u20133297 (2017)","DOI":"10.1109\/CVPR.2017.486"},{"key":"38_CR22","doi-asserted-by":"publisher","first-page":"959","DOI":"10.1109\/TIP.2019.2937757","volume":"29","author":"Q Ke","year":"2019","unstructured":"Ke, Q., Bennamoun, M., Rahmani, H., An, S., Sohel, F., Boussaid, F.: Learning latent global network for skeleton-based action prediction. IEEE Trans. Image Process. 29, 959\u2013970 (2019)","journal-title":"IEEE Trans. Image Process."},{"key":"38_CR23","unstructured":"Kong, Y., Fu, Y.: Human action recognition and prediction: a survey. arXiv preprint arXiv:1806.11230 (2018)"},{"key":"38_CR24","doi-asserted-by":"crossref","unstructured":"Kong, Y., Gao, S., Sun, B., Fu, Y.: Action prediction from videos via memorizing hard-to-predict samples. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32 (2018)","DOI":"10.1609\/aaai.v32i1.12324"},{"key":"38_CR25","doi-asserted-by":"crossref","unstructured":"Kong, Y., Kit, D., Fu, Y.: A discriminative model with multiple temporal scales for action prediction. In: European conference on computer vision. pp. 596\u2013611. Springer (2014)","DOI":"10.1007\/978-3-319-10602-1_39"},{"key":"38_CR26","doi-asserted-by":"crossref","unstructured":"Kong, Y., Tao, Z., Fu, Y.: Deep sequential context networks for action prediction. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1473\u20131481 (2017)","DOI":"10.1109\/CVPR.2017.390"},{"issue":"3","key":"38_CR27","doi-asserted-by":"publisher","first-page":"539","DOI":"10.1109\/TPAMI.2018.2882805","volume":"42","author":"Y Kong","year":"2018","unstructured":"Kong, Y., Tao, Z., Fu, Y.: Adversarial action prediction networks. IEEE Trans. Pattern Anal. Mach. Intell. 42(3), 539\u2013553 (2018)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"1","key":"38_CR28","doi-asserted-by":"publisher","first-page":"14","DOI":"10.1109\/TPAMI.2015.2430335","volume":"38","author":"HS Koppula","year":"2015","unstructured":"Koppula, H.S., Saxena, A.: Anticipating human activities using object affordances for reactive robotic response. IEEE Trans. Pattern Anal. Mach. Intell. 38(1), 14\u201329 (2015)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"38_CR29","doi-asserted-by":"crossref","unstructured":"Li, H., Wu, Z., Shrivastava, A., Davis, L.S.: 2D or not 2d? adaptive 3d convolution selection for efficient video recognition. In: CVPR, pp. 6155\u20136164 (2021)","DOI":"10.1109\/CVPR46437.2021.00609"},{"key":"38_CR30","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"420","DOI":"10.1007\/978-3-030-58621-8_25","volume-title":"Computer Vision \u2013 ECCV 2020","author":"T Li","year":"2020","unstructured":"Li, T., Liu, J., Zhang, W., Duan, L.: HARD-Net: hardness-AwaRe discrimination network for 3D early activity prediction. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12356, pp. 420\u2013436. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58621-8_25"},{"key":"38_CR31","doi-asserted-by":"crossref","unstructured":"Lin, J., Gan, C., Han, S.: TSM: temporal shift module for efficient video understanding. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 7083\u20137093 (2019)","DOI":"10.1109\/ICCV.2019.00718"},{"issue":"10","key":"38_CR32","doi-asserted-by":"publisher","first-page":"2684","DOI":"10.1109\/TPAMI.2019.2916873","volume":"42","author":"J Liu","year":"2019","unstructured":"Liu, J., Shahroudy, A., Perez, M., Wang, G., Duan, L.Y., Kot, A.C.: NTU RGB+ D 120: a large-scale benchmark for 3d human activity understanding. IEEE Trans. Pattern Anal. Mach. Intell. 42(10), 2684\u20132701 (2019)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"6","key":"38_CR33","doi-asserted-by":"publisher","first-page":"1453","DOI":"10.1109\/TPAMI.2019.2898954","volume":"42","author":"J Liu","year":"2019","unstructured":"Liu, J., Shahroudy, A., Wang, G., Duan, L.Y., Kot, A.C.: Skeleton-based online action prediction using scale selection network. IEEE Trans. Pattern Anal. Mach. Intell. 42(6), 1453\u20131467 (2019)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"38_CR34","doi-asserted-by":"crossref","unstructured":"Liu, Z., Zhang, H., Chen, Z., Wang, Z., Ouyang, W.: Disentangling and unifying graph convolutions for skeleton-based action recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 143\u2013152 (2020)","DOI":"10.1109\/CVPR42600.2020.00022"},{"key":"38_CR35","doi-asserted-by":"crossref","unstructured":"Ma, S., Sigal, L., Sclaroff, S.: Learning activity progression in LSTMs for activity detection and early detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1942\u20131950 (2016)","DOI":"10.1109\/CVPR.2016.214"},{"key":"38_CR36","unstructured":"Mavrogiannis, A., Chandra, R., Manocha, D.: B-gap: Behavior-guided action prediction for autonomous navigation. arXiv preprint arXiv:2011.03748 (2020)"},{"key":"38_CR37","unstructured":"Mullapudi, R.T., Mark, W.R., Shazeer, N., Fatahalian, K.: Hydranets: Specialized dynamic architectures for efficient inference. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8080\u20138089 (2018)"},{"key":"38_CR38","doi-asserted-by":"crossref","unstructured":"Nguyen, X.S.: GeomNet: a neural network based on Riemannian geometries of SPD matrix space and Cholesky space for 3d skeleton-based interaction recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 13379\u201313389 (2021)","DOI":"10.1109\/ICCV48922.2021.01313"},{"key":"38_CR39","doi-asserted-by":"crossref","unstructured":"Pang, G., Wang, X., Hu, J., Zhang, Q., Zheng, W.S.: DbdNet: learning bi-directional dynamics for early action prediction. In: IJCAI, pp. 897\u2013903 (2019)","DOI":"10.24963\/ijcai.2019\/126"},{"issue":"6","key":"38_CR40","doi-asserted-by":"publisher","first-page":"1281","DOI":"10.1007\/s10514-017-9692-3","volume":"42","author":"B Reily","year":"2018","unstructured":"Reily, B., Han, F., Parker, L.E., Zhang, H.: Skeleton-based bio-inspired human activity prediction for real-time human-robot interaction. Auton. Robot. 42(6), 1281\u20131298 (2018)","journal-title":"Auton. Robot."},{"key":"38_CR41","doi-asserted-by":"crossref","unstructured":"Sadegh Aliakbarian, M., Sadat Saleh, F., Salzmann, M., Fernando, B., Petersson, L., Andersson, L.: Encouraging lSTMs to anticipate actions very early. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 280\u2013289 (2017)","DOI":"10.1109\/ICCV.2017.39"},{"key":"38_CR42","doi-asserted-by":"crossref","unstructured":"Shahroudy, A., Liu, J., Ng, T.T., Wang, G.: NTU RGB+ D: a large scale dataset for 3d human activity analysis. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1010\u20131019 (2016)","DOI":"10.1109\/CVPR.2016.115"},{"key":"38_CR43","unstructured":"Shazeer, N., Mirhoseini, A., Maziarz, K., Davis, A., Le, Q., Hinton, G., Dean, J.: Outrageously large neural networks: the sparsely-gated mixture-of-experts layer (2017)"},{"key":"38_CR44","doi-asserted-by":"crossref","unstructured":"Shi, L., Zhang, Y., Cheng, J., Lu, H.: Skeleton-based action recognition with directed graph neural networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7912\u20137921 (2019)","DOI":"10.1109\/CVPR.2019.00810"},{"key":"38_CR45","doi-asserted-by":"crossref","unstructured":"Shi, L., Zhang, Y., Cheng, J., Lu, H.: Two-stream adaptive graph convolutional networks for skeleton-based action recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12026\u201312035 (2019)","DOI":"10.1109\/CVPR.2019.01230"},{"key":"38_CR46","doi-asserted-by":"crossref","unstructured":"Shi, L., Zhang, Y., Cheng, J., Lu, H.: AdaSGN: adapting joint number and model size for efficient skeleton-based action recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 13413\u201313422 (2021)","DOI":"10.1109\/ICCV48922.2021.01316"},{"key":"38_CR47","unstructured":"Shu, J., et al.: Meta-Weight-Net: Learning an explicit mapping for sample weighting. In: : Proceedings of the 33rd International Conference on Neural Information Processing System (2019)"},{"key":"38_CR48","doi-asserted-by":"crossref","unstructured":"Song, Y.F., Zhang, Z., Shan, C., Wang, L.: Stronger, faster and more explainable: a graph convolutional baseline for skeleton-based action recognition. In: Proceedings of the 28th ACM International Conference on Multimedia, pp. 1625\u20131633 (2020)","DOI":"10.1145\/3394171.3413802"},{"key":"38_CR49","unstructured":"Soomro, K., Zamir, A.R., Shah, M.: Ucf101: a dataset of 101 human actions classes from videos in the wild. arXiv preprint arXiv:1212.0402 (2012)"},{"key":"38_CR50","doi-asserted-by":"crossref","unstructured":"Tran, V., Balasubramanian, N., Hoai, M.: Progressive knowledge distillation for early action recognition. In: 2021 IEEE International Conference on Image Processing (ICIP), pp. 2583\u20132587. IEEE (2021)","DOI":"10.1109\/ICIP42928.2021.9506507"},{"key":"38_CR51","doi-asserted-by":"crossref","unstructured":"Veit, A., Belongie, S.: Convolutional networks with adaptive inference graphs. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 3\u201318 (2018)","DOI":"10.1007\/978-3-030-01246-5_1"},{"key":"38_CR52","doi-asserted-by":"crossref","unstructured":"Wang, W., Chang, F., Liu, C., Li, G., Wang, B.: GA-Net: a guidance aware network for skeleton-based early activity recognition. IEEE Trans. Multim. Early Access (2021)","DOI":"10.1109\/TMM.2021.3137745"},{"key":"38_CR53","doi-asserted-by":"crossref","unstructured":"Wang, X., Yu, F., Dou, Z.Y., Darrell, T., Gonzalez, J.E.: SkipNet: learning dynamic routing in convolutional networks. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 409\u2013424 (2018)","DOI":"10.1007\/978-3-030-01261-8_25"},{"key":"38_CR54","doi-asserted-by":"crossref","unstructured":"Wang, X., Hu, J.F., Lai, J.H., Zhang, J., Zheng, W.S.: Progressive teacher-student learning for early action prediction. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3556\u20133565 (2019)","DOI":"10.1109\/CVPR.2019.00367"},{"issue":"12","key":"38_CR55","doi-asserted-by":"publisher","first-page":"4626","DOI":"10.1109\/TCSVT.2020.2976789","volume":"30","author":"J Weng","year":"2020","unstructured":"Weng, J., Jiang, X., Zheng, W.L., Yuan, J.: Early action recognition with category exclusion using policy-based reinforcement learning. IEEE Trans. Circuits Syst. Video Technol. 30(12), 4626\u20134638 (2020)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"5","key":"38_CR56","doi-asserted-by":"publisher","first-page":"1484","DOI":"10.1007\/s11263-020-01409-9","volume":"129","author":"X Wu","year":"2021","unstructured":"Wu, X., Wang, R., Hou, J., Lin, H., Luo, J.: Spatial-temporal relation reasoning for action prediction in videos. Int. J. Comput. Vision 129(5), 1484\u20131505 (2021)","journal-title":"Int. J. Comput. Vision"},{"key":"38_CR57","doi-asserted-by":"crossref","unstructured":"Wu, X., Zhao, J., Wang, R.: Anticipating future relations via graph growing for action prediction. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, pp. 2952\u20132960 (2021)","DOI":"10.1609\/aaai.v35i4.16402"},{"issue":"11","key":"38_CR58","doi-asserted-by":"publisher","first-page":"2965","DOI":"10.1007\/s11263-021-01508-1","volume":"129","author":"Z Wu","year":"2021","unstructured":"Wu, Z., Li, H., Zheng, Y., Xiong, C., Jiang, Y., Davis, L.S.: A coarse-to-fine framework for resource efficient video recognition. Int. J. Comput. Vision 129(11), 2965\u20132977 (2021)","journal-title":"Int. J. Comput. Vision"},{"key":"38_CR59","doi-asserted-by":"crossref","unstructured":"Wu, Z., et al.: Blockdrop: Dynamic inference paths in residual networks. In: CVPR, pp. 8817\u20138826 (2018)","DOI":"10.1109\/CVPR.2018.00919"},{"key":"38_CR60","doi-asserted-by":"crossref","unstructured":"Xie, S., Sun, C., Huang, J., Tu, Z., Murphy, K.: Rethinking spatiotemporal feature learning: speed-accuracy trade-offs in video classification. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 305\u2013321 (2018)","DOI":"10.1007\/978-3-030-01267-0_19"},{"key":"38_CR61","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"531","DOI":"10.1007\/978-3-030-58452-8_31","volume-title":"Computer Vision \u2013 ECCV 2020","author":"Z Xie","year":"2020","unstructured":"Xie, Z., Zhang, Z., Zhu, X., Huang, G., Lin, S.: Spatially adaptive inference with stochastic feature sampling and interpolation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 531\u2013548. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_31"},{"key":"38_CR62","doi-asserted-by":"crossref","unstructured":"Xu, W., Yu, J., Miao, Z., Wan, L., Ji, Q.: Prediction-CGAN: human action prediction with conditional generative adversarial networks. In: Proceedings of the 27th ACM International Conference on Multimedia, pp. 611\u2013619 (2019)","DOI":"10.1145\/3343031.3351073"},{"key":"38_CR63","doi-asserted-by":"crossref","unstructured":"Yan, R., Tang, J., Shu, X., Li, Z., Tian, Q.: Participation-contributed temporal dynamic model for group activity recognition. In: Proceedings of the 26th ACM International Conference on Multimedia, pp. 1292\u20131300 (2018)","DOI":"10.1145\/3240508.3240572"},{"key":"38_CR64","doi-asserted-by":"crossref","unstructured":"Yan, R., Xie, L., Tang, J., Shu, X., Tian, Q.: HiGCIN: hierarchical graph-based cross inference network for group activity recognition. IEEE Trans. Pattern Aanal. Mach. Intell. Early Access (2020)","DOI":"10.1109\/TPAMI.2020.3034233"},{"key":"38_CR65","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"208","DOI":"10.1007\/978-3-030-58598-3_13","volume-title":"Computer Vision \u2013 ECCV 2020","author":"R Yan","year":"2020","unstructured":"Yan, R., Xie, L., Tang, J., Shu, X., Tian, Q.: Social adaptive module for\u00a0weakly-supervised group activity recognition. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12353, pp. 208\u2013224. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58598-3_13"},{"key":"38_CR66","unstructured":"Yang, B., Bender, G., Le, Q.V., Ngiam, J.: CondConv: conditionally parameterized convolutions for efficient inference. In: Advances in Neural Information Processing Systems 32 (NeurIPS 2019) (2019)"},{"key":"38_CR67","doi-asserted-by":"crossref","unstructured":"Ye, F., Pu, S., Zhong, Q., Li, C., Xie, D., Tang, H.: Dynamic GCN: context-enriched topology learning for skeleton-based action recognition. In: Proceedings of the 28th ACM International Conference on Multimedia, pp. 55\u201363 (2020)","DOI":"10.1145\/3394171.3413941"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19830-4_38","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,10,22]],"date-time":"2022-10-22T00:03:12Z","timestamp":1666396992000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19830-4_38"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031198298","9783031198304"],"references-count":67,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19830-4_38","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"22 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}