{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T22:08:46Z","timestamp":1740175726190,"version":"3.37.3"},"reference-count":31,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2022,5,19]],"date-time":"2022-05-19T00:00:00Z","timestamp":1652918400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2022,5,19]],"date-time":"2022-05-19T00:00:00Z","timestamp":1652918400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"name":"The Key Research and Development Program of Shaanxi Province-Key Industry Innovation Chain (Group) -Industrial Field","award":["No. 2019ZDLGY10-06"],"award-info":[{"award-number":["No. 2019ZDLGY10-06"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Complex Intell. Syst."],"published-print":{"date-parts":[[2022,12]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>With the growing popularity of eSports, video highlight detection, which encapsulates the most informative parts in a few seconds, has become a critical part of live competition. However, learning the spatial\u2013temporal dependency efficiently and discriminatively in video highlight detection for league of legends (LoL) is a critical problem. In this study, to address these existing problems, we propose a novel discriminative and efficient non-local attention network (DENAN) for LoL highlight detection. In particular, both spatial and temporal dependencies are learned using an end-to-end lightweight trainable framework. An auxiliary triplet loss is used in discriminative training to learn robust LoL video feature representations and improve DENAN\u2019s performance. Our experimental results on the NLACS and LMS datasets show the effectiveness of our method in terms of performance and computation cost.<\/jats:p>","DOI":"10.1007\/s40747-022-00762-1","type":"journal-article","created":{"date-parts":[[2022,5,19]],"date-time":"2022-05-19T13:03:41Z","timestamp":1652965421000},"page":"5377-5386","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Discriminative and efficient non-local attention network for league of legends highlight detection"],"prefix":"10.1007","volume":"8","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-8200-7710","authenticated-orcid":false,"given":"Qian","family":"Wan","sequence":"first","affiliation":[]},{"given":"Aruna","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Guoshuai","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Le","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Jiaji","family":"Wu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,5,19]]},"reference":[{"doi-asserted-by":"crossref","unstructured":"Xiong B, Kalantidis Y, Ghadiyaram D, Grauman K (2019) Less is more: learning highlight detection from video duration. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 1258\u20131267","key":"762_CR1","DOI":"10.1109\/CVPR.2019.00135"},{"doi-asserted-by":"crossref","unstructured":"Nepal S, Srinivasan U, Reynolds G (2001) Automatic detection of \u201cgoal\u201d segments in basketball videos. In: Proceedings of the ninth ACM international conference on multimedia, pp 261\u2013269","key":"762_CR2","DOI":"10.1145\/500141.500181"},{"doi-asserted-by":"crossref","unstructured":"Rui Y, Gupta A, Acero A (2000) Automatically extracting highlights for TV baseball programs. In: Proceedings of the eighth ACM international conference on multimedia, pp 105\u2013115","key":"762_CR3","DOI":"10.1145\/354384.354443"},{"issue":"5","key":"762_CR4","doi-asserted-by":"publisher","first-page":"907","DOI":"10.1109\/TMM.2005.854410","volume":"7","author":"Y-F Ma","year":"2005","unstructured":"Ma Y-F, Hua X-S, Lie Lu, Zhang H-J (2005) A generic framework of user attention model and its application in video summarization. IEEE Trans Multimed 7(5):907\u2013919. https:\/\/doi.org\/10.1109\/TMM.2005.854410","journal-title":"IEEE Trans Multimed"},{"unstructured":"Song Y, Vallmiyana J, Stent A et al (2015) Tvsum: summarizing web videos using titles. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5179\u20135187","key":"762_CR5"},{"doi-asserted-by":"publisher","unstructured":"Zhang K, Chao W-L, Sha F, Grauman K (2016) Video summarization with long short-term memory. In: Lecture notes in computer science Eur conference on computer vision. Springer, pp 766\u2013782. https:\/\/doi.org\/10.1007\/978-3-319-46478-7_47","key":"762_CR6","DOI":"10.1007\/978-3-319-46478-7_47"},{"doi-asserted-by":"crossref","unstructured":"Yao T, Mei T, Rui Y (2016) Highlight detection with pairwise deep ranking for first-person video summarization. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 982\u2013990","key":"762_CR7","DOI":"10.1109\/CVPR.2016.112"},{"doi-asserted-by":"crossref","unstructured":"Zhou K, Qiao Y, Xiang T (2018) Deep reinforcement learning for unsupervised video summarization with diversity-representativeness reward. In: Thirty-second AAAI conference on artificial intelligence","key":"762_CR8","DOI":"10.1609\/aaai.v32i1.12255"},{"issue":"8","key":"762_CR9","doi-asserted-by":"publisher","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","volume":"9","author":"S Hochreiter","year":"1997","unstructured":"Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735\u20131780. https:\/\/doi.org\/10.1162\/neco.1997.9.8.1735","journal-title":"Neural Comput"},{"doi-asserted-by":"crossref","unstructured":"Zhao B, Li X, Lu X (2017) Hierarchical recurrent neural network for video summarization. In: Proceedings of the 25th ACM international conference on multimedia, pp 863\u2013871","key":"762_CR10","DOI":"10.1145\/3123266.3123328"},{"issue":"6","key":"762_CR11","doi-asserted-by":"publisher","first-page":"1709","DOI":"10.1109\/TCSVT.2019.2904996","volume":"30","author":"Z Ji","year":"2019","unstructured":"Ji Z, Xiong K, Pang Y, Li X (2019) Video summarization with attention-based encoder\u2013decoder networks. IEEE Trans Circuits Syst Video Technol 30(6):1709\u20131717","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"doi-asserted-by":"crossref","unstructured":"Fajtl J, Sokeh HS, Argyriou V, Monekosso D, Remagnino P (2018) Summarizing videos with attention. In: Asian Conference on Computer Vision. Springer, Cham, pp 39\u201354","key":"762_CR32","DOI":"10.1007\/978-3-030-21074-8_4"},{"doi-asserted-by":"crossref","unstructured":"Wang X, Girshick R, Gupta A et al (2018) Non-local neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7794\u20137803","key":"762_CR12","DOI":"10.1109\/CVPR.2018.00813"},{"doi-asserted-by":"publisher","unstructured":"Ma N, Zhang X, Zheng H-T, Sun J (2018) Shufflenet v2: practical guidelines for efficient CNN architecture design. In: Lecture notes in computer science proceedings of the European conference on computer vision (ECCV), pp 122\u2013138. https:\/\/doi.org\/10.1007\/978-3-030-01264-9_8","key":"762_CR13","DOI":"10.1007\/978-3-030-01264-9_8"},{"key":"762_CR14","first-page":"45","volume":"2008","author":"D Borth","year":"2008","unstructured":"Borth D, Ulges A, Schulze C et al (2008) Keyframe extraction for video tagging and summarization. Informatiktage 2008:45\u201348","journal-title":"Informatiktage"},{"doi-asserted-by":"crossref","unstructured":"Lin Y-L, Morariu VI, Hsu W (2015) Summarizing while recording: context-based highlight detection for egocentric videos. In: Proceedings of the IEEE international conference on computer vision workshops, pp 51\u201359","key":"762_CR15","DOI":"10.1109\/ICCVW.2015.65"},{"doi-asserted-by":"crossref","unstructured":"Zhao Z, Jiang S, Huang Q et al (2006) Highlight summarization in sports video based on replay detection. In: Vol. 2006 IEEE international conference on multimedia and expo. IEEE Publications, pp 1613\u20131616","key":"762_CR16","DOI":"10.1109\/ICME.2006.262855"},{"doi-asserted-by":"crossref","unstructured":"Hsieh L-C, Lee C-W, Chiu T-H et al (2012) Live semantic sport highlight detection based on analyzing tweets of twitter. In: Vol. 2012 IEEE international conference on multimedia and expo. IEEE Publications, pp 949\u2013954","key":"762_CR17","DOI":"10.1109\/ICME.2012.135"},{"unstructured":"Song Y (2016) Real-time video highlights for yahoo esports. arXiv preprint http:\/\/arxiv.org\/abs\/1611.08780","key":"762_CR18"},{"unstructured":"Wang L, Sun Z, Yao W et al (2019) \u2018Unsupervised multistream highlight detection for the game\u2019 honor of kings. arXiv preprint http:\/\/arxiv.org\/abs\/1910.06189","key":"762_CR19"},{"doi-asserted-by":"crossref","unstructured":"Ma Y-F, Lu L, Zhang H-J et al (2002) A user attention model for video summarization. In: Proceedings of the tenth ACM international conference on multimedia, pp 533\u2013542","key":"762_CR20","DOI":"10.1145\/641007.641116"},{"issue":"1","key":"762_CR21","doi-asserted-by":"publisher","first-page":"34","DOI":"10.1016\/j.image.2012.10.002","volume":"28","author":"N Ejaz","year":"2013","unstructured":"Ejaz N, Mehmood I, Wook Baik SW (2013) Efficient visual attention based framework for extracting key frames from videos. Signal Process Image Commun 28(1):34\u201344. https:\/\/doi.org\/10.1016\/j.image.2012.10.002","journal-title":"Signal Process Image Commun"},{"issue":"10","key":"762_CR22","doi-asserted-by":"publisher","first-page":"2693","DOI":"10.1109\/TMM.2018.2815998","volume":"20","author":"Y Jiao","year":"2018","unstructured":"Jiao Y, Li Z, Huang S et al (2018) Three-dimensional attention-based deep ranking model for video highlight detection. IEEE Trans Multimed 20(10):2693\u20132705. https:\/\/doi.org\/10.1109\/TMM.2018.2815998","journal-title":"IEEE Trans Multimed"},{"unstructured":"Vaswani A, Shazeer N, Parmar N et al (2017) Attention is all you need. Adv Neural Inf Process Syst 5998\u20136008","key":"762_CR23"},{"doi-asserted-by":"crossref","unstructured":"Buades A, Coll B, Morel J-M (2005) A non-local algorithm for image denoising. In: Vol. 2 IEEE comput. Soc. conference on computer vision and pattern recognition (CVPR\u201905). IEEE Publications, pp 60\u201365","key":"762_CR24","DOI":"10.1109\/CVPR.2005.38"},{"unstructured":"Hermans A, Beyer L, Leibe B (2017) In defense of the triplet loss for person re-identification. arXiv preprint http:\/\/arxiv.org\/abs\/1703.07737","key":"762_CR25"},{"doi-asserted-by":"crossref","unstructured":"Zhang X, Zhou X, Lin M et al (2018) Shufflenet: an extremely efficient convolutional neural network for mobile devices. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 6848\u20136856","key":"762_CR26","DOI":"10.1109\/CVPR.2018.00716"},{"unstructured":"Cheng-Yang Fu MB, Lee J, Berg AC (2017) Video highlight prediction using audience chat reactions. In: EMNLP","key":"762_CR27"},{"doi-asserted-by":"crossref","unstructured":"Mahasseni B, Lam M, Todorovic S (2017) Unsupervised video summarization with adversarial lstm networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 202\u2013211","key":"762_CR28","DOI":"10.1109\/CVPR.2017.318"},{"doi-asserted-by":"crossref","unstructured":"Qin Z, Li Z, Zhang Z et al (2019) Thundernet: towards real-time generic object detection. arXiv preprint http:\/\/arxiv.org\/abs\/1903.11752","key":"762_CR29","DOI":"10.1109\/ICCV.2019.00682"},{"doi-asserted-by":"crossref","unstructured":"Szegedy C, Liu W, Jia Y et al (2015) Going deeper with convolutions. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1\u20139","key":"762_CR30","DOI":"10.1109\/CVPR.2015.7298594"}],"container-title":["Complex &amp; Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-022-00762-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s40747-022-00762-1\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-022-00762-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,10,27]],"date-time":"2022-10-27T12:22:18Z","timestamp":1666873338000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s40747-022-00762-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,5,19]]},"references-count":31,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2022,12]]}},"alternative-id":["762"],"URL":"https:\/\/doi.org\/10.1007\/s40747-022-00762-1","relation":{},"ISSN":["2199-4536","2198-6053"],"issn-type":[{"type":"print","value":"2199-4536"},{"type":"electronic","value":"2198-6053"}],"subject":[],"published":{"date-parts":[[2022,5,19]]},"assertion":[{"value":"9 December 2021","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 April 2022","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 May 2022","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"On behalf of all authors, the corresponding author states that there is no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}