{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,27]],"date-time":"2026-01-27T11:39:09Z","timestamp":1769513949202,"version":"3.49.0"},"reference-count":50,"publisher":"IEEE","license":[{"start":{"date-parts":[[2021,1,10]],"date-time":"2021-01-10T00:00:00Z","timestamp":1610236800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2021,1,10]],"date-time":"2021-01-10T00:00:00Z","timestamp":1610236800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2021,1,10]],"date-time":"2021-01-10T00:00:00Z","timestamp":1610236800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100010190","name":"GENCI-IDRIS","doi-asserted-by":"publisher","award":["2019-AD011011269"],"award-info":[{"award-number":["2019-AD011011269"]}],"id":[{"id":"10.13039\/501100010190","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021,1,10]]},"DOI":"10.1109\/icpr48806.2021.9413153","type":"proceedings-article","created":{"date-parts":[[2021,5,6]],"date-time":"2021-05-06T02:15:54Z","timestamp":1620267354000},"page":"631-638","source":"Crossref","is-referenced-by-count":7,"title":["ActionSpotter: Deep Reinforcement Learning Framework for Temporal Action Spotting in Videos"],"prefix":"10.1109","author":[{"given":"Guillaume","family":"Vaudaux-Ruth","sequence":"first","affiliation":[]},{"given":"Adrien","family":"Chan-Hon-Tong","sequence":"additional","affiliation":[]},{"given":"Catherine","family":"Achard","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/ICME.2019.00224"},{"key":"ref38","article-title":"Rethinking the faster r-cnn architecture for temporal action localization","author":"seybold","year":"2018","journal-title":"CVPR 2018"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.392"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.155"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.119"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.342"},{"key":"ref37","article-title":"BSN: boundary sensitive network for temporal action proposal generation","author":"lin","year":"2018","journal-title":"CoRR"},{"key":"ref36","article-title":"Cascaded boundary regression for temporal action detection","author":"gao","year":"2017","journal-title":"BMVC"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.317"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.617"},{"key":"ref28","author":"haarnoja","year":"2018","journal-title":"Soft actor-critic Off-policy maximum entropy deep reinforcement learning with a stochastic actor"},{"key":"ref27","author":"laud","year":"2004","journal-title":"Theory and Application of Reward Shaping in Reinforcement Learning"},{"key":"ref29","author":"sutton","year":"1998","journal-title":"Introduction to Reinforcement Learning"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1007\/s10462-012-9356-9"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00633"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2018.00223"},{"key":"ref22","article-title":"Recycle-gan: Unsupervised video retargeting","author":"bansal","year":"2018","journal-title":"Proceedings of the European Conference on Computer Vision (ECCV)"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/CRV50864.2020.00038"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-009-0275-4"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2007.4409010"},{"key":"ref26","author":"chung","year":"2014","journal-title":"Empirical evaluation of gated recurrent neural networks on sequence modeling"},{"key":"ref25","article-title":"Activi-tynet: A large-scale video benchmark for human activity understanding","author":"fabian caba heilbron","year":"2015","journal-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1109\/ICME.2019.00224"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/MMSP.2016.7813397"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1016\/j.ins.2017.12.020"},{"key":"ref40","author":"jiang","year":"2014","journal-title":"THUMOS Challenge Action Recognition with A Large Number of Classes"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.367"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2014.04.018"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01240-3_16"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00044"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00137"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00632"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.214"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00564"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2014.05.010"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2017.05.008"},{"key":"ref6","article-title":"A novel video key-frame-extraction algorithm based on perceived motion energy model","author":"liu","year":"2003","journal-title":"IEEE Transactions on Circuits and Systems for Video Technology"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.293"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1016\/j.image.2012.10.002"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1016\/j.jvcir.2013.08.003"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00399"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46478-7_47"},{"key":"ref46","author":"narayan","year":"2019","journal-title":"3c-net Category count and center loss for weakly-supervised action localization"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01225-0_35"},{"key":"ref48","article-title":"Bsn: Boundary sensitive network for temporal action proposal generation","author":"lin","year":"2018","journal-title":"Proceedings of the European Conference on Computer Vision (ECCV)"},{"key":"ref47","author":"he","year":"2020","journal-title":"Bi-direction feature pyramid temporal action detection network"},{"key":"ref42","article-title":"Two-stream convolutional networks for action recognition in videos","author":"simonyan","year":"2014","journal-title":"Advances in Neural Information Processing Systems 27"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298698"},{"key":"ref44","article-title":"W-TALC: weakly-supervised temporal activity localization and classification","author":"paul","year":"2018","journal-title":"CoRR"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.502"}],"event":{"name":"2020 25th International Conference on Pattern Recognition (ICPR)","location":"Milan, Italy","start":{"date-parts":[[2021,1,10]]},"end":{"date-parts":[[2021,1,15]]}},"container-title":["2020 25th International Conference on Pattern Recognition (ICPR)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9411940\/9411911\/09413153.pdf?arnumber=9413153","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,5,10]],"date-time":"2022-05-10T15:40:53Z","timestamp":1652197253000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9413153\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,1,10]]},"references-count":50,"URL":"https:\/\/doi.org\/10.1109\/icpr48806.2021.9413153","relation":{},"subject":[],"published":{"date-parts":[[2021,1,10]]}}}