{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,12]],"date-time":"2025-05-12T10:45:14Z","timestamp":1747046714109},"reference-count":34,"publisher":"IEEE","license":[{"start":{"date-parts":[[2021,6,6]],"date-time":"2021-06-06T00:00:00Z","timestamp":1622937600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2021,6,6]],"date-time":"2021-06-06T00:00:00Z","timestamp":1622937600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2021,6,6]],"date-time":"2021-06-06T00:00:00Z","timestamp":1622937600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021,6,6]]},"DOI":"10.1109\/icassp39728.2021.9415101","type":"proceedings-article","created":{"date-parts":[[2021,5,13]],"date-time":"2021-05-13T19:53:45Z","timestamp":1620935625000},"source":"Crossref","is-referenced-by-count":10,"title":["Agent-Environment Network for Temporal Action Proposal Generation"],"prefix":"10.1109","author":[{"given":"Viet-Khoa","family":"Vo-Ho","sequence":"first","affiliation":[]},{"given":"Ngan","family":"Le","sequence":"additional","affiliation":[]},{"given":"Kashu","family":"Kamazaki","sequence":"additional","affiliation":[]},{"given":"Akihiro","family":"Sugimoto","sequence":"additional","affiliation":[]},{"given":"Minh-Triet","family":"Tran","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.593"},{"key":"ref32","article-title":"THUMOS challenge: Action recognition with a large number of classes","author":"jiang","year":"2014"},{"key":"ref31","first-page":"961","article-title":"Activitynet: A large-scale video benchmark for human activity understanding","author":"ghanem","year":"2015","journal-title":"CVPR"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2019.2953187"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.675"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.341"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00124"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.211"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.119"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.392"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.317"},{"key":"ref16","first-page":"3","article-title":"Bsn: Boundary sensitive network for temporal action proposal generation","author":"lin","year":"2018","journal-title":"ECCV"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00372"},{"key":"ref18","first-page":"91","article-title":"Faster r-cnn: Towards real-time object detection with region proposal networks","author":"ren","year":"2015","journal-title":"NIPS"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.324"},{"key":"ref28","article-title":"Temporal convolution based action proposal: Submission to activitynet 2017","author":"lin","year":"2017","journal-title":"arXiv preprint arXiv 1707 06892"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33018957"},{"key":"ref27","article-title":"Msr asia msm at activitynet challenge 2017: Trimmed action recognition, temporal action proposals and densecaptioning events in videos","author":"yao","year":"2017","journal-title":"cvprw"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.112"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW.2017.121"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01216-8_5"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2019.2940007"},{"key":"ref8","first-page":"568","article-title":"Two-stream convolutional networks for action recognition in videos","author":"simonyan","year":"2014","journal-title":"NIPS"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2012.59"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.512"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.213"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00630"},{"key":"ref20","article-title":"Yolov3: An incremental improvement","author":"redmon","year":"2018","journal-title":"ArXiv"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.322"},{"key":"ref21","article-title":"The kinetics human action video dataset","author":"kay","year":"2017","journal-title":"arXiv preprint arXiv 1705 06950"},{"key":"ref24","article-title":"Microsoft coco: Common objects in context","author":"lin","year":"2014","journal-title":"ECCV"},{"key":"ref23","first-page":"5998","article-title":"Attention is all you need","author":"vaswani","year":"2017","journal-title":"NIPS"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.610"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00399"}],"event":{"name":"ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","location":"Toronto, ON, Canada","start":{"date-parts":[[2021,6,6]]},"end":{"date-parts":[[2021,6,11]]}},"container-title":["ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9413349\/9413350\/09415101.pdf?arnumber=9415101","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,5,10]],"date-time":"2022-05-10T15:40:58Z","timestamp":1652197258000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9415101\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,6,6]]},"references-count":34,"URL":"https:\/\/doi.org\/10.1109\/icassp39728.2021.9415101","relation":{},"subject":[],"published":{"date-parts":[[2021,6,6]]}}}