{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,9,20]],"date-time":"2024-09-20T16:40:02Z","timestamp":1726850402580},"reference-count":26,"publisher":"IEEE","license":[{"start":{"date-parts":[[2020,10,1]],"date-time":"2020-10-01T00:00:00Z","timestamp":1601510400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2020,10,1]],"date-time":"2020-10-01T00:00:00Z","timestamp":1601510400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2020,10,1]],"date-time":"2020-10-01T00:00:00Z","timestamp":1601510400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020,10]]},"DOI":"10.1109\/icip40778.2020.9191256","type":"proceedings-article","created":{"date-parts":[[2020,9,30]],"date-time":"2020-09-30T20:45:18Z","timestamp":1601498718000},"page":"266-270","source":"Crossref","is-referenced-by-count":3,"title":["Egok360: A 360 Egocentric Kinetic Human Activity Video Dataset"],"prefix":"10.1109","author":[{"given":"Keshav","family":"Bhandari","sequence":"first","affiliation":[]},{"given":"Mario A.","family":"DeLaGarza","sequence":"additional","affiliation":[]},{"given":"Ziliang","family":"Zong","sequence":"additional","affiliation":[]},{"given":"Hugo","family":"Latapie","sequence":"additional","affiliation":[]},{"given":"Yan","family":"Yan","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref10","first-page":"357","article-title":"Robot-centric activity recognition from first-person rgb-d videos","author":"lu","year":"2015","journal-title":"WACV"},{"key":"ref11","first-page":"1346","article-title":"Discovering important people and objects for egocentric video summarization","author":"lee","year":"2012","journal-title":"CVPR"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1145\/3083187.3083218"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1145\/3083187.3083219"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.153"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1145\/3083187.3083210"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00154"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.226"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01228-1_38"},{"key":"ref19","article-title":"Charades-ego: A large-scale dataset of paired third and first person videos","author":"sigurdsson","year":"2018","journal-title":"arXiv preprint arXiv 1804 09023"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2011.6126543"},{"key":"ref3","first-page":"arxiv:1212.0402","article-title":"UCF101: A Dataset of 101 Human Actions Classes From Videos in The Wild","author":"soomro","year":"2012","journal-title":"ArXiv e-prints"},{"key":"ref6","first-page":"arxiv:1406.2199","article-title":"Two-Stream Convolutional Networks for Action Recognition in Videos","author":"simonyan","year":"2014","journal-title":"ArXiv e-prints"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.502"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/NCVPRIPG.2015.7489943"},{"key":"ref7","first-page":"arxiv:1705.06950","article-title":"The Kinetics Human Action Video Dataset","author":"kay","year":"2017","journal-title":"ArXiv e-prints"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2017.07.003"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.287"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01234-2_27"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2012.6248010"},{"key":"ref22","first-page":"arxiv:1512.03385","article-title":"Deep Residual Learning for Image Recognition","author":"he","year":"2015","journal-title":"ArXiv e-prints"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.179"},{"key":"ref24","first-page":"arxiv:1512.04150","article-title":"Learning Deep Features for Discriminative Localization","author":"zhou","year":"2015","journal-title":"ArXiv e-prints"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1016\/j.cviu.2017.07.001"},{"key":"ref26","first-page":"arxiv:1708.00919","article-title":"Learning Spherical Convolution for Fast Features from 360 &#x00B0; Imagery","author":"su","year":"2017","journal-title":"ArXiv eprint"},{"key":"ref25","first-page":"arxiv:1812.03115","article-title":"Kernel Transformer Networks for Compact Spherical Convolution","author":"su","year":"2018","journal-title":"ArXiv e-prints"}],"event":{"name":"2020 IEEE International Conference on Image Processing (ICIP)","start":{"date-parts":[[2020,10,25]]},"location":"Abu Dhabi, United Arab Emirates","end":{"date-parts":[[2020,10,28]]}},"container-title":["2020 IEEE International Conference on Image Processing (ICIP)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9184803\/9190635\/09191256.pdf?arnumber=9191256","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,6,28]],"date-time":"2022-06-28T00:12:45Z","timestamp":1656375165000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9191256\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,10]]},"references-count":26,"URL":"https:\/\/doi.org\/10.1109\/icip40778.2020.9191256","relation":{},"subject":[],"published":{"date-parts":[[2020,10]]}}}