{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,20]],"date-time":"2026-01-20T01:12:40Z","timestamp":1768871560409,"version":"3.49.0"},"reference-count":62,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2019,1,1]],"date-time":"2019-01-01T00:00:00Z","timestamp":1546300800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2019,1,1]],"date-time":"2019-01-01T00:00:00Z","timestamp":1546300800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2019,1,1]],"date-time":"2019-01-01T00:00:00Z","timestamp":1546300800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61672523"],"award-info":[{"award-number":["61672523"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61902347"],"award-info":[{"award-number":["61902347"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U1609215"],"award-info":[{"award-number":["U1609215"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100004731","name":"Natural Science Foundation of Zhejiang Province","doi-asserted-by":"publisher","award":["LQ19F020002"],"award-info":[{"award-number":["LQ19F020002"]}],"id":[{"id":"10.13039\/501100004731","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Fundamental Research Funds for the Central Universities and the Research Funds of Renmin University of China","award":["18XNLG19"],"award-info":[{"award-number":["18XNLG19"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Knowl. Data Eng."],"published-print":{"date-parts":[[2019]]},"DOI":"10.1109\/tkde.2019.2947442","type":"journal-article","created":{"date-parts":[[2019,10,15]],"date-time":"2019-10-15T22:19:02Z","timestamp":1571177942000},"page":"1-1","source":"Crossref","is-referenced-by-count":6,"title":["Feature Re-Learning with Data Augmentation for Video Relevance Prediction"],"prefix":"10.1109","author":[{"given":"Jianfeng","family":"Dong","sequence":"first","affiliation":[]},{"given":"Xun","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Leimin","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Chaoxi","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Gang","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Xirong","family":"Li","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","first-page":"2121","article-title":"DeViSE: A deep visual-semantic embedding model","author":"frome","year":"2013","journal-title":"Proc Int Conf Neural Inf Process"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2018.2872898"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1145\/3240508.3266441"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.510"},{"key":"ref31","article-title":"YouTube-8M: A large-scale video classification benchmark","author":"abu-el-haija","year":"2016"},{"key":"ref30","first-page":"1","article-title":"Towards minimal necessary data: The case for analyzing training data requirements of recommender algorithms","author":"larson","year":"2017","journal-title":"Proc FATREC Workshop Responsible Recommendation"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1145\/3240508.3266434"},{"key":"ref36","first-page":"1097","article-title":"ImageNet classification with deep convolutional neural networks","author":"krizhevsky","year":"2012","journal-title":"Proc Int Conf Neural Inf Process"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1145\/3219819.3219856"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1145\/3220162.3220168"},{"key":"ref60","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-015-0816-y"},{"key":"ref62","doi-asserted-by":"publisher","DOI":"10.1109\/ISM.2018.000-3"},{"key":"ref61","article-title":"Adam: A method for stochastic optimization","author":"kingma","year":"2014","journal-title":"Proc Int Conf Learn Representations"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW.2017.121"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW.2017.49"},{"key":"ref29","first-page":"1109","article-title":"Large scale online learning of image similarity through ranking","volume":"11","author":"chechik","year":"2010","journal-title":"J Mach Learn Res"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2018.2831682"},{"key":"ref1","article-title":"Content-based video relevance prediction challenge: Data, protocol, and baseline","author":"liu","year":"2018"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1145\/3231742"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2016.2559947"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2017.2716819"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2018.2789887"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-540-74553-2_37"},{"key":"ref26","article-title":"FusedLSTM: Fusing frame-level and video-level features for content-based video relevance prediction","author":"bhalgat","year":"2018"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2013.2271746"},{"key":"ref50","first-page":"97","article-title":"Unsupervised representation learning with deep convolutional neural network for remote sensing images","author":"radford","year":"2017","journal-title":"Proc Int Conf Image Graphics"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.223"},{"key":"ref59","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298682"},{"key":"ref58","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.572"},{"key":"ref57","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00957"},{"key":"ref56","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2018.2832602"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.497"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1145\/2964284.2984064"},{"key":"ref53","first-page":"3483","article-title":"Learning structured output representation using deep conditional generative models","author":"sohn","year":"2015","journal-title":"Proc Int Conf Neural Inf Process"},{"key":"ref52","article-title":"End to end learning for self-driving cars","author":"bojarski","year":"2016"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1145\/1277741.1277899"},{"key":"ref11","article-title":"Personalized video recommendation using rich contents from videos","author":"du","year":"2018","journal-title":"IEEE Trans Knowl Data Eng"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298932"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1145\/1282280.1282290"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1145\/3240323.3240407"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/INFOP.2015.7489372"},{"key":"ref15","doi-asserted-by":"crossref","first-page":"251","DOI":"10.1007\/978-3-319-51811-4_21","article-title":"Near-duplicate video retrieval by aggregating intermediate CNN layers","author":"kordopatis-zilos","year":"2017","journal-title":"Proc Int Conf Multimedia Model"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2012.2187181"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1145\/3231737"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-32251-9_56"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2018.07.031"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2012.2185041"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2016.2601260"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1145\/2501654.2501658"},{"key":"ref5","article-title":"Deep hashing with category mask for fast video retrieval","author":"liu","year":"2017"},{"key":"ref8","first-page":"2579","article-title":"Visualizing data using t-SNE","volume":"9","author":"van de maaten","year":"2008","journal-title":"J Mach Learn Res"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2012.92"},{"key":"ref49","first-page":"2672","article-title":"Generative adversarial nets","author":"goodfellow","year":"2014","journal-title":"Proc Int Conf Neural Inf Process"},{"key":"ref9","first-page":"714","article-title":"Recommendation as classification: Using social and content-based information in recommendation","author":"basu","year":"1998","journal-title":"Proc 15th Nat \/10th Conf Artif Intell \/Innovative Appl Artif Intell"},{"key":"ref46","article-title":"The effectiveness of data augmentation in image classification using deep learning","author":"perez","year":"2017","journal-title":"Convolutional Neural Network for Visual Recognition"},{"key":"ref45","doi-asserted-by":"crossref","DOI":"10.21105\/joss.00432","article-title":"Augmentor: An image augmentation library for machine learning","author":"bloice","year":"2017"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-93040-4_28"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.405"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1145\/3123266.3123326"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2018.2796248"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2006.100"},{"key":"ref43","first-page":"1","article-title":"VSE++: Improving visual-semantic embeddings with hard negatives","author":"faghri","year":"2018","journal-title":"Proc Brit Mach Vis Conf"}],"container-title":["IEEE Transactions on Knowledge and Data Engineering"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/69\/4358933\/08869872.pdf?arnumber=8869872","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,5,10]],"date-time":"2022-05-10T14:50:48Z","timestamp":1652194248000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8869872\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019]]},"references-count":62,"URL":"https:\/\/doi.org\/10.1109\/tkde.2019.2947442","relation":{},"ISSN":["1041-4347","1558-2191","2326-3865"],"issn-type":[{"value":"1041-4347","type":"print"},{"value":"1558-2191","type":"electronic"},{"value":"2326-3865","type":"electronic"}],"subject":[],"published":{"date-parts":[[2019]]}}}