{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,6]],"date-time":"2025-02-06T05:09:38Z","timestamp":1738818578020,"version":"3.37.0"},"reference-count":22,"publisher":"IEEE","license":[{"start":{"date-parts":[[2024,9,18]],"date-time":"2024-09-18T00:00:00Z","timestamp":1726617600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,9,18]],"date-time":"2024-09-18T00:00:00Z","timestamp":1726617600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,9,18]]},"DOI":"10.1109\/cbmi62980.2024.10859220","type":"proceedings-article","created":{"date-parts":[[2025,2,4]],"date-time":"2025-02-04T18:30:52Z","timestamp":1738693852000},"page":"1-6","source":"Crossref","is-referenced-by-count":0,"title":["Finding Video Shots for Immersive Journalism Through Text-to-Video Search"],"prefix":"10.1109","author":[{"given":"Lyndon","family":"Nixon","sequence":"first","affiliation":[{"name":"Modul Technology,Vienna,Austria"}]},{"given":"Damianos","family":"Galanopoulos","sequence":"additional","affiliation":[{"name":"CERTH-ITI,Thessaloniki,Greece"}]},{"given":"Vasileios","family":"Mezaris","sequence":"additional","affiliation":[{"name":"CERTH-ITI,Thessaloniki,Greece"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1162\/PRES_a_00005"},{"key":"ref2","article-title":"What is immersive journalism?","author":"Gynnild","year":"2020","journal-title":"Immersive Journalism as Storytelling, OAPEN open access library"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1080\/1461670X.2022.2161931"},{"key":"ref4","article-title":"Progress and Prospects in 3D Generative AI: A Technical Overview including 3D human","author":"Bai","year":"2024","journal-title":"arXiv preprint"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.02100"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.17645\/mac.v8i3.3031"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-05710-7_12"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1145\/3343031.3351046"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-25069-9_40"},{"key":"ref10","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","volume-title":"International conference on machine learning","author":"Radford","year":"2021"},{"key":"ref11","first-page":"529","article-title":"Slip: Selfsupervision meets language-image pre-training","volume-title":"European conference on computer vision","author":"Mu"},{"key":"ref12","first-page":"12888","article-title":"Blip: Bootstrapping language-image pre-training for unified vision-language understanding and generation","volume-title":"International conference on machine learning","author":"Li"},{"key":"ref13","first-page":"19730","article-title":"Blip-2: Bootstrapping language-image pre-training with frozen image encoders and large language models","volume-title":"International conference on machine learning","author":"Li"},{"key":"ref14","first-page":"36","article-title":"Improving clip training with language rewrites","author":"Fan","year":"2024","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref15","article-title":"TRECVID 2023 - a Series of Evaluation Tracks in Video Understanding","volume-title":"Proceedings of TRECVID 2023. NIST, USA","author":"George","year":"2023"},{"key":"ref16","article-title":"ITI-CERTH participation in AVS Task of TRECVID 2023, Proceedings of TRECVID 2023","volume-title":"NIST, USA","author":"Galanopoulos","year":"2023"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298698"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.502"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00468"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.571"},{"key":"ref21","article-title":"Whu-Nercms At Trecvid2023: AdHoc Video Search (AVS) And Deep Video Understanding (DVU) Tasks","volume-title":"Proceedings of TRECVID 2023. NIST, USA","author":"He","year":"2023"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00817"}],"event":{"name":"2024 International Conference on Content-Based Multimedia Indexing (CBMI)","start":{"date-parts":[[2024,9,18]]},"location":"Reykjavik, Iceland","end":{"date-parts":[[2024,9,20]]}},"container-title":["2024 International Conference on Content-Based Multimedia Indexing (CBMI)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10858870\/10858871\/10859220.pdf?arnumber=10859220","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,2,5]],"date-time":"2025-02-05T06:14:52Z","timestamp":1738736092000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10859220\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,9,18]]},"references-count":22,"URL":"https:\/\/doi.org\/10.1109\/cbmi62980.2024.10859220","relation":{},"subject":[],"published":{"date-parts":[[2024,9,18]]}}}