{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,8]],"date-time":"2026-03-08T01:38:26Z","timestamp":1772933906917,"version":"3.50.1"},"reference-count":30,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,12,8]],"date-time":"2025-12-08T00:00:00Z","timestamp":1765152000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,12,8]],"date-time":"2025-12-08T00:00:00Z","timestamp":1765152000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,12,8]]},"DOI":"10.1109\/bigdata66926.2025.11401853","type":"proceedings-article","created":{"date-parts":[[2026,3,6]],"date-time":"2026-03-06T20:57:57Z","timestamp":1772830677000},"page":"5255-5262","source":"Crossref","is-referenced-by-count":0,"title":["AVATAAR: Agentic Video Answering via Temporal Adaptive Alignment and Reasoning"],"prefix":"10.1109","author":[{"given":"Fang-Chun","family":"Yeh","sequence":"first","affiliation":[{"name":"S&#x0026;P Global Inc.,Commercial Data Science,New York,United States Of America"}]},{"given":"Urjitkumar","family":"Patel","sequence":"additional","affiliation":[{"name":"S&#x0026;P Global Inc.,Commercial Data Science,New York,United States Of America"}]},{"given":"Chinmay","family":"Gondhalekar","sequence":"additional","affiliation":[{"name":"S&#x0026;P Global Inc.,Commercial Data Science,New York,United States Of America"}]}],"member":"263","reference":[{"key":"ref1","volume-title":"A comprehensive study of deep video action recognition","author":"Zhu","year":"2020"},{"key":"ref2","volume-title":"Videomultiagents: A multi-agent framework for video question answering","author":"Kugo","year":"2025"},{"key":"ref3","volume-title":"Gemini 2.5: Pushing the frontier with advanced reasoning, multimodality, long context, and next generation agentic capabilities","author":"Comanici","year":"2025"},{"key":"ref4","volume-title":"On the opportunities and risks of foundation models","author":"Bommasani","year":"2022"},{"key":"ref5","article-title":"Cinepile: A long video question answering dataset and benchmark","author":"Rawal","year":"2024","journal-title":"arXiv preprint arXiv"},{"key":"ref6","article-title":"Learning transferable visual models from natural language supervision","author":"Radford","year":"2021","journal-title":"ICML"},{"key":"ref7","first-page":"4904","article-title":"Scaling up visual and vision-language representation learning with noisy text supervision","volume-title":"International conference on machine learning","author":"Jia","year":"2021"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i9.26263"},{"key":"ref9","article-title":"Roberta: A robustly optimized bert pretraining approach","author":"Liu","year":"2019","journal-title":"arXiv preprint arXiv"},{"key":"ref10","volume-title":"Flamingo: a visual language model for few-shot learning","author":"Alayrac","year":"2022"},{"key":"ref11","volume-title":"Blip-2: Bootstrapping languageimage pre-training with frozen image encoders and large language models","author":"Li","year":"2023"},{"key":"ref12","article-title":"Internvideo: General video foundation models via generative and discriminative learning","author":"Wang","year":"2024","journal-title":"CVPR"},{"key":"ref13","article-title":"Video-llama: An instruction-tuned audio-visual language model for video understanding","author":"Zhao","year":"2023","journal-title":"arXiv preprint arXiv"},{"key":"ref14","article-title":"Video-llava: Learning united visual representation for video understanding with 11 ms","author":"Zhang","year":"2023","journal-title":"arXiv preprint arXiv"},{"key":"ref15","volume-title":"Gpt-4o system card","author":"Hurst","year":"2024"},{"key":"ref16","first-page":"2025","volume-title":"Claude 3 model card","year":"2024"},{"key":"ref17","volume-title":"The llama 3 herd of models","author":"Grattafiori","year":"2024"},{"key":"ref18","volume-title":"Chameleon: Mixed-modal early-fusion foundation models","year":"2025"},{"key":"ref19","article-title":"Retrievalaugmented generation for knowledge-intensive nlp tasks","author":"Lewis","year":"2020","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.550"},{"key":"ref21","article-title":"Reflexion: Language agents with verbal reinforcement learning","author":"Shinn","year":"2023","journal-title":"arXiv preprint arXiv"},{"key":"ref22","article-title":"A survey on agentic retrieval-augmented generation","author":"Singh","year":"2025","journal-title":"arXiv preprint arXiv"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/ICAIC60265.2024.10433839"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/BigData62323.2024.10825891"},{"key":"ref25","article-title":"MultiFinRAG: An optimized multimodal retrieval-augmented generation (RAG) framework for financial question answering","author":"Gondhalekar","year":"2025","journal-title":"arXiv preprint arXiv"},{"key":"ref26","article-title":"Videorag: Retrieval-augmented video question answering at scale","author":"Chen","year":"2024","journal-title":"arXiv preprint arXiv"},{"key":"ref27","first-page":"1110","article-title":"Dig into multi-modal cues for video retrieval with hierarchical alignment","volume-title":"Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence (IJCAI-21)","author":"Wang","year":"2021"},{"key":"ref28","article-title":"Robust speech recognition via large-scale weak supervision","author":"Radford","year":"2022","journal-title":"arXiv preprint arXiv"},{"key":"ref29","volume-title":"Activitynetqa: A dataset for understanding complex web videos via question answering","author":"Yu","year":"2019"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00965"}],"event":{"name":"2025 IEEE International Conference on Big Data (BigData)","location":"Macau, China","start":{"date-parts":[[2025,12,8]]},"end":{"date-parts":[[2025,12,11]]}},"container-title":["2025 IEEE International Conference on Big Data (BigData)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11400704\/11400712\/11401853.pdf?arnumber=11401853","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T06:54:55Z","timestamp":1772866495000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11401853\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,8]]},"references-count":30,"URL":"https:\/\/doi.org\/10.1109\/bigdata66926.2025.11401853","relation":{},"subject":[],"published":{"date-parts":[[2025,12,8]]}}}