{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,9]],"date-time":"2026-04-09T06:26:33Z","timestamp":1775715993852,"version":"3.50.1"},"publisher-location":"Cham","reference-count":29,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783030983574","type":"print"},{"value":"9783030983581","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-030-98358-1_18","type":"book-chapter","created":{"date-parts":[[2022,3,14]],"date-time":"2022-03-14T12:06:56Z","timestamp":1647259616000},"page":"217-228","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":7,"title":["LLQA - Lifelog Question Answering Dataset"],"prefix":"10.1007","author":[{"given":"Ly-Duyen","family":"Tran","sequence":"first","affiliation":[]},{"given":"Thanh Cong","family":"Ho","sequence":"additional","affiliation":[]},{"given":"Lan Anh","family":"Pham","sequence":"additional","affiliation":[]},{"given":"Binh","family":"Nguyen","sequence":"additional","affiliation":[]},{"given":"Cathal","family":"Gurrin","sequence":"additional","affiliation":[]},{"given":"Liting","family":"Zhou","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,3,15]]},"reference":[{"key":"18_CR1","doi-asserted-by":"crossref","unstructured":"Anderson, P., et al.: Bottom-up and top-down attention for image captioning and visual question answering, pp. 6077\u20136086 (2018)","DOI":"10.1109\/CVPR.2018.00636"},{"key":"18_CR2","unstructured":"Bao, H., et al.: Unilmv2: pseudo-masked language models for unified language model pre-training. In: International Conference on Machine Learning, pp. 642\u2013652. PMLR (2020)"},{"issue":"1","key":"18_CR3","first-page":"101","volume":"176","author":"V Bush","year":"1945","unstructured":"Bush, V., et al.: As we may think. The atlantic monthly 176(1), 101\u2013108 (1945)","journal-title":"The atlantic monthly"},{"key":"18_CR4","doi-asserted-by":"crossref","unstructured":"Byrne, D., Kelliher, A., Jones, G.J.: Life editing: third-party perspectives on lifelog content. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 1501\u20131510 (2011)","DOI":"10.1145\/1978942.1979162"},{"key":"18_CR5","unstructured":"Castro, S., Azab, M., Stroud, J., Noujaim, C., Wang, R., Deng, J., Mihalcea, R.: Lifeqa: a real-life dataset for video question answering. In: Proceedings of the 12th Language Resources and Evaluation Conference, pp. 4352\u20134358 (2020)"},{"key":"18_CR6","unstructured":"Doherty, A., Smeaton, A.: Automatically segmenting LifeLog data into events"},{"key":"18_CR7","doi-asserted-by":"crossref","unstructured":"Fan, C.: EgoVQA - an egocentric video question answering benchmark dataset. In: 2019 IEEE\/CVF International Conference on Computer Vision Workshop (ICCVW), pp. 4359\u20134366 (Oct 2019), iSSN: 2473\u20139944","DOI":"10.1109\/ICCVW.2019.00536"},{"key":"18_CR8","doi-asserted-by":"crossref","unstructured":"Fukui, A., Park, D.H., Yang, D., Rohrbach, A., Darrell, T., Rohrbach, M.: Multimodal Compact Bilinear Pooling for Visual Question Answering and Visual Grounding. arXiv:1606.01847 [cs], September 2016","DOI":"10.18653\/v1\/D16-1044"},{"key":"18_CR9","doi-asserted-by":"crossref","unstructured":"Gao, Y., Bing, L., Li, P., King, I., Lyu, M.R.: Generating distractors for reading comprehension questions from real examinations. In: AAAI-19 AAAI Conference on Artificial Intelligence (2019)","DOI":"10.1609\/aaai.v33i01.33016423"},{"key":"18_CR10","doi-asserted-by":"publisher","first-page":"89","DOI":"10.1145\/1107458.1107460","volume":"49","author":"J Gemmell","year":"2006","unstructured":"Gemmell, J., Bell, C., Lueder, R.: Mylifebits: a personal database for everything. Commun. ACM 49, 89\u201395 (2006)","journal-title":"Commun. ACM"},{"key":"18_CR11","unstructured":"Gurrin, C., et al.: Overview of the NTCIR-14 lifelog-3 task. In: Proceedings of the 14th NTCIR Conference, p. 13. NII (2019)"},{"key":"18_CR12","doi-asserted-by":"crossref","unstructured":"Gurrin, C., et al.: Introduction to the third annual lifelog search challenge (LSC\u201920). In: Proceedings of the 2020 International Conference on Multimedia Retrieval, ICMR 2020, pp. 584\u2013585. Association for Computing Machinery","DOI":"10.1145\/3372278.3388043"},{"key":"18_CR13","doi-asserted-by":"crossref","unstructured":"Gurrin, C., Smeaton, A.F., Doherty, A.R., et al.: Lifelogging: personal big data. Found. Trends Inform. Retrieval 8(1), 1\u2013125 (2014)","DOI":"10.1561\/1500000033"},{"key":"18_CR14","doi-asserted-by":"crossref","unstructured":"Hu, R., Andreas, J., Rohrbach, M., Darrell, T., Saenko, K.: Learning to reason: end-to-end module networks for visual question answering. arXiv:1704.05526 [cs], Septrmber 2017. arXiv: 1704.05526 version: 3","DOI":"10.1109\/ICCV.2017.93"},{"key":"18_CR15","unstructured":"Jang, Y., Song, Y., Yu, Y., Kim, Y., Kim, G.: TGIF-QA: toward spatio-temporal reasoning in visual question answering"},{"key":"18_CR16","unstructured":"Jia, Y., et al.: Caffe: convolutional architecture for fast feature embedding"},{"key":"18_CR17","doi-asserted-by":"crossref","unstructured":"Lei, J., Yu, L., Bansal, M., Berg, T.L.: TVQA: localized, compositional video question answering. arXiv:1809.01696 [cs] (May 2019), arXiv: 1809.01696","DOI":"10.18653\/v1\/D18-1167"},{"key":"18_CR18","doi-asserted-by":"crossref","unstructured":"Lei, J., Yu, L., Berg, T.L., Bansal, M.: TVQA+: spatio-temporal grounding for video question answering. arXiv:1904.11574 [cs], May 2020. arXiv: 1904.11574","DOI":"10.18653\/v1\/2020.acl-main.730"},{"key":"18_CR19","doi-asserted-by":"crossref","unstructured":"Loko\u010d, J., et al.: Is the reign of interactive search eternal? findings from the video browser showdown 2020. ACM Trans. Multimedia Comput. Commun. Appl. 17(3), July 2021","DOI":"10.1145\/3445031"},{"key":"18_CR20","doi-asserted-by":"crossref","unstructured":"Nguyen, T.N., et al.: Lifeseeker 3.0: An interactive lifelog search engine for lsc\u201921. In: Proceedings of the 4th Annual on Lifelog Search Challenge, pp. 41\u201346 (2021)","DOI":"10.1145\/3463948.3469065"},{"key":"18_CR21","unstructured":"Ninh, V.T., Le, T.K., Zhou, L., Piras, L., Riegler, M.: Overview of ImageCLEFlifelog 2020: Lifelog moment retrieval and sport performance lifelog. In: CLEF (Working Notes), p. 17 (2020)"},{"key":"18_CR22","unstructured":"Pennington, J., Socher, R., Manning, C.D.: GloVe: global vectors for word representation. In: Empirical Methods in Natural Language Processing (EMNLP), pp. 1532\u20131543"},{"key":"18_CR23","doi-asserted-by":"publisher","first-page":"249","DOI":"10.1162\/tacl_a_00266","volume":"7","author":"S Reddy","year":"2019","unstructured":"Reddy, S., Chen, D., Manning, C.D.: CoQA: a conversational question answering challenge. Trans. Assoc. Comput. Linguist. 7, 249\u2013266 (2019)","journal-title":"Trans. Assoc. Comput. Linguist."},{"key":"18_CR24","doi-asserted-by":"crossref","unstructured":"Sellen, A.J., Whittaker, S.: Beyond total capture: a constructive critique of lifelogging 53(5), 70\u201377","DOI":"10.1145\/1735223.1735243"},{"key":"18_CR25","doi-asserted-by":"crossref","unstructured":"Speer, R., Chin, J., Havasi, C.: Conceptnet 5.5: an open multilingual graph of general knowledge. In: Thirty-First AAAI Conference on Artificial Intelligence (2017)","DOI":"10.1609\/aaai.v31i1.11164"},{"key":"18_CR26","doi-asserted-by":"crossref","unstructured":"Tran, L.D., Nguyen, M.D., Thanh Binh, N., Lee, H., Gurrin, C.: Mysc\u00e9al 2.0: a revised experimental interactive lifelog retrieval system for lsc\u201921. In: Proceedings of the 4th Annual on Lifelog Search Challenge, pp. 11\u201316 (2021)","DOI":"10.1145\/3463948.3469064"},{"key":"18_CR27","doi-asserted-by":"crossref","unstructured":"Trotman, A., Geva, S., Kamps, J.: Report on the sigir 2007 workshop on focused retrieval. In: ACM SIGIR Forum, vol. 41, pp. 97\u2013103. ACM, New York (2007)","DOI":"10.1145\/1328964.1328981"},{"key":"18_CR28","doi-asserted-by":"crossref","unstructured":"Xu, D., et al.: Video question answering via gradually refined attention over appearance and motion. In: Proceedings of the 25th ACM International Conference on Multimedia, MM 2017, pp. 1645\u20131653. Association for Computing Machinery, event-place: Mountain View, California, USA","DOI":"10.1145\/3123266.3123427"},{"key":"18_CR29","doi-asserted-by":"crossref","unstructured":"Ye, Y., Zhao, Z., Li, Y., Chen, L., Xiao, J., Zhuang, Y.: Video question answering via attribute-augmented attention network learning. In: Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 829\u2013832 (2017)","DOI":"10.1145\/3077136.3080655"}],"container-title":["Lecture Notes in Computer Science","MultiMedia Modeling"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-98358-1_18","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,2,19]],"date-time":"2024-02-19T10:10:12Z","timestamp":1708337412000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-98358-1_18"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783030983574","9783030983581"],"references-count":29,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-98358-1_18","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"15 March 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MMM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Multimedia Modeling","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Phu Quoc","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Vietnam","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6 June 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"10 June 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"mmm2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Conftool Pro","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"212","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"84","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"40% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}