{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,9]],"date-time":"2026-04-09T20:26:21Z","timestamp":1775766381750,"version":"3.50.1"},"publisher-location":"Cham","reference-count":45,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783030872397","type":"print"},{"value":"9783030872403","type":"electronic"}],"license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021]]},"DOI":"10.1007\/978-3-030-87240-3_7","type":"book-chapter","created":{"date-parts":[[2021,9,23]],"date-time":"2021-09-23T07:44:03Z","timestamp":1632383043000},"page":"64-74","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":103,"title":["Multiple Meta-model Quantifying for Medical Visual Question Answering"],"prefix":"10.1007","author":[{"given":"Tuong","family":"Do","sequence":"first","affiliation":[]},{"given":"Binh X.","family":"Nguyen","sequence":"additional","affiliation":[]},{"given":"Erman","family":"Tjiputra","sequence":"additional","affiliation":[]},{"given":"Minh","family":"Tran","sequence":"additional","affiliation":[]},{"given":"Quang D.","family":"Tran","sequence":"additional","affiliation":[]},{"given":"Anh","family":"Nguyen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,9,21]]},"reference":[{"key":"7_CR1","unstructured":"Abacha, A.B., Gayen, S., Lau, J.J., Rajaraman, S., Demner-Fushman, D.: NLM at ImageCLEF 2018 visual question answering in the medical domain. In: CEUR Workshop Proceedings (2018)"},{"key":"7_CR2","unstructured":"Abacha, A.B., Hasan, S.A., Datla, V.V., Liu, J., Demner-Fushman, D., M\u00fcller, H.: VQA-Med: overview of the medical visual question answering task at ImageCLEF 2019. In: CLEF (Working Notes) (2019)"},{"key":"7_CR3","doi-asserted-by":"crossref","unstructured":"Bar, Y., Diamant, I., Wolf, L., Greenspan, H.: Deep learning with non-medical training used for chest pathology identification. In: Medical Imaging: Computer-Aided Diagnosis (2015)","DOI":"10.1117\/12.2083124"},{"key":"7_CR4","doi-asserted-by":"crossref","unstructured":"Chi, W., et al.: Collaborative robot-assisted endovascular catheterization with generative adversarial imitation learning. In: ICRA (2020)","DOI":"10.1109\/ICRA40945.2020.9196912"},{"key":"7_CR5","doi-asserted-by":"crossref","unstructured":"Do, T., Do, T.T., Tran, H., Tjiputra, E., Tran, Q.D.: Compact trilinear interaction for visual question answering. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00048"},{"key":"7_CR6","unstructured":"Finn, C., Abbeel, P., Levine, S.: Model-agnostic meta-learning for fast adaptation of deep networks. In: ICML (2017)"},{"key":"7_CR7","doi-asserted-by":"crossref","unstructured":"Fukui, A., Park, D.H., Yang, D., Rohrbach, A., Darrell, T., Rohrbach, M.: Multimodal compact bilinear pooling for visual question answering and visual grounding. In: EMNLP (2016)","DOI":"10.18653\/v1\/D16-1044"},{"key":"7_CR8","doi-asserted-by":"crossref","unstructured":"Gupta, D., Suman, S., Ekbal, A.: Hierarchical deep multi-modal network for medical visual question answering. Expert Syst. Appl. (2021)","DOI":"10.1016\/j.eswa.2020.113993"},{"key":"7_CR9","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"7_CR10","doi-asserted-by":"crossref","unstructured":"He, X., Zhang, Y., Mou, L., Xing, E., Xie, P.: PathVQA: 30000+ questions for medical visual question answering. arXiv preprint arXiv:2003.10286 (2020)","DOI":"10.36227\/techrxiv.13127537.v1"},{"key":"7_CR11","unstructured":"Hsu, K., Levine, S., Finn, C.: Unsupervised learning via meta-learning. In: ICLR (2019)"},{"issue":"8","key":"7_CR12","doi-asserted-by":"publisher","first-page":"1389","DOI":"10.1007\/s11548-020-02205-z","volume":"15","author":"B Huang","year":"2020","unstructured":"Huang, B., et al.: Tracking and visualization of the sensing area for a tethered laparoscopic gamma probe. Int. J. Comput. Assist. Radiol. Surg. 15(8), 1389\u20131397 (2020). https:\/\/doi.org\/10.1007\/s11548-020-02205-z","journal-title":"Int. J. Comput. Assist. Radiol. Surg."},{"key":"7_CR13","doi-asserted-by":"crossref","unstructured":"Huang, B., et al.: Self-supervised generative adversarial network for depth estimation in laparoscopic images. In: MICCAI (2021)","DOI":"10.1007\/978-3-030-87202-1_22"},{"key":"7_CR14","unstructured":"Khodadadeh, S., B\u00f6l\u00f6ni, L., Shah, M.: Unsupervised meta-learning for few-shot image classification. In: NIPS (2019)"},{"key":"7_CR15","unstructured":"Kim, J.H., Jun, J., Zhang, B.T.: Bilinear attention networks. In: NIPS (2018)"},{"key":"7_CR16","unstructured":"Koch, G., Zemel, R., Salakhutdinov, R.: Siamese neural networks for one-shot image recognition. In: ICML Deep Learning Workshop (2015)"},{"key":"7_CR17","unstructured":"Kornuta, T., Rajan, D., Shivade, C., Asseman, A., Ozcan, A.S.: Leveraging medical visual question answering with supporting facts. arXiv:1905.12008 (2019)"},{"key":"7_CR18","doi-asserted-by":"crossref","unstructured":"Lau, J.J., Gayen, S., Abacha, A.B., Demner-Fushman, D.: A dataset of clinically generated visual questions and answers about radiology images. Nature (2018)","DOI":"10.1038\/sdata.2018.251"},{"key":"7_CR19","unstructured":"Liu, S., Ding, H., Zhou, X.: Shengyan at VQA-Med 2020: an encoder-decoder model for medical domain visual question answering task. CLEF (2020)"},{"key":"7_CR20","doi-asserted-by":"crossref","unstructured":"Lubna, A., Kalady, S., Lijiya, A.: MoBVQA: a modality based medical image visual question answering system. In: TENCON (2019)","DOI":"10.1109\/TENCON.2019.8929456"},{"key":"7_CR21","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"546","DOI":"10.1007\/978-3-030-00928-1_62","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2018","author":"G Maicas","year":"2018","unstructured":"Maicas, G., Bradley, A.P., Nascimento, J.C., Reid, I., Carneiro, G.: Training medical image analysis systems like radiologists. In: Frangi, A.F., Schnabel, J.A., Davatzikos, C., Alberola-L\u00f3pez, C., Fichtinger, G. (eds.) MICCAI 2018. LNCS, vol. 11070, pp. 546\u2013554. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-00928-1_62"},{"key":"7_CR22","unstructured":"Munkhdalai, T., Yu, H.: Meta networks. In: ICML (2017)"},{"key":"7_CR23","unstructured":"Nguyen, A.: Scene understanding for autonomous manipulation with deep learning. arXiv preprint arXiv:1903.09761 (2019)"},{"key":"7_CR24","doi-asserted-by":"crossref","unstructured":"Nguyen, A., et al.: End-to-end real-time catheter segmentation with optical flow-guided warping during endovascular intervention. In: ICRA (2020)","DOI":"10.1109\/ICRA40945.2020.9197307"},{"key":"7_CR25","doi-asserted-by":"crossref","unstructured":"Nguyen, A., Nguyen, N., Tran, K., Tjiputra, E., Tran, Q.: Autonomous navigation in complex environments with deep multimodal fusion network. In: IROS (2020)","DOI":"10.1109\/IROS45743.2020.9341494"},{"key":"7_CR26","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"522","DOI":"10.1007\/978-3-030-32251-9_57","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2019","author":"BD Nguyen","year":"2019","unstructured":"Nguyen, B.D., Do, T.-T., Nguyen, B.X., Do, T., Tjiputra, E., Tran, Q.D.: Overcoming data limitation in medical visual question answering. In: Shen, D., et al. (eds.) MICCAI 2019. LNCS, vol. 11767, pp. 522\u2013530. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-32251-9_57"},{"key":"7_CR27","unstructured":"Nichol, A., Achiam, J., Schulman, J.: On first-order meta-learning algorithms. arXiv preprint arXiv:1803.02999 (2018)"},{"key":"7_CR28","unstructured":"Peng, Y., Liu, F., Rosen, M.P.: UMass at ImageCLEF medical visual question answering (Med-VQA) 2018 task. In: CEUR Workshop Proceedings (2018)"},{"key":"7_CR29","doi-asserted-by":"crossref","unstructured":"Pennington, J., Socher, R., Manning, C.D.: GloVe: global vectors for word representation. In: EMNLP (2014)","DOI":"10.3115\/v1\/D14-1162"},{"key":"7_CR30","unstructured":"Ravi, S., Larochelle, H.: Optimization as a model for few-shot learning. In: ICLR (2017)"},{"key":"7_CR31","doi-asserted-by":"publisher","first-page":"50626","DOI":"10.1109\/ACCESS.2020.2980024","volume":"8","author":"F Ren","year":"2020","unstructured":"Ren, F., Zhou, Y.: CGMVQA: a new classification and generative model for medical visual question answering. IEEE Access 8, 50626\u201350636 (2020)","journal-title":"IEEE Access"},{"issue":"3","key":"7_CR32","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","volume":"115","author":"O Russakovsky","year":"2015","unstructured":"Russakovsky, O., et al.: ImageNet large scale visual recognition challenge. Int. J. Comput. Vis. 115(3), 211\u2013252 (2015). https:\/\/doi.org\/10.1007\/s11263-015-0816-y","journal-title":"Int. J. Comput. Vis."},{"key":"7_CR33","unstructured":"Santoro, A., Bartunov, S., Botvinick, M., Wierstra, D., Lillicrap, T.: Meta-learning with memory-augmented neural networks. In: ICML (2016)"},{"key":"7_CR34","unstructured":"Schmidhuber, J.: Evolutionary principles in self-referential learning (1987)"},{"key":"7_CR35","unstructured":"Shi, L., Liu, F., Rosen, M.P.: Deep multimodal learning for medical visual question answering. In: CLEF (Working Notes) (2019)"},{"key":"7_CR36","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. In: ICLR (2015)"},{"key":"7_CR37","unstructured":"Snell, J., Swersky, K., Zemel, R.S.: Prototypical networks for few-shot learning. In: NIPS (2017)"},{"key":"7_CR38","doi-asserted-by":"crossref","unstructured":"Sung, F., Yang, Y., Zhang, L., Xiang, T., Torr, P.H., Hospedales, T.M.: Learning to compare: relation network for few-shot learning. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00131"},{"key":"7_CR39","unstructured":"Vinyals, O., Blundell, C., Lillicrap, T., Kavukcuoglu, K., Wierstra, D.: Matching networks for one shot learning. In: NIPS (2016)"},{"key":"7_CR40","first-page":"2856","volume":"39","author":"MH Vu","year":"2020","unstructured":"Vu, M.H., L\u00f6fstedt, T., Nyholm, T., Sznitman, R.: A question-centric model for visual question answering in medical imaging. IEEE TMI 39, 2856\u20132868 (2020)","journal-title":"IEEE TMI"},{"key":"7_CR41","unstructured":"Vu, M., Sznitman, R., Nyholm, T., L\u00f6fstedt, T.: Ensemble of streamlined bilinear visual question answering models for the ImageCLEF 2019 challenge in the medical domain. In: Conference and Labs of the Evaluation Forum (2019)"},{"key":"7_CR42","unstructured":"Wang, Y.X., Hebert, M.: Learning from small sample sets by combining unsupervised meta-training with CNNs. In: NIPS (2016)"},{"key":"7_CR43","doi-asserted-by":"crossref","unstructured":"Yang, Z., He, X., Gao, J., Deng, L., Smola, A.J.: Stacked attention networks for image question answering. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.10"},{"key":"7_CR44","doi-asserted-by":"crossref","unstructured":"Zhan, L.M., Liu, B., Fan, L., Chen, J., Wu, X.M.: Medical visual question answering via conditional reasoning. In: ACM International Conference on Multimedia (2020)","DOI":"10.1145\/3394171.3413761"},{"key":"7_CR45","unstructured":"Zhou, Y., Kang, X., Ren, F.: Employing Inception-Resnet-v2 and Bi-LSTM for medical domain visual question answering. In: CEUR Workshop Proceedings (2018)"}],"container-title":["Lecture Notes in Computer Science","Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2021"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-87240-3_7","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,12,4]],"date-time":"2021-12-04T23:03:39Z","timestamp":1638659019000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-87240-3_7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"ISBN":["9783030872397","9783030872403"],"references-count":45,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-87240-3_7","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021]]},"assertion":[{"value":"21 September 2021","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MICCAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Medical Image Computing and Computer-Assisted Intervention","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Strasbourg","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"France","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 September 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1 October 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"24","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"miccai2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/miccai2021.org\/en\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1622","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"531","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"33% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"The conference was held virtually.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}