{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,12]],"date-time":"2025-10-12T03:26:22Z","timestamp":1760239582409,"version":"build-2065373602"},"publisher-location":"Cham","reference-count":18,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783032075017","type":"print"},{"value":"9783032075024","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,10,12]],"date-time":"2025-10-12T00:00:00Z","timestamp":1760227200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,10,12]],"date-time":"2025-10-12T00:00:00Z","timestamp":1760227200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-3-032-07502-4_1","type":"book-chapter","created":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T10:43:20Z","timestamp":1760179400000},"page":"1-9","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["GMAT: Grounded Multi-agent Clinical Description Generation for\u00a0Text Encoder in\u00a0Vision-Language MIL for\u00a0Whole Slide Image Classification"],"prefix":"10.1007","author":[{"given":"Ngoc Bui Lam","family":"Quang","sequence":"first","affiliation":[]},{"given":"Nam Le Nguyen","family":"Binh","sequence":"additional","affiliation":[]},{"given":"Thanh-Huy","family":"Nguyen","sequence":"additional","affiliation":[]},{"given":"Le Thien Phuc","family":"Nguyen","sequence":"additional","affiliation":[]},{"given":"Quan","family":"Nguyen","sequence":"additional","affiliation":[]},{"given":"Ulas","family":"Bagci","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,10,12]]},"reference":[{"key":"1_CR1","unstructured":"Ilse, M., Tomczak, J., Welling, M.: Attention-based deep multiple instance learning, In: International Conference on Machine Learning, pp.\u00a02127\u20132136, PMLR (2018)"},{"issue":"6","key":"1_CR2","doi-asserted-by":"publisher","first-page":"555","DOI":"10.1038\/s41551-020-00682-w","volume":"5","author":"MY Lu","year":"2021","unstructured":"Lu, M.Y., Williamson, D.F., Chen, T.Y., Chen, R.J., Barbieri, M., Mahmood, F.: Data-efficient and weakly supervised computational pathology on whole-slide images. Nat. Biomed. Eng. 5(6), 555\u2013570 (2021)","journal-title":"Nat. Biomed. Eng."},{"key":"1_CR3","first-page":"2136","volume":"34","author":"Z Shao","year":"2021","unstructured":"Shao, Z., Bian, H., Chen, Y., Wang, Y., Zhang, J., Ji, X., et al.: Transmil: transformer based correlated multiple instance learning for whole slide image classification. Adv. Neural. Inf. Process. Syst. 34, 2136\u20132147 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"1_CR4","doi-asserted-by":"crossref","unstructured":"Chen, R.J., et al.: Scaling vision transformers to gigapixel images via hierarchical self-supervised learning,\u2019 In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp.\u00a016144\u201316155 (2022)","DOI":"10.1109\/CVPR52688.2022.01567"},{"key":"1_CR5","doi-asserted-by":"crossref","unstructured":"Li, B., Li, Y., Eliceiri, K.W.: Dual-stream multiple instance learning network for whole slide image classification with self-supervised contrastive learning, In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp.\u00a014318\u201314328 (2021)","DOI":"10.1109\/CVPR46437.2021.01409"},{"key":"1_CR6","unstructured":"Fourkioti, O., De Vries, M., Bakal, C.: CAMIL: context-aware multiple instance learning for cancer detection and subtyping in whole slide images, In: The Twelfth International Conference on Learning Representations (2024)"},{"key":"1_CR7","unstructured":"Zhang, H., et al.: DTFD-MIL: Double-tier feature distillation multiple instance learning for histopathology whole slide image classification."},{"key":"1_CR8","doi-asserted-by":"crossref","unstructured":"Jafarinia, H., Alipanah, A., Razavi, S., Mirzaie, N., Rohban, M.H.: Snuffy: Efficient whole slide image classifier (2024)","DOI":"10.1007\/978-3-031-73024-5_15"},{"key":"1_CR9","doi-asserted-by":"crossref","unstructured":"Qu, L., Luo, X., Liu, S., Wang, M., Song, Z.: Dgmil: distribution guided multiple instance learning for whole slide image classification, In: International Conference on Medical Image Computing and Computer-assisted Intervention, pp.\u00a024\u201334, Springer (2022)","DOI":"10.1007\/978-3-031-16434-7_3"},{"key":"1_CR10","unstructured":"Nguyen, A.-T., et al.: Mgpath: vision-language model with multi-granular prompt learning for few-shot WSI classification (2025)"},{"key":"1_CR11","unstructured":"Nguyen, A.-T., et al.: Few-shot whole slide pathology classification with multi-granular vision-language models, In: ICLR 2025 Workshop on Foundation Models in the Wild (2025)"},{"key":"1_CR12","doi-asserted-by":"crossref","unstructured":"Han, M., Qu, L., Yang, D., Zhang, X., Wang, X., Zhang, L.: Mscpt: few-shot whole slide image classification with multi-scale and context-focused prompt tuning, IEEE Transactions on Medical Imaging, pp.\u00a01\u20131 (2025)","DOI":"10.1109\/TMI.2025.3564976"},{"key":"1_CR13","unstructured":"Zhang, S., et al.: Biomedclip: a multimodal biomedical foundation model pretrained from fifteen million scientific image-text pairs, arXiv preprint arXiv:2403.xxxxx (2024)"},{"issue":"9","key":"1_CR14","doi-asserted-by":"publisher","first-page":"2307","DOI":"10.1038\/s41591-023-02504-3","volume":"29","author":"Z Huang","year":"2023","unstructured":"Huang, Z., Bianchi, F., Yuksekgonul, M., Montine, T.J., Zou, J.: A visual-language foundation model for pathology image analysis using medical twitter. Nat. Med. 29(9), 2307\u20132316 (2023)","journal-title":"Nat. Med."},{"key":"1_CR15","doi-asserted-by":"crossref","unstructured":"Lu, M.Y., et al.: Visual language pretrained multiple instance zero-shot transfer for histopathology images, In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp.\u00a019764\u201319775 (2023)","DOI":"10.1109\/CVPR52729.2023.01893"},{"key":"1_CR16","doi-asserted-by":"crossref","unstructured":"Shi, J., Li, C., Gong, T., Zheng, Y., Fu, H.: Vila-mil: Dual-scale vision-language multiple instance learning for whole slide image classification, In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp.\u00a011248\u201311258 (2024)","DOI":"10.1109\/CVPR52733.2024.01069"},{"key":"1_CR17","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2022.102559","volume":"81","author":"X Wang","year":"2022","unstructured":"Wang, X., et al.: Transformer-based unsupervised contrastive learning for histopathological image classification. Med. Image Anal. 81, 102559 (2022)","journal-title":"Med. Image Anal."},{"key":"1_CR18","unstructured":"Xu, H., et\u00a0al.: A whole-slide foundation model for digital pathology from real-world data, Nature, pp.\u00a01\u20138 (2024)"}],"container-title":["Lecture Notes in Computer Science","Emerging LLM\/LMM Applications in Medical Imaging"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-032-07502-4_1","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T10:43:24Z","timestamp":1760179404000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-032-07502-4_1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,12]]},"ISBN":["9783032075017","9783032075024"],"references-count":18,"URL":"https:\/\/doi.org\/10.1007\/978-3-032-07502-4_1","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,10,12]]},"assertion":[{"value":"12 October 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ELAMI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Workshop on Emerging LLM\/LMM Applications in Medical Imaging","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Daejeon","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Korea (Republic of)","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 September 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 September 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"elami2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/hula-ai.github.io\/LLM-MI-MICCAI-2025\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}