{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T17:45:11Z","timestamp":1777657511267,"version":"3.51.4"},"publisher-location":"Cham","reference-count":80,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031734137","type":"print"},{"value":"9783031734144","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,10,25]],"date-time":"2024-10-25T00:00:00Z","timestamp":1729814400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,10,25]],"date-time":"2024-10-25T00:00:00Z","timestamp":1729814400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-73414-4_2","type":"book-chapter","created":{"date-parts":[[2024,10,24]],"date-time":"2024-10-24T17:02:54Z","timestamp":1729789374000},"page":"19-37","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":19,"title":["Multistain Pretraining for\u00a0Slide Representation Learning in\u00a0Pathology"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-3832-1390","authenticated-orcid":false,"given":"Guillaume","family":"Jaume","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6378-6010","authenticated-orcid":false,"given":"Anurag","family":"Vaidya","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9432-2793","authenticated-orcid":false,"given":"Andrew","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Andrew","family":"H. Song","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9356-9156","authenticated-orcid":false,"given":"Richard","family":"J. Chen","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3194-1973","authenticated-orcid":false,"given":"Sharifa","family":"Sahai","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1380-6746","authenticated-orcid":false,"given":"Dandan","family":"Mo","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4836-444X","authenticated-orcid":false,"given":"Emilio","family":"Madrigal","sequence":"additional","affiliation":[]},{"given":"Long","family":"Phi Le","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7587-1562","authenticated-orcid":false,"given":"Faisal","family":"Mahmood","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,10,25]]},"reference":[{"key":"2_CR1","unstructured":"AIDPATH DB (2024). https:\/\/mitel.dimi.uniud.it\/aidpath-db\/app\/login.php. Accessed 6 Mar 2024"},{"issue":"9","key":"2_CR2","doi-asserted-by":"publisher","first-page":"1773","DOI":"10.1038\/s41591-022-01981-2","volume":"28","author":"JN Acosta","year":"2022","unstructured":"Acosta, J.N., Falcone, G.J., Rajpurkar, P., Topol, E.J.: Multimodal biomedical AI. Nat. Med. 28(9), 1773\u20131784 (2022). https:\/\/doi.org\/10.1038\/s41591-022-01981-2","journal-title":"Nat. Med."},{"key":"2_CR3","unstructured":"Akbarnejad, A., Ray, N., Barnes, P.J., Bigras, G.: Predicting Ki67, ER, PR, and HER2 Statuses from H &E-stained Breast Cancer Images. arXiv (2023)"},{"key":"2_CR4","first-page":"23716","volume":"35","author":"JB Alayrac","year":"2022","unstructured":"Alayrac, J.B., et al.: Flamingo: a visual language model for few-shot learning. Adv. Neural. Inf. Process. Syst. 35, 23716\u201323736 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"2_CR5","doi-asserted-by":"crossref","unstructured":"Anand, D., et al.: Deep learning to estimate human epidermal growth factor receptor 2 status from hematoxylin and eosin-stained breast tissue images. J. Pathol. Inform. 11 (2020)","DOI":"10.4103\/jpi.jpi_10_20"},{"key":"2_CR6","doi-asserted-by":"crossref","unstructured":"Aryal, M., Yahyasoltani, N.: Context-aware self-supervised learning of whole slide images. arXiv (2023)","DOI":"10.1109\/TAI.2024.3365779"},{"key":"2_CR7","unstructured":"Azizi, S., et\u00a0al.: Robust and data-efficient generalization of self-supervised machine learning for diagnostic imaging. Nat. Biomed. Eng., 1\u201324 (2023)"},{"key":"2_CR8","doi-asserted-by":"crossref","unstructured":"Bachmann, R., Mizrahi, D., Atanov, A., Zamir, A.: MultiMAE: multi-modal multi-task masked autoencoders. arXiv (2022)","DOI":"10.1007\/978-3-031-19836-6_20"},{"key":"2_CR9","unstructured":"Campanella, G., et al.: Computational pathology at health system scale \u2013 self-supervised foundation models from three billion images. arXiv (2023)"},{"key":"2_CR10","doi-asserted-by":"crossref","unstructured":"Caron, M., et al.: Emerging properties in self-supervised vision transformers. arXiv preprint arXiv:2104.14294 (2021)","DOI":"10.1109\/ICCV48922.2021.00951"},{"key":"2_CR11","unstructured":"Chen, L., Gan, Z., Cheng, Y., Li, L., Carin, L., Liu, J.: Graph optimal transport for cross-domain alignment. In: International Conference on Machine Learning, pp. 1542\u20131553. PMLR (2020)"},{"key":"2_CR12","doi-asserted-by":"crossref","unstructured":"Chen, R.J., et al.: Scaling vision transformers to gigapixel images via hierarchical self-supervised learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 16144\u201316155 (2022)","DOI":"10.1109\/CVPR52688.2022.01567"},{"key":"2_CR13","doi-asserted-by":"crossref","unstructured":"Chen, R.J., et al.: Towards a general-purpose foundation model for computational pathology. Nat. Med. (2024)","DOI":"10.1038\/s41591-024-02857-3"},{"issue":"8","key":"2_CR14","doi-asserted-by":"publisher","first-page":"865","DOI":"10.1016\/j.ccell.2022.07.004","volume":"40","author":"RJ Chen","year":"2022","unstructured":"Chen, R.J., et al.: Pan-cancer integrative histology-genomic analysis via multimodal deep learning. Cancer Cell 40(8), 865\u2013878 (2022)","journal-title":"Cancer Cell"},{"key":"2_CR15","unstructured":"Chen, T., Kornblith, S., Norouzi, M., Hinton, G.: A simple framework for contrastive learning of visual representations. In: International Conference on Machine Learning, pp. 1597\u20131607. PMLR (2020)"},{"key":"2_CR16","doi-asserted-by":"publisher","first-page":"104","DOI":"10.1007\/978-3-030-58577-8_7","volume-title":"Computer Vision \u2013 ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XXX","author":"Y-C Chen","year":"2020","unstructured":"Chen, Y.-C., et al.: UNITER: universal image-text representation learning. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) Computer Vision \u2013 ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XXX, pp. 104\u2013120. Springer International Publishing, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58577-8_7"},{"key":"2_CR17","doi-asserted-by":"crossref","unstructured":"Couture, H.D., et al.: Image analysis with deep learning to predict breast cancer grade, ER status, histologic subtype, and intrinsic subtype. NPJ Breast Cancer 4(30), 1\u20138 (2018)","DOI":"10.1038\/s41523-018-0079-1"},{"key":"2_CR18","unstructured":"Dosovitskiy, A., et\u00a0al.: An image is worth 16$$\\,\\times \\,$$16 words: transformers for image recognition at scale. In: International Conference on Learning Representations (2021)"},{"issue":"1","key":"2_CR19","doi-asserted-by":"publisher","first-page":"44","DOI":"10.1038\/s41379-021-00911-w","volume":"35","author":"S Farahmand","year":"2022","unstructured":"Farahmand, S., et al.: Deep learning trained on hematoxylin and eosin tumor region of Interest predicts HER2 status and trastuzumab treatment response in HER2+ breast cancer. Mod. Pathol. 35(1), 44\u201351 (2022)","journal-title":"Mod. Pathol."},{"key":"2_CR20","doi-asserted-by":"crossref","unstructured":"Filiot, A., et al.: Scaling self-supervised learning for histopathology with masked image modeling. medRxiv (2023)","DOI":"10.1101\/2023.07.21.23292757"},{"key":"2_CR21","doi-asserted-by":"crossref","unstructured":"Gamper, J., Rajpoot, N.: Multiple instance captioning: learning representations from histopathology textbooks and articles. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 16549\u201316559 (2021)","DOI":"10.1109\/CVPR46437.2021.01628"},{"key":"2_CR22","doi-asserted-by":"publisher","first-page":"401","DOI":"10.1038\/s42256-022-00471-x","volume":"4","author":"P Ghahremani","year":"2022","unstructured":"Ghahremani, P., et al.: Deep learning-inferred multiplex immunofluorescence for immunohistochemical image quantification. Nat. Mach. Intell. 4, 401\u2013412 (2022)","journal-title":"Nat. Mach. Intell."},{"issue":"7","key":"2_CR23","doi-asserted-by":"publisher","DOI":"10.1001\/jamanetworkopen.2019.7700","volume":"2","author":"M Gil Shamai","year":"2019","unstructured":"Gil Shamai, M.: Artificial intelligence algorithms to assess hormonal status from tissue microarrays in patients with breast. JAMA Netw. Open 2(7), e197700 (2019)","journal-title":"JAMA Netw. Open"},{"key":"2_CR24","doi-asserted-by":"crossref","unstructured":"Hua, S., Yan, F., Shen, T., Zhang, X.: PathoDuet: foundation models for pathological slide analysis of H &E and IHC stains (2023)","DOI":"10.1016\/j.media.2024.103289"},{"key":"2_CR25","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1038\/s41591-023-02504-3","volume":"29","author":"Z Huang","year":"2023","unstructured":"Huang, Z., Bianchi, F., Yuksekgonul, M., Montine, T., Zou, J.: A visual-language foundation model for pathology image analysis using medical twitter. Nat. Med. 29, 1\u201310 (2023)","journal-title":"Nat. Med."},{"key":"2_CR26","unstructured":"Ilse, M., Tomczak, J., Welling, M.: Attention-based deep multiple instance learning. In: International Conference on Machine Learning, pp. 2127\u20132136. PMLR (2018)"},{"key":"2_CR27","unstructured":"Jaegle, A., et al.: Perceiver IO: a general architecture for structured inputs & outputs. In: International Conference on Learning Representations (2022)"},{"key":"2_CR28","unstructured":"Jaume, G., et al.: HEST-1k: a dataset for spatial transcriptomics and histology image analysis. arXiv (2024)"},{"key":"2_CR29","doi-asserted-by":"crossref","unstructured":"Jaume, G., et al.: Transcriptomics-guided slide representation learning in computational pathology. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2024)","DOI":"10.1109\/CVPR52733.2024.00920"},{"key":"2_CR30","doi-asserted-by":"crossref","unstructured":"Jaume, G., Vaidya, A., Chen, R., Williamson, D., Liang, P., Mahmood, F.: Modeling dense multimodal interactions between biological pathways and histology for survival prediction. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2024)","DOI":"10.1109\/CVPR52733.2024.01100"},{"key":"2_CR31","unstructured":"Jia, C., et al.: Scaling up visual and vision-language representation learning with noisy text supervision. In: International Conference on Machine Learning, pp. 4904\u20134916. PMLR (2021)"},{"key":"2_CR32","doi-asserted-by":"crossref","unstructured":"Jiang, S., Hondelink, L., Suriawinata, A.A., Hassanpour, S.: Masked pre-training of transformers for histology image analysis. arXiv preprint arXiv:2304.07434 (2023)","DOI":"10.1016\/j.jpi.2024.100386"},{"key":"2_CR33","doi-asserted-by":"crossref","unstructured":"Kang, M., Song, H., Park, S., Yoo, D., Pereira, S.: Benchmarking self-supervised learning on diverse pathology datasets. In: 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3344\u20133354 (2023)","DOI":"10.1109\/CVPR52729.2023.00326"},{"issue":"8","key":"2_CR34","doi-asserted-by":"publisher","first-page":"789","DOI":"10.1038\/s43018-020-0087-6","volume":"1","author":"JN Kather","year":"2020","unstructured":"Kather, J.N., et al.: Pan-cancer image-based detection of clinically actionable genetic alterations. Nat. Cancer 1(8), 789\u2013799 (2020)","journal-title":"Nat. Cancer"},{"key":"2_CR35","doi-asserted-by":"publisher","first-page":"164","DOI":"10.1016\/j.compbiomed.2019.05.020","volume":"110","author":"FD Khameneh","year":"2019","unstructured":"Khameneh, F.D., Razavi, S., Kamasak, M.: Automated segmentation of cell membranes to evaluate HER2 status in whole slide images using a modified deep learning network. Comput. Biol. Med. 110, 164\u2013174 (2019)","journal-title":"Comput. Biol. Med."},{"key":"2_CR36","unstructured":"Kim, M.: Differentiable expectation-maximization for set representation learning. In: International Conference on Learning Representations (2022)"},{"key":"2_CR37","doi-asserted-by":"crossref","unstructured":"Koohbanani, N.A., Unnikrishnan, B., Khurram, S.A., Krishnaswamy, P., Rajpoot, N.: Self-path: self-supervision for classification of pathology images with limited annotations. IEEE Trans. Med. Imaging (2021)","DOI":"10.1109\/TMI.2021.3056023"},{"key":"2_CR38","doi-asserted-by":"crossref","unstructured":"Krishnan, R., Rajpurkar, P., Topol, E.J.: Self-supervised learning in medicine and healthcare. Nat. Biomed. Eng. (2022)","DOI":"10.1038\/s41551-022-00914-1"},{"key":"2_CR39","doi-asserted-by":"crossref","unstructured":"Lazard, T., Lerousseau, M., Decenci\u00e8re, E., Walter, T.: Giga-SSL: self-supervised learning for gigapixel images. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4304\u20134313 (2023)","DOI":"10.1109\/CVPRW59228.2023.00453"},{"key":"2_CR40","doi-asserted-by":"crossref","unstructured":"Lee, Y., Park, J., Oh, S., et al.: Derivation of prognostic contextual histopathological features from whole-slide images of tumours via graph deep learning. Nat. Biomed. Eng. (2022)","DOI":"10.1038\/s41551-022-00923-0"},{"key":"2_CR41","doi-asserted-by":"crossref","unstructured":"Li, B., Li, Y., Eliceiri, K.W.: Dual-stream multiple instance learning network for whole slide image classification with self-supervised contrastive learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 14318\u201314328 (2021)","DOI":"10.1109\/CVPR46437.2021.01409"},{"key":"2_CR42","doi-asserted-by":"crossref","unstructured":"Li, H., et al.: Task-specific fine-tuning via variational information bottleneck for weakly-supervised pathology whole slide image classification. In: 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2023)","DOI":"10.1109\/CVPR52729.2023.00720"},{"key":"2_CR43","unstructured":"Li, J., Li, D., Savarese, S., Hoi, S.: BLIP-2: bootstrapping language-image pre-training with frozen image encoders and large language models. arXiv preprint arXiv:2301.12597 (2023)"},{"key":"2_CR44","first-page":"9694","volume":"34","author":"J Li","year":"2021","unstructured":"Li, J., Selvaraju, R., Gotmare, A., Joty, S., Xiong, C., Hoi, S.C.H.: Align before fuse: vision and language representation learning with momentum distillation. Adv. Neural. Inf. Process. Syst. 34, 9694\u20139705 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"2_CR45","doi-asserted-by":"crossref","unstructured":"Li, Y., Fan, H., Hu, R., Feichtenhofer, C., He, K.: Scaling language-image pre-training via masking. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 23390\u201323400 (2023)","DOI":"10.1109\/CVPR52729.2023.02240"},{"key":"2_CR46","unstructured":"Liang, P.P., et al.: High-modality multimodal transformer: quantifying modality & interaction heterogeneity for high-modality representation learning. Trans. Mach. Learn. Res. (2023)"},{"key":"2_CR47","doi-asserted-by":"crossref","unstructured":"Lu, M., et al.: Towards a visual-language foundation model for computational pathology. Nat. Med. (2024)","DOI":"10.1038\/s41591-024-02856-4"},{"issue":"6","key":"2_CR48","doi-asserted-by":"publisher","first-page":"555","DOI":"10.1038\/s41551-020-00682-w","volume":"5","author":"MY Lu","year":"2021","unstructured":"Lu, M.Y., Williamson, D.F., Chen, T.Y., Chen, R.J., Barbieri, M., Mahmood, F.: Data-efficient and weakly supervised computational pathology on whole-slide images. Nat. Biomed. Eng. 5(6), 555\u2013570 (2021)","journal-title":"Nat. Biomed. Eng."},{"key":"2_CR49","doi-asserted-by":"crossref","unstructured":"Mukashyaka, P., Sheridan, T., Pour, A., Chuang, J.: Sampler: unsupervised representations for rapid analysis of whole slide tissue images. eBioMedicine 99, 104908 (2024)","DOI":"10.1016\/j.ebiom.2023.104908"},{"issue":"5727","key":"2_CR50","first-page":"1","volume":"11","author":"N Naik","year":"2020","unstructured":"Naik, N., et al.: Deep learning-enabled breast cancer hormonal receptor status determination from base-level H &E stains. Nat. Commun. 11(5727), 1\u20138 (2020)","journal-title":"Nat. Commun."},{"key":"2_CR51","unstructured":"Pramanick, S., et al.: VoLTA: vision-language transformer with weakly-supervised local-feature alignment. arXiv preprint arXiv:2210.04135 (2022)"},{"key":"2_CR52","unstructured":"Pramanick, S., et al.: VoLTA: Vision-language transformer with weakly-supervised local-feature alignment. TMLR (2023)"},{"key":"2_CR53","doi-asserted-by":"crossref","unstructured":"Qaiser, T., et al.: HER2 challenge contest: a detailed assessment of automated HER2 scoring algorithms in whole slide images of breast cancer tissues. Histopathology 72(2), 227\u2013238 (2018)","DOI":"10.1111\/his.13333"},{"key":"2_CR54","unstructured":"Radford, A., et\u00a0al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning, pp. 8748\u20138763. PMLR (2021)"},{"issue":"7275","key":"2_CR55","first-page":"1","volume":"10","author":"RR Rawat","year":"2020","unstructured":"Rawat, R.R., et al.: Deep learned tissue fingerprints classify breast cancers by ER\/PR\/Her2 status from H &E images. Sci. Rep. 10(7275), 1\u201313 (2020)","journal-title":"Sci. Rep."},{"key":"2_CR56","unstructured":"Shaikovski, G., et al.: PRISM: a multi-modal generative foundation model for slide-level histopathology. arXiv (2024)"},{"issue":"6753","key":"2_CR57","first-page":"1","volume":"13","author":"G Shamai","year":"2022","unstructured":"Shamai, G., et al.: Deep learning-based image analysis predicts PD-L1 status from H &E-stained histopathology images in breast cancer. Nat. Commun. 13(6753), 1\u201313 (2022)","journal-title":"Nat. Commun."},{"key":"2_CR58","first-page":"2136","volume":"34","author":"Z Shao","year":"2021","unstructured":"Shao, Z., Bian, H., Chen, Y., Wang, Y., Zhang, J., Ji, X., et al.: TransMIL: transformer based correlated multiple instance learning for whole slide image classification. Adv. Neural. Inf. Process. Syst. 34, 2136\u20132147 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"2_CR59","doi-asserted-by":"crossref","unstructured":"Singh, A., et al.: FLAVA: a foundational language and vision alignment model. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 15638\u201315650 (2022)","DOI":"10.1109\/CVPR52688.2022.01519"},{"key":"2_CR60","doi-asserted-by":"crossref","unstructured":"Song, A.H., Chen, R.J., Ding, T., Williamson, D.F., Jaume, G., Mahmood, F.: Morphological prototyping for unsupervised slide representation learning in computational pathology. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (2024)","DOI":"10.1109\/CVPR52733.2024.01099"},{"key":"2_CR61","unstructured":"Song, A.H., Chen, R.J., Jaume, G., Vaidya, A.J., Baras, A., Mahmood, F.: Multimodal prototyping for cancer survival prediction. In: Forty-first International Conference on Machine Learning (2024)"},{"key":"2_CR62","doi-asserted-by":"crossref","unstructured":"Song, A.H., et al.: Artificial intelligence for digital and computational pathology. Nat. Rev. Bioeng. (2023)","DOI":"10.1038\/s44222-023-00096-8"},{"key":"2_CR63","doi-asserted-by":"crossref","unstructured":"Tavolara, T., Gurcan, M., Niazi, M.: Contrastive multiple instance learning: an unsupervised framework for learning slide-level representations of whole slide histopathology images without labels. Cancers 14, 5778 (2022)","DOI":"10.3390\/cancers14235778"},{"issue":"45938","key":"2_CR64","first-page":"1","volume":"7","author":"ME Vandenberghe","year":"2017","unstructured":"Vandenberghe, M.E., Scott, M.L.J., Scorer, P.W., S\u00f6derberg, M., Balcerzak, D., Barker, C.: Relevance of deep learning to facilitate the diagnosis of HER2 status in breast cancer. Sci. Rep. 7(45938), 1\u201311 (2017)","journal-title":"Sci. Rep."},{"key":"2_CR65","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Neural Information Processing Systems (NeurIPS) (2017)"},{"key":"2_CR66","unstructured":"Vorontsov, E., et al.: Virchow: a million-slide digital pathology foundation model (2023)"},{"key":"2_CR67","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2023.102743","volume":"85","author":"QD Vu","year":"2023","unstructured":"Vu, Q.D., Rajpoot, K., Raza, S.E.A., Rajpoot, N.: Handcrafted Histological Transformer (H2T): unsupervised representation of whole slide images. Med. Image Anal. 85, 102743 (2023)","journal-title":"Med. Image Anal."},{"issue":"1","key":"2_CR68","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s12859-023-05474-y","volume":"24","author":"J Wang","year":"2023","unstructured":"Wang, J., Zhu, X., Chen, K., Hao, L., Liu, Y.: HAHNet: a convolutional neural network for HER2 status classification of breast cancer. BMC Bioinf. 24(1), 1\u201316 (2023)","journal-title":"BMC Bioinf."},{"key":"2_CR69","doi-asserted-by":"crossref","unstructured":"Wang, W., et\u00a0al.: Image as a foreign language: BEiT pretraining for vision and vision-language tasks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 19175\u201319186 (2023)","DOI":"10.1109\/CVPR52729.2023.01838"},{"key":"2_CR70","doi-asserted-by":"publisher","unstructured":"Wang, X., Yang, S., Zhang, J., Wang, M., Zhang, J., Huang, J., Yang, W., Han, X.: TransPath: transformer-based self-supervised learning for histopathological image classification. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. pp. 186\u2013195. Springer (2021). https:\/\/doi.org\/10.1007\/978-3-030-87237-3_18","DOI":"10.1007\/978-3-030-87237-3_18"},{"key":"2_CR71","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2022.102559","volume":"81","author":"X Wang","year":"2022","unstructured":"Wang, X., et al.: Transformer-based unsupervised contrastive learning for histopathological image classification. Med. Image Anal. 81, 102559 (2022)","journal-title":"Med. Image Anal."},{"key":"2_CR72","doi-asserted-by":"publisher","first-page":"562","DOI":"10.1038\/s41597-023-02422-6","volume":"10","author":"P Weitz","year":"2023","unstructured":"Weitz, P., Valkonen, M., Solorzano, L., et al.: A multi-stain breast cancer histological whole-slide-image data set from routine diagnostics. Sci. Data 10, 562 (2023)","journal-title":"Sci. Data"},{"issue":"1","key":"2_CR73","doi-asserted-by":"publisher","first-page":"562","DOI":"10.1038\/s41597-023-02422-6","volume":"10","author":"P Weitz","year":"2023","unstructured":"Weitz, P., et al.: A multi-stain breast cancer histological whole-slide-image data set from routine diagnostics. Sci. Data 10(1), 562 (2023)","journal-title":"Sci. Data"},{"key":"2_CR74","unstructured":"Xiang, J., Zhang, J.: Exploring low-rank property in multiple instance learning for whole slide image classification. In: The Eleventh International Conference on Learning Representations (2022)"},{"key":"2_CR75","unstructured":"Xiang, J., Zhang, J.: Exploring low-rank property in multiple instance learning for whole slide image classification. In: The Eleventh International Conference on Learning Representations (2023)"},{"key":"2_CR76","doi-asserted-by":"publisher","DOI":"10.3389\/fonc.2021.759007","volume":"11","author":"F Xu","year":"2021","unstructured":"Xu, F., et al.: Predicting axillary lymph node metastasis in early breast cancer using deep learning on primary tumor biopsy slides. Front. Oncol. 11, 759007 (2021)","journal-title":"Front. Oncol."},{"key":"2_CR77","doi-asserted-by":"crossref","unstructured":"Xu, H., et al.: A whole-slide foundation model for digital pathology from real-world data. Nature (2024)","DOI":"10.1038\/s41586-024-07441-w"},{"key":"2_CR78","unstructured":"Yu, J., Wang, Z., Vasudevan, V., Yeung, L., Seyedhosseini, M., Wu, Y.: CoCa: contrastive captioners are image-text foundation models. Trans. Mach. Learn. Res. (2022)"},{"key":"2_CR79","doi-asserted-by":"publisher","unstructured":"Yu, Z., Lin, T., Xu, Y.: SLPD: slide-level prototypical distillation for WSIs. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 259\u2013269. Springer (2023). https:\/\/doi.org\/10.1007\/978-3-031-43907-0_25","DOI":"10.1007\/978-3-031-43907-0_25"},{"key":"2_CR80","unstructured":"Zhou, J., et al.: Image BERT pre-training with online tokenizer. In: International Conference on Learning Representations (2022)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-73414-4_2","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,24]],"date-time":"2024-10-24T17:04:10Z","timestamp":1729789450000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-73414-4_2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,25]]},"ISBN":["9783031734137","9783031734144"],"references-count":80,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-73414-4_2","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,10,25]]},"assertion":[{"value":"25 October 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}