{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,31]],"date-time":"2026-01-31T08:14:52Z","timestamp":1769847292660,"version":"3.49.0"},"publisher-location":"Cham","reference-count":39,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031164330","type":"print"},{"value":"9783031164347","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-16434-7_40","type":"book-chapter","created":{"date-parts":[[2022,9,15]],"date-time":"2022-09-15T15:03:08Z","timestamp":1663254188000},"page":"409-419","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":15,"title":["Fast FF-to-FFPE Whole Slide Image Translation via\u00a0Laplacian Pyramid and\u00a0Contrastive Learning"],"prefix":"10.1007","author":[{"given":"Lei","family":"Fan","sequence":"first","affiliation":[]},{"given":"Arcot","family":"Sowmya","sequence":"additional","affiliation":[]},{"given":"Erik","family":"Meijering","sequence":"additional","affiliation":[]},{"given":"Yang","family":"Song","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,9,16]]},"reference":[{"issue":"12","key":"40_CR1","doi-asserted-by":"publisher","first-page":"2481","DOI":"10.1109\/TPAMI.2016.2644615","volume":"39","author":"V Badrinarayanan","year":"2017","unstructured":"Badrinarayanan, V., Kendall, A., Cipolla, R.: SegNet: a deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 39(12), 2481\u20132495 (2017)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"40_CR2","unstructured":"Brown, R.W.: Histologic Preparations: Common Problems and their Solutions. College of American Pathologists (2009)"},{"issue":"4","key":"40_CR3","doi-asserted-by":"publisher","first-page":"532","DOI":"10.1109\/TCOM.1983.1095851","volume":"3","author":"PJ Burt","year":"1983","unstructured":"Burt, P.J., Adelson, E.H.: The Laplacian pyramid as a compact image code. IEEE Trans. Commun. 3(4), 532\u2013540 (1983)","journal-title":"IEEE Trans. Commun."},{"key":"40_CR4","unstructured":"Chen, T., Kornblith, S., Norouzi, M., Hinton, G.: A simple framework for contrastive learning of visual representations. In: International Conference on Machine Learning (ICML), pp. 1597\u20131607. PMLR (2020)"},{"key":"40_CR5","doi-asserted-by":"crossref","unstructured":"Chen, X., He, K.: Exploring simple siamese representation learning. In: CVPR, pp. 15750\u201315758 (2021)","DOI":"10.1109\/CVPR46437.2021.01549"},{"key":"40_CR6","doi-asserted-by":"publisher","unstructured":"Cong, C., et al.: Semi-supervised adversarial learning for stain normalisation in histopathology images. In: de Bruijne, M., et al. (eds.) MICCAI 2021. LNCS, vol. 12908, pp. 581\u2013591. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-87237-3_56","DOI":"10.1007\/978-3-030-87237-3_56"},{"key":"40_CR7","unstructured":"Denton, E., Chintala, S., Szlam, A., Fergus, R.: Deep generative image models using a laplacian pyramid of adversarial networks. In: NeurIPS, vol. 28, pp. 1486\u20131494 (2015)"},{"issue":"5","key":"40_CR8","doi-asserted-by":"publisher","first-page":"554","DOI":"10.1038\/s41374-021-00718-y","volume":"102","author":"K Falahkheirkhah","year":"2022","unstructured":"Falahkheirkhah, K., Guo, T., Hwang, M., et al.: A generative adversarial approach to facilitate archival-quality histopathologic diagnoses from frozen tissue sections. Lab. Investig. 102(5), 554\u2013559 (2022)","journal-title":"Lab. Investig."},{"key":"40_CR9","doi-asserted-by":"publisher","unstructured":"Fan, L., Sowmya, A., Meijering, E., Song, Y.: Learning visual features by colorization for slide-consistent survival prediction from whole slide images. In: de Bruijne, M., et al. (eds.) MICCAI 2021. LNCS, vol. 12908, pp. 592\u2013601. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-87237-3_57","DOI":"10.1007\/978-3-030-87237-3_57"},{"key":"40_CR10","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., et al.: Generative adversarial nets. In: NeurIPS, vol. 27 (2014)"},{"key":"40_CR11","doi-asserted-by":"crossref","unstructured":"He, K., Fan, H., Wu, Y., Xie, S., Girshick, R.: Momentum contrast for unsupervised visual representation learning. In: CVPR, pp. 9729\u20139738 (2020)","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"40_CR12","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"40_CR13","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: GANs trained by a two time-scale update rule converge to a local nash equilibrium. In: NeurIPS, vol. 30 (2017)"},{"key":"40_CR14","doi-asserted-by":"crossref","unstructured":"Huang, X., Belongie, S.: Arbitrary style transfer in real-time with adaptive instance normalization. In: ICCV, pp. 1501\u20131510 (2017)","DOI":"10.1109\/ICCV.2017.167"},{"issue":"1","key":"40_CR15","first-page":"4","volume":"13","author":"H Jaafar","year":"2006","unstructured":"Jaafar, H.: Intra-operative frozen section consultation: concepts, applications and limitations. Malaysian J. Med. Sci. MJMS 13(1), 4 (2006)","journal-title":"Malaysian J. Med. Sci. MJMS"},{"issue":"7471","key":"40_CR16","doi-asserted-by":"publisher","first-page":"333","DOI":"10.1038\/nature12634","volume":"502","author":"C Kandoth","year":"2013","unstructured":"Kandoth, C., McLellan, M.D., Vandin, F., et al.: Mutational landscape and significance across 12 major cancer types. Nature 502(7471), 333\u2013339 (2013)","journal-title":"Nature"},{"issue":"7","key":"40_CR17","doi-asserted-by":"publisher","first-page":"1054","DOI":"10.1038\/s41591-019-0462-y","volume":"25","author":"JN Kather","year":"2019","unstructured":"Kather, J.N., Pearson, A.T., Halama, N., et al.: Deep learning can predict microsatellite instability directly from histology in gastrointestinal cancer. Nat. Med. 25(7), 1054\u20131056 (2019)","journal-title":"Nat. Med."},{"key":"40_CR18","unstructured":"Khosla, P., Teterwak, P., Wang, C., et al.: Supervised contrastive learning. In: NeurIPS, vol. 33, pp. 18661\u201318673 (2020)"},{"key":"40_CR19","unstructured":"Kingma, D.P., Ba, J.: Adam: A method for stochastic optimization. In: International Conference for Learning Representations (ICLR) (2015)"},{"key":"40_CR20","doi-asserted-by":"crossref","unstructured":"Lai, W.S., Huang, J.B., Ahuja, N., Yang, M.H.: Deep Laplacian pyramid networks for fast and accurate super-resolution. In: CVPR, pp. 624\u2013632 (2017)","DOI":"10.1109\/CVPR.2017.618"},{"key":"40_CR21","doi-asserted-by":"crossref","unstructured":"Liang, J., Zeng, H., Zhang, L.: High-resolution photorealistic image translation in real-time: a Laplacian pyramid translation network. In: CVPR, pp. 9392\u20139400 (2021)","DOI":"10.1109\/CVPR46437.2021.00927"},{"key":"40_CR22","doi-asserted-by":"publisher","DOI":"10.1109\/TKDE.2021.3090866","author":"X Liu","year":"2021","unstructured":"Liu, X., Zhang, F., Hou, Z., et al.: Self-supervised learning: generative or contrastive. IEEE Trans. Knowl. Data Eng. (2021). https:\/\/doi.org\/10.1109\/TKDE.2021.3090866","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"40_CR23","doi-asserted-by":"crossref","unstructured":"Macenko, M., Niethammer, M., Marron, J.S., et al.: A method for normalizing histology slides for quantitative analysis. In: IEEE International Symposium on Biomedical Imaging: From Nano to Macro (ISBI), pp. 1107\u20131110. IEEE (2009)","DOI":"10.1109\/ISBI.2009.5193250"},{"key":"40_CR24","doi-asserted-by":"crossref","unstructured":"Mao, X., Li, Q., Xie, H., et al.: Least squares generative adversarial networks. In: ICCV, pp. 2794\u20132802 (2017)","DOI":"10.1109\/ICCV.2017.304"},{"key":"40_CR25","unstructured":"Ozyoruk, K.B., Can, S., Gokceler, G.I., et al.: Deep learning-based frozen section to FFPE translation. arXiv preprint arXiv:2107.11786 (2021)"},{"key":"40_CR26","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2021.3109419","author":"Y Pang","year":"2021","unstructured":"Pang, Y., Lin, J., Qin, T., Chen, Z.: Image-to-image translation: methods and applications. IEEE Trans. Multim. (2021). https:\/\/doi.org\/10.1109\/TMM.2021.3109419","journal-title":"IEEE Trans. Multim."},{"key":"40_CR27","doi-asserted-by":"publisher","unstructured":"Park, T., Efros, A.A., Zhang, R., Zhu, J.-Y.: Contrastive learning for unpaired image-to-image translation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12354, pp. 319\u2013345. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58545-7_19","DOI":"10.1007\/978-3-030-58545-7_19"},{"key":"40_CR28","unstructured":"Paszke, A., Gross, S., Massa, F., et al.: PyTorch: an imperative style, high-performance deep learning library. In: NeurIPS, pp. 8024\u20138035 (2019)"},{"key":"40_CR29","unstructured":"Rolls, G.O., Farmer, N.J., Hall, J.B.: Artifacts in histological and cytological preparations. Leica Microsystems (2008)"},{"key":"40_CR30","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., et al.: MobileNetv2: Inverted residuals and linear bottlenecks. In: CVPR, pp. 4510\u20134520 (2018)","DOI":"10.1109\/CVPR.2018.00474"},{"key":"40_CR31","doi-asserted-by":"crossref","unstructured":"Shaban, M.T., Baur, C., Navab, N., Albarqouni, S.: StainGAN: stain style transfer for digital histological images. In: IEEE International Symposium on Biomedical Imaging: From Nano to Macro (ISBI), pp. 953\u2013956. IEEE (2019)","DOI":"10.1109\/ISBI.2019.8759152"},{"issue":"2","key":"40_CR32","doi-asserted-by":"publisher","first-page":"279","DOI":"10.4103\/jomfp.JOMFP_125_15","volume":"22","author":"SA Taqi","year":"2018","unstructured":"Taqi, S.A., Sami, S.A., Sami, L.B., Zaki, S.A.: A review of artifacts in histopathology. J. Oral Maxillof. Pathol. 22(2), 279 (2018)","journal-title":"J. Oral Maxillof. Pathol."},{"key":"40_CR33","doi-asserted-by":"crossref","unstructured":"Wang, T.C., Liu, M.Y., Zhu, J.Y., et al.: High-resolution image synthesis and semantic manipulation with conditional GANs. In: CVPR, pp. 8798\u20138807 (2018)","DOI":"10.1109\/CVPR.2018.00917"},{"key":"40_CR34","doi-asserted-by":"publisher","unstructured":"Wolterink, J.M., Dinkla, A.M., Savenije, M.H.F., Seevinck, P.R., van den Berg, C.A.T., I\u0161gum, I.: Deep MR to CT synthesis using unpaired data. In: Tsaftaris, S.A., Gooya, A., Frangi, A.F., Prince, J.L. (eds.) SASHIMI 2017. LNCS, vol. 10557, pp. 14\u201323. Springer, Cham (2017). https:\/\/doi.org\/10.1007\/978-3-319-68127-6_2","DOI":"10.1007\/978-3-319-68127-6_2"},{"key":"40_CR35","doi-asserted-by":"publisher","unstructured":"Yang, H., et al.: Unpaired brain MR-to-CT synthesis using a structure-constrained CycleGAN. In: Stoyanov, D., et al. (eds.) DLMIA\/ML-CDS -2018. LNCS, vol. 11045, pp. 174\u2013182. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-00889-5_20","DOI":"10.1007\/978-3-030-00889-5_20"},{"key":"40_CR36","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2019.101552","volume":"58","author":"X Yi","year":"2019","unstructured":"Yi, X., Walia, E., Babyn, P.: Generative adversarial network in medical imaging: a review. Med. Image Anal. 58, 101552 (2019)","journal-title":"Med. Image Anal."},{"key":"40_CR37","unstructured":"Zhang, H., Cisse, M., Dauphin, Y.N., Lopez-Paz, D.: Mixup: beyond empirical risk minimization. In: International Conference for Learning Representations (ICLR) (2018)"},{"issue":"6","key":"40_CR38","doi-asserted-by":"publisher","first-page":"1856","DOI":"10.1109\/TMI.2019.2959609","volume":"39","author":"Z Zhou","year":"2020","unstructured":"Zhou, Z., Siddiquee, M.M.R., Tajbakhsh, N., Liang, J.: UNet++: redesigning skip connections to exploit multiscale features in image segmentation. IEEE Trans. Med. Imag. 39(6), 1856\u20131867 (2020)","journal-title":"IEEE Trans. Med. Imag."},{"key":"40_CR39","doi-asserted-by":"crossref","unstructured":"Zhu, J.Y., Park, T., Isola, P., Efros, A.A.: Unpaired Image-to-Image translation using cycle-consistent adversarial networks. In: ICCV, pp. 2223\u20132232 (2017)","DOI":"10.1109\/ICCV.2017.244"}],"container-title":["Lecture Notes in Computer Science","Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-16434-7_40","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,13]],"date-time":"2024-03-13T11:46:29Z","timestamp":1710330389000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-16434-7_40"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031164330","9783031164347"],"references-count":39,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-16434-7_40","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"16 September 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MICCAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Medical Image Computing and Computer-Assisted Intervention","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Singapore","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Singapore","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 September 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 September 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"25","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"miccai2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft Conference","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1831","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"574","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"31% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}