{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T14:44:49Z","timestamp":1742913889298,"version":"3.40.3"},"publisher-location":"Cham","reference-count":25,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031442155"},{"type":"electronic","value":"9783031442162"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-44216-2_17","type":"book-chapter","created":{"date-parts":[[2023,9,21]],"date-time":"2023-09-21T07:02:58Z","timestamp":1695279778000},"page":"203-215","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Transformer Based Prototype Learning for\u00a0Weakly-Supervised Histopathology Tissue Semantic Segmentation"],"prefix":"10.1007","author":[{"given":"Jinwen","family":"She","sequence":"first","affiliation":[]},{"given":"Yanxu","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Andy J.","family":"Ma","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,9,22]]},"reference":[{"key":"17_CR1","doi-asserted-by":"crossref","unstructured":"Chan, L., Hosseini, M.S., Rowsell, C., Plataniotis, K.N., Damaskinos, S.: HistoSegNet: semantic segmentation of histological tissue type in whole slide images. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.01076"},{"key":"17_CR2","doi-asserted-by":"crossref","unstructured":"Chan, P.P., Chen, K., Xu, L., Hu, X., Yeung, D.S.: Weakly supervised semantic segmentation with patch-based metric learning enhancement. In: ICANN (2021)","DOI":"10.1007\/978-3-030-86365-4_38"},{"key":"17_CR3","doi-asserted-by":"crossref","unstructured":"Chang, Y.T., Wang, Q., Hung, W.C., Piramuthu, R., Tsai, Y.H., Yang, M.H.: Weakly-supervised semantic segmentation via sub-category exploration. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00901"},{"key":"17_CR4","doi-asserted-by":"crossref","unstructured":"Chattopadhay, A., Sarkar, A., Howlader, P., Balasubramanian, V.N.: Grad-CAM++: generalized gradient-based visual explanations for deep convolutional networks. In: WACV (2018)","DOI":"10.1109\/WACV.2018.00097"},{"key":"17_CR5","unstructured":"Chen, J., et al.: Transunet: transformers make strong encoders for medical image segmentation. arXiv preprint arXiv:2102.04306 (2021)"},{"key":"17_CR6","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2022.102615","volume":"82","author":"J Chen","year":"2022","unstructured":"Chen, J., Frey, E.C., He, Y., Segars, W.P., Li, Y., Du, Y.: TransMorph: transformer for unsupervised medical image registration. Med. Image Anal. 82, 102615 (2022)","journal-title":"Med. Image Anal."},{"key":"17_CR7","unstructured":"Dosovitskiy, A., et al.: An image is worth $$16 \\times 16$$ words: transformers for image recognition at scale. ICLR (2021)"},{"key":"17_CR8","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2022.102487","volume":"80","author":"C Han","year":"2022","unstructured":"Han, C., et al.: Multi-layer pseudo-supervision for histopathology tissue semantic segmentation using patch-level classification labels. Med. Image Anal. 80, 102487 (2022)","journal-title":"Med. Image Anal."},{"key":"17_CR9","doi-asserted-by":"crossref","unstructured":"Hanahan, D., Weinberg, R.A.: Hallmarks of cancer: the next generation. Cell 144(5), 646\u2013674 (2011)","DOI":"10.1016\/j.cell.2011.02.013"},{"key":"17_CR10","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.90"},{"issue":"11","key":"17_CR11","first-page":"2376","volume":"36","author":"Z Jia","year":"2017","unstructured":"Jia, Z., Huang, X., Eric, I., Chang, C., Xu, Y.: Constrained deep weak supervision for histopathology image segmentation. IEEE TMI 36(11), 2376\u20132388 (2017)","journal-title":"IEEE TMI"},{"key":"17_CR12","doi-asserted-by":"crossref","unstructured":"Jiang, P.T., Hou, Q., Cao, Y., Cheng, M.M., Wei, Y., Xiong, H.K.: Integral object mining via online attention accumulation. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00216"},{"key":"17_CR13","doi-asserted-by":"crossref","unstructured":"Kweon, H., Yoon, S.H., Kim, H., Park, D., Yoon, K.J.: Unlocking the potential of ordinary classifier: class-specific adversarial erasing framework for weakly supervised semantic segmentation. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00691"},{"key":"17_CR14","doi-asserted-by":"publisher","unstructured":"Li, Y., Yu, Y., Zou, Y., Xiang, T., Li, X.: Online easy example mining for weakly-supervised gland segmentation from histology images. In: MICCAI (2022). https:\/\/doi.org\/10.1007\/978-3-031-16440-8_55","DOI":"10.1007\/978-3-031-16440-8_55"},{"key":"17_CR15","doi-asserted-by":"publisher","unstructured":"Li, Z., et al.: TFCNs: a CNN-transformer hybrid network for medical image segmentation. In: Pimenidis, E., Angelov, P., Jayne, C., Papaleonidas, A., Aydin, M. (eds.) Artificial Neural Networks and Machine Learning \u2013 ICANN 2022. ICANN 2022. LNCS, vol. 13532, pp. 781\u2013792. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-15937-4_65","DOI":"10.1007\/978-3-031-15937-4_65"},{"key":"17_CR16","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.media.2019.03.014","volume":"55","author":"T Qaiser","year":"2019","unstructured":"Qaiser, T., et al.: Fast and accurate tumor segmentation of histology images using persistent homology and deep convolutional features. Med. Image Anal. 55, 1\u201314 (2019)","journal-title":"Med. Image Anal."},{"key":"17_CR17","doi-asserted-by":"publisher","unstructured":"Qian, Z., et al.: Transformer based multiple instance learning for weakly supervised histopathology image segmentation. In: Wang, L., Dou, Q., Fletcher, P.T., Speidel, S., Li, S. (eds.) Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2022. MICCAI 2022. LNCS, vol. 13432, pp. 160\u2013170. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-16434-7_16","DOI":"10.1007\/978-3-031-16434-7_16"},{"key":"17_CR18","unstructured":"Touvron, H., Cord, M., Douze, M., Massa, F., Sablayrolles, A., J\u00e9gou, H.: Training data-efficient image transformers & distillation through attention. In: ICML (2021)"},{"key":"17_CR19","doi-asserted-by":"crossref","unstructured":"Van Rijthoven, M., Balkenhol, M., Silina, K., Van Der Laak, J., Ciompi, F.: HookNet: multi-resolution convolutional neural networks for semantic segmentation in histopathology whole-slide images. Med. Image Anal. 68, 101890 (2021)","DOI":"10.1016\/j.media.2020.101890"},{"key":"17_CR20","unstructured":"Vaswani, A., et al.: Attention is all you need. NeurIPS (2017)"},{"key":"17_CR21","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2022.102559","volume":"81","author":"X Wang","year":"2022","unstructured":"Wang, X., et al.: Transformer-based unsupervised contrastive learning for histopathological image classification. Med. Image Anal. 81, 102559 (2022)","journal-title":"Med. Image Anal."},{"key":"17_CR22","doi-asserted-by":"crossref","unstructured":"Wang, Y., Zhang, J., Kan, M., Shan, S., Chen, X.: Self-supervised equivariant attention mechanism for weakly supervised semantic segmentation. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.01229"},{"key":"17_CR23","doi-asserted-by":"crossref","unstructured":"Xu, G., et al.: Camel: a weakly supervised learning framework for histopathology image segmentation. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.01078"},{"key":"17_CR24","doi-asserted-by":"crossref","unstructured":"Xu, L., Ouyang, W., Bennamoun, M., Boussaid, F., Xu, D.: Multi-class token transformer for weakly supervised semantic segmentation. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.00427"},{"key":"17_CR25","doi-asserted-by":"crossref","unstructured":"Zhou, B., Khosla, A., Lapedriza, A., Oliva, A., Torralba, A.: Learning deep features for discriminative localization. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.319"}],"container-title":["Lecture Notes in Computer Science","Artificial Neural Networks and Machine Learning \u2013 ICANN 2023"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-44216-2_17","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,9,21]],"date-time":"2023-09-21T07:05:58Z","timestamp":1695279958000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-44216-2_17"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031442155","9783031442162"],"references-count":25,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-44216-2_17","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"22 September 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICANN","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Artificial Neural Networks","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Heraklion","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Greece","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26 September 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"32","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icann2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/e-nns.org\/icann2023\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"easyacademia.org","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"947","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"426","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"22","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"45% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.4","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"type of other papers accepted  : 9 Abstract","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}