{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,15]],"date-time":"2026-04-15T20:17:30Z","timestamp":1776284250987,"version":"3.50.1"},"reference-count":31,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,7,3]],"date-time":"2025-07-03T00:00:00Z","timestamp":1751500800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,7,3]],"date-time":"2025-07-03T00:00:00Z","timestamp":1751500800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/100009647","name":"Ministry of Health","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100009647","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,7,3]]},"DOI":"10.1109\/icecet63943.2025.11472161","type":"proceedings-article","created":{"date-parts":[[2026,4,9]],"date-time":"2026-04-09T19:42:35Z","timestamp":1775763755000},"page":"1-6","source":"Crossref","is-referenced-by-count":0,"title":["Self-Supervised Pathology Foundation Models based Feature Encoder and Multiple Instance Learning for Breast and Skin Cancer Diagnosis"],"prefix":"10.1109","author":[{"given":"Umm E","family":"Farwa","sequence":"first","affiliation":[{"name":"Inje University,Institute of Digital Anti-Aging Healthcare,Gimhae,Korea,50834"}]},{"given":"Sikandar","family":"Ali","sequence":"additional","affiliation":[{"name":"Inje University,Institute of Digital Anti-Aging Healthcare,Gimhae,Korea,50834"}]},{"given":"Hee Cheol","family":"Kim","sequence":"additional","affiliation":[{"name":"u-AHRC Inje University,College of AI Convergence, Institute of Digital Anti-Aging Healthcare,Gimhae,Korea,50834"}]}],"member":"263","reference":[{"key":"ref1","article-title":"Foundation Models in Computational Pathology: A Review of Challenges, Opportunities, and Impact","author":"Bilal","year":"2025"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.31662\/jmaj.2024-0206"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/RBME.2025.3531360"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1186\/s13040-024-00414-9"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.7759\/cureus.69818"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.15212\/RADSCI-2023-0018"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1016\/j.labinv.2023.100255"},{"key":"ref8","article-title":"Virchow2: Scaling self-supervised mixed magnification models in pathology","author":"Zimmermann","year":"2024"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1148\/radiol.240597"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/JBHI.2022.3181823"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1038\/s41746-023-00879-8"},{"key":"ref12","article-title":"Review on multi-modal AI models to integrate imaging and omics data","author":"Vincenzo","year":"2024"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1016\/j.compbiomed.2023.107356"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1016\/j.mcpdig.2024.02.003"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1001\/jama.2022.6421"},{"issue":"1","key":"ref16","first-page":"471","article-title":"Foundation models in medical imaging: Revolutionizing diagnostic accuracy and efficiency","volume":"4","author":"Prabhod","year":"2024","journal-title":"Journal of Artificial Intelligence Research and Applications"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.54660\/.IJMRGE.2023.4.6.1141-1154"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1038\/s41591-024-02857-3"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1038\/s41591-024-02856-4"},{"key":"ref20","article-title":"EXAONEPath 1.0 Patch-level Foundation Model for Pathology","author":"Yun","year":"2024"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00326"},{"key":"ref22","article-title":"PRISM: A Multimodal Generative Foundation Model for Slide-Level Histopathology","author":"Shaikovski","year":"2024"},{"key":"ref23","article-title":"Virchow: A Million-Slide Digital Pathology Foundation Model","author":"Vorontsov","year":"2023"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1038\/s41746-023-00811-0"},{"key":"ref25","article-title":"Learning transferable visual models from natural language supervision","volume-title":"Int. Conf. Mach. Learn. (ICML)","author":"Radford"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.18653\/vl\/N19-142"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1007\/s12530-024-09581-w"},{"key":"ref28","article-title":"A simple framework for contrastive learning of visual representations","volume-title":"Int. Conf. Mach. Learn. (ICML)","author":"Chen"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00950"},{"key":"ref30","article-title":"DINOv2: Learning robust visual features without supervision","author":"Oquab","year":"2023"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1038\/sdata.2018.161"}],"event":{"name":"2025 5th International Conference on Electrical, Computer and Energy Technologies (ICECET)","location":"Paris, France","start":{"date-parts":[[2025,7,3]]},"end":{"date-parts":[[2025,7,6]]}},"container-title":["2025 5th International Conference on Electrical, Computer and Energy Technologies (ICECET)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11471878\/11471697\/11472161.pdf?arnumber=11472161","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,4,15]],"date-time":"2026-04-15T19:24:04Z","timestamp":1776281044000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11472161\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,3]]},"references-count":31,"URL":"https:\/\/doi.org\/10.1109\/icecet63943.2025.11472161","relation":{},"subject":[],"published":{"date-parts":[[2025,7,3]]}}}