{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,27]],"date-time":"2026-01-27T01:12:12Z","timestamp":1769476332447,"version":"3.49.0"},"reference-count":29,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,12,18]],"date-time":"2025-12-18T00:00:00Z","timestamp":1766016000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,12,18]],"date-time":"2025-12-18T00:00:00Z","timestamp":1766016000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62173279"],"award-info":[{"award-number":["62173279"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Sichuan Science and Technology Program","award":["2021JDJQ0012"],"award-info":[{"award-number":["2021JDJQ0012"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Machine Vision and Applications"],"published-print":{"date-parts":[[2026,1]]},"DOI":"10.1007\/s00138-025-01775-9","type":"journal-article","created":{"date-parts":[[2025,12,18]],"date-time":"2025-12-18T13:18:00Z","timestamp":1766063880000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Balanced sampling-driven multi-modal active learning framework for breast cancer segmentation"],"prefix":"10.1007","volume":"37","author":[{"given":"Aisen","family":"Yang","sequence":"first","affiliation":[]},{"given":"Jun","family":"Li","sequence":"additional","affiliation":[]},{"given":"Na","family":"Qin","sequence":"additional","affiliation":[]},{"given":"Deqing","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Xuhui","family":"Song","sequence":"additional","affiliation":[]},{"given":"Jian","family":"Shu","sequence":"additional","affiliation":[]},{"given":"Lulu","family":"Xu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,12,18]]},"reference":[{"issue":"9","key":"1775_CR1","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3472291","volume":"54","author":"P Ren","year":"2021","unstructured":"Ren, P., Xiao, Y., Chang, X., Huang, P.-Y., Li, Z., Gupta, B.B., Chen, X., Wang, X.: A survey of deep active learning. ACM comput. surv. (CSUR) 54(9), 1\u201340 (2021)","journal-title":"ACM comput. surv. (CSUR)"},{"key":"1775_CR2","doi-asserted-by":"crossref","unstructured":"Beluch, W.H., Genewein, T., N\u00fcrnberger, A., K\u00f6hler, J.M.: The power of ensembles for active learning in image classification. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 9368\u20139377 (2018)","DOI":"10.1109\/CVPR.2018.00976"},{"key":"1775_CR3","doi-asserted-by":"crossref","unstructured":"Smailagic, A., Noh, H.Y., Costa, P., Walawalkar, D., Khandelwal, K., Mirshekari, M., Fagert, J., Galdr\u00e1n, A., Xu, S.: Medal: Deep active learning sampling method for medical image analysis. arXiv preprint arXiv:1809.09287 (2018)","DOI":"10.1109\/ICMLA.2018.00078"},{"issue":"29","key":"1775_CR4","doi-asserted-by":"publisher","first-page":"20483","DOI":"10.1007\/s11042-019-07988-1","volume":"79","author":"I Bakkouri","year":"2020","unstructured":"Bakkouri, I., Afdel, K.: Computer-aided diagnosis (cad) system based on multi-layer feature fusion network for skin lesion recognition in dermoscopy images. Multimed. Tools and Appl. 79(29), 20483\u201320518 (2020)","journal-title":"Multimed. Tools and Appl."},{"issue":"10","key":"1775_CR5","doi-asserted-by":"publisher","first-page":"12939","DOI":"10.1007\/s11042-018-6267-z","volume":"78","author":"I Bakkouri","year":"2019","unstructured":"Bakkouri, I., Afdel, K.: Multi-scale cnn based on region proposals for efficient breast abnormality recognition. Multimed. Tools and Appl. 78(10), 12939\u201312960 (2019)","journal-title":"Multimed. Tools and Appl."},{"key":"1775_CR6","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2021.102062","volume":"71","author":"S Budd","year":"2021","unstructured":"Budd, S., Robinson, E.C., Kainz, B.: A survey on active learning and human-in-the-loop deep learning for medical image analysis. Med. Image Anal. 71, 102062 (2021)","journal-title":"Med. Image Anal."},{"key":"1775_CR7","unstructured":"Sener, O., Savarese, S.: Active learning for convolutional neural networks: a core-set approach. arXiv preprint arXiv:1708.00489 (2017)"},{"key":"1775_CR8","doi-asserted-by":"crossref","unstructured":"Ozdemir, F., Peng, Z., Tanner, C., Fuernstahl, P., Goksel, O.: Active learning for segmentation by optimizing content information for maximal entropy. In: Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support: 4th International Workshop, DLMIA 2018, and 8th International Workshop, ML-CDS 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, September 20, 2018, Proceedings 4, pp. 183\u2013191 (2018). Springer","DOI":"10.1007\/978-3-030-00889-5_21"},{"key":"1775_CR9","doi-asserted-by":"crossref","unstructured":"Miao, J., Chen, C., Zhang, K., Chuai, J., Li, Q., Heng, P.-A.: Cross prompting consistency with segment anything model for semi-supervised medical image segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 167\u2013177 (2024). Springer","DOI":"10.1007\/978-3-031-72120-5_16"},{"key":"1775_CR10","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2023.102918","volume":"89","author":"MA Mazurowski","year":"2023","unstructured":"Mazurowski, M.A., Dong, H., Gu, H., Yang, J., Konz, N., Zhang, Y.: Segment anything model for medical image analysis: an experimental study. Med. Image Anal. 89, 102918 (2023)","journal-title":"Med. Image Anal."},{"issue":"4","key":"1775_CR11","doi-asserted-by":"publisher","first-page":"447","DOI":"10.1007\/s11633-022-1410-8","volume":"20","author":"X Wang","year":"2023","unstructured":"Wang, X., Chen, G., Qian, G., Gao, P., Wei, X.-Y., Wang, Y., Tian, Y., Gao, W.: Large-scale multi-modal pre-trained models: A comprehensive survey. Mach. Intel. Res. 20(4), 447\u2013482 (2023)","journal-title":"Mach. Intel. Res."},{"key":"1775_CR12","unstructured":"Awais, M., Naseer, M., Khan, S., Anwer, R.M., Cholakkal, H., Shah, M., Yang, M.-H., Khan, F.S.: Foundational models defining a new era in vision: a survey and outlook. arXiv preprint arXiv:2307.13721 (2023)"},{"key":"1775_CR13","doi-asserted-by":"publisher","DOI":"10.1016\/j.compbiomed.2024.108238","volume":"171","author":"Y Zhang","year":"2024","unstructured":"Zhang, Y., Shen, Z., Jiao, R.: Segment anything model for medical image segmentation: Current applications and future directions. Comput. Biol. Med. 171, 108238 (2024)","journal-title":"Comput. Biol. Med."},{"key":"1775_CR14","volume-title":"MLCA2F: multi-level context attentional feature fusion for COVID-19 lesion segmentation from CT scans","author":"I Bakkouri","year":"2022","unstructured":"Bakkouri, I., Afdel, K.: MLCA2F: multi-level context attentional feature fusion for COVID-19 lesion segmentation from CT scans. Signal, Image and Video Processing (2022)"},{"issue":"8","key":"1775_CR15","doi-asserted-by":"publisher","first-page":"10743","DOI":"10.1007\/s11042-022-12242-2","volume":"81","author":"I Bakkouri","year":"2022","unstructured":"Bakkouri, I., Afdel, K., Benois-Pineau, J., Initiative, G.C.F.T.A.D.N.: Bg-3dm2f: bidirectional gated 3d multi-scale feature fusion for alzheimer s disease diagnosis. Multimed. Tools and Appl. 81(8), 10743\u201310776 (2022)","journal-title":"Multimed. Tools and Appl."},{"key":"1775_CR16","doi-asserted-by":"crossref","unstructured":"Lahoud, F., S\u00fcsstrunk, S.: Zero-learning fast medical image fusion. In: 2019 22th International Conference on Information Fusion (FUSION), pp. 1\u20138 (2019). IEEE","DOI":"10.23919\/FUSION43075.2019.9011178"},{"key":"1775_CR17","doi-asserted-by":"publisher","first-page":"53","DOI":"10.1016\/0377-0427(87)90125-7","volume":"20","author":"PJ Rousseeuw","year":"1987","unstructured":"Rousseeuw, P.J.: Silhouettes: a graphical aid to the interpretation and validation of cluster analysis. J. Comput. Appl. Math. 20, 53\u201365 (1987)","journal-title":"J. Comput. Appl. Math."},{"issue":"2","key":"1775_CR18","doi-asserted-by":"publisher","first-page":"203","DOI":"10.1038\/s41592-020-01008-z","volume":"18","author":"F Isensee","year":"2021","unstructured":"Isensee, F., Jaeger, P.F., Kohl, S.A., Petersen, J., Maier-Hein, K.H.: nnu-net: a self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 18(2), 203\u2013211 (2021)","journal-title":"Nat. Methods"},{"key":"1775_CR19","unstructured":"Wu, J., Ji, W., Liu, Y., Fu, H., Xu, M., Xu, Y., Jin, Y.: Medical sam adapter: Adapting segment anything model for medical image segmentation. arXiv preprint arXiv:2304.12620 (2023)"},{"issue":"2","key":"1775_CR20","doi-asserted-by":"publisher","first-page":"526","DOI":"10.1016\/j.csda.2005.10.006","volume":"51","author":"F Leisch","year":"2006","unstructured":"Leisch, F.: A toolbox for k-centroids cluster analysis. Comput. Stat. Data Anal. 51(2), 526\u2013544 (2006)","journal-title":"Comput. Stat. Data Anal."},{"issue":"4","key":"1775_CR21","first-page":"589","volume":"6","author":"T Luo","year":"2005","unstructured":"Luo, T., Kramer, K., Goldgof, D.B., Hall, L.O., Samson, S., Remsen, A., Hopkins, T., Cohn, D.: Active learning to recognize multiple types of plankton. J. Mach. Learn. Res. 6(4), 589 (2005)","journal-title":"J. Mach. Learn. Res."},{"issue":"11","key":"1775_CR22","first-page":"2579","volume":"9","author":"L Maaten","year":"2008","unstructured":"Maaten, L., Hinton, G.: Visualizing data using t-sne. J. Mach. Learn. Res. 9(11), 2579 (2008)","journal-title":"J. Mach. Learn. Res."},{"key":"1775_CR23","unstructured":"Campbell, T., Broderick, T.: Bayesian coreset construction via greedy iterative geodesic ascent. In: International Conference on Machine Learning, pp. 698\u2013706 (2018). PMLR"},{"key":"1775_CR24","doi-asserted-by":"crossref","unstructured":"Zhou, Z., Rahman\u00a0Siddiquee, M.M., Tajbakhsh, N., Liang, J.: Unet++: A nested u-net architecture for medical image segmentation. In: International Workshop on Deep Learning in Medical Image Analysis, pp. 3\u201311 (2018). Springer","DOI":"10.1007\/978-3-030-00889-5_1"},{"key":"1775_CR25","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2024.103280","volume":"97","author":"J Chen","year":"2024","unstructured":"Chen, J., Mei, J., Li, X., Lu, Y., Yu, Q., Wei, Q., Luo, X., Xie, Y., Adeli, E., Wang, Y., et al.: Transunet: rethinking the u-net architecture design for medical image segmentation through the lens of transformers. Med. Image Anal. 97, 103280 (2024)","journal-title":"Med. Image Anal."},{"key":"1775_CR26","unstructured":"Kirsch, A., Van Amersfoort, J., Gal, Y.: Batchbald: Efficient and diverse batch acquisition for deep bayesian active learning. Advances in neural information processing systems 32 (2019)"},{"key":"1775_CR27","doi-asserted-by":"crossref","unstructured":"Dolz, J., Desrosiers, C., Ben Ayed, I.: Ivd-net: Intervertebral disc localization and segmentation in mri with a multi-modal unet. In: International Workshop and Challenge on Computational Methods and Clinical Applications for Spine Imaging, pp. 130\u2013143 (2018). Springer","DOI":"10.1007\/978-3-030-13736-6_11"},{"key":"1775_CR28","doi-asserted-by":"publisher","first-page":"74","DOI":"10.1016\/j.neunet.2019.08.025","volume":"121","author":"N Ibtehaz","year":"2020","unstructured":"Ibtehaz, N., Rahman, M.S.: Multiresunet: rethinking the u-net architecture for multimodal biomedical image segmentation. Neural Netw. 121, 74\u201387 (2020)","journal-title":"Neural Netw."},{"issue":"10","key":"1775_CR29","doi-asserted-by":"publisher","first-page":"4780","DOI":"10.1109\/JBHI.2023.3294278","volume":"27","author":"J Li","year":"2023","unstructured":"Li, J., Zhu, H., Chen, T., Qian, X.: Generalizable pancreas segmentation via a dual self-supervised learning framework. IEEE J. Biomed. Health Inform. 27(10), 4780 (2023)","journal-title":"IEEE J. Biomed. Health Inform."}],"container-title":["Machine Vision and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00138-025-01775-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00138-025-01775-9","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00138-025-01775-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,26]],"date-time":"2026-01-26T15:07:56Z","timestamp":1769440076000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00138-025-01775-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,18]]},"references-count":29,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,1]]}},"alternative-id":["1775"],"URL":"https:\/\/doi.org\/10.1007\/s00138-025-01775-9","relation":{},"ISSN":["0932-8092","1432-1769"],"issn-type":[{"value":"0932-8092","type":"print"},{"value":"1432-1769","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,12,18]]},"assertion":[{"value":"22 June 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 November 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"1 December 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 December 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Informed consent was obtained from all individual participants included in the study.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to participate"}}],"article-number":"17"}}