{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,26]],"date-time":"2026-03-26T15:58:30Z","timestamp":1774540710927,"version":"3.50.1"},"reference-count":47,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2024,7,17]],"date-time":"2024-07-17T00:00:00Z","timestamp":1721174400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,7,17]],"date-time":"2024-07-17T00:00:00Z","timestamp":1721174400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61771220"],"award-info":[{"award-number":["61771220"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62271226"],"award-info":[{"award-number":["62271226"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Cluster Comput"],"published-print":{"date-parts":[[2024,12]]},"DOI":"10.1007\/s10586-024-04679-x","type":"journal-article","created":{"date-parts":[[2024,7,17]],"date-time":"2024-07-17T10:01:58Z","timestamp":1721210518000},"page":"14231-14245","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["G-SAM: GMM-based segment anything model for medical image classification and segmentation"],"prefix":"10.1007","volume":"27","author":[{"given":"Xiaoxiao","family":"Liu","sequence":"first","affiliation":[]},{"given":"Yan","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Shigang","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Jian","family":"Wei","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,7,17]]},"reference":[{"key":"4679_CR1","doi-asserted-by":"publisher","first-page":"70","DOI":"10.1109\/RBME.2022.3185292","volume":"16","author":"TA Soomro","year":"2022","unstructured":"Soomro, T.A., Zheng, L., Afifi, A.J., Ali, A., Soomro, S., Yin, M., Gao, J.: Image segmentation for MR brain tumor detection using machine learning: a review. IEEE Rev. Biomed. Eng. 16, 70\u201390 (2022). https:\/\/doi.org\/10.1109\/RBME.2022.3185292","journal-title":"IEEE Rev. Biomed. Eng."},{"issue":"7802","key":"4679_CR2","doi-asserted-by":"publisher","first-page":"252","DOI":"10.1038\/s41586-020-2145-8","volume":"580","author":"D Ouyang","year":"2020","unstructured":"Ouyang, D., He, B., Ghorbani, A., Yuan, N., Ebinger, J., Langlotz, C.P., Heidenreich, P.A., Harrington, R.A., Liang, D.H., Ashley, E.A., et al.: Video-based AI for beat-to-beat assessment of cardiac function. Nature 580(7802), 252\u2013256 (2020). https:\/\/doi.org\/10.1038\/s41586-020-2145-8","journal-title":"Nature"},{"key":"4679_CR3","doi-asserted-by":"publisher","first-page":"1026","DOI":"10.1016\/j.media.2022.102685","volume":"83","author":"S Graham","year":"2023","unstructured":"Graham, S., Vu, Q.D., Jahanifar, M., Raza, S.E.A., Minhas, F., Snead, D., Rajpoot, N.: One model is all you need: multi-task learning enables simultaneous histology image segmentation and classification. Med. Image Anal. 83, 1026 (2023). https:\/\/doi.org\/10.1016\/j.media.2022.102685","journal-title":"Med. Image Anal."},{"issue":"6","key":"4679_CR4","doi-asserted-by":"publisher","first-page":"1542","DOI":"10.1109\/TMI.2021.3060497","volume":"40","author":"KB Girum","year":"2021","unstructured":"Girum, K.B., Cr\u00e9hange, G., Lalande, A.: Learning with context feedback loop for robust medical image segmentation. IEEE Trans. Med. Imaging 40(6), 1542\u20131554 (2021). https:\/\/doi.org\/10.1109\/TMI.2021.3060497","journal-title":"IEEE Trans. Med. Imaging"},{"key":"4679_CR5","doi-asserted-by":"publisher","DOI":"10.1007\/s10586-023-04245-x","author":"A Manocha","year":"2024","unstructured":"Manocha, A., Sood, S.K., Bhatia, M.: Edge intelligence-assisted smart healthcare solution for health pandemic: a federated environment approach. Clust. Comput. (2024). https:\/\/doi.org\/10.1007\/s10586-023-04245-x","journal-title":"Clust. Comput."},{"key":"4679_CR6","doi-asserted-by":"publisher","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-net: convolutional networks for biomedical image segmentation, in: Medical image computing and computer-assisted intervention\u2013MICCAI 2015: 18th international conference, Munich, Germany, October 5\u20139, 2015, proceedings, part III 18. pp. 234\u2013241. Springer (2015). https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"4679_CR7","doi-asserted-by":"publisher","unstructured":"Huang, H., Lin, L., Tong, R., Hu, H., Zhang, Q., Iwamoto, Y., Han, X., Chen, Y.-W., Wu, J.: Unet 3+: a full-scale connected unet for medical image segmentation, in: ICASSP 2020\u20132020 IEEE international conference on acoustics, speech and signal processing (ICASSP), pp. 1055\u20131059. IEEE (2020). https:\/\/doi.org\/10.1109\/ICASSP40776.2020.9053405","DOI":"10.1109\/ICASSP40776.2020.9053405"},{"key":"4679_CR8","doi-asserted-by":"publisher","first-page":"102797","DOI":"10.1016\/j.media.2023.102797","volume":"86","author":"F Bougourzi","year":"2023","unstructured":"Bougourzi, F., Distante, C., Dornaika, F., Taleb-Ahmed, A.: Pdatt-UNet: pyramid dual-decoder attention UNet for covid-19 infection segmentation from CT-scans. Med. Image Anal. 86, 102797 (2023). https:\/\/doi.org\/10.1016\/j.media.2023.102797","journal-title":"Med. Image Anal."},{"key":"4679_CR9","doi-asserted-by":"publisher","DOI":"10.1007\/s10586-024-04292-y","author":"Z Chen","year":"2024","unstructured":"Chen, Z., Zhu, H., Liu, Y., Gao, X.: Msca-UNet: multi-scale channel attention-based UNet for segmentation of medical ultrasound images. Clust. Comput. (2024). https:\/\/doi.org\/10.1007\/s10586-024-04292-y","journal-title":"Clust. Comput."},{"issue":"5","key":"4679_CR10","doi-asserted-by":"publisher","first-page":"3145","DOI":"10.1007\/s10586-022-03951-2","volume":"26","author":"H Abdel-Nabi","year":"2023","unstructured":"Abdel-Nabi, H., Ali, M., Awajan, A., Daoud, M., Alazrai, R., Suganthan, P.N., Ali, T.: A comprehensive review of the deep learning-based tumor analysis approaches in histopathological images: segmentation, classification and multi-learning tasks. Clust. Comput. 26(5), 3145\u20133185 (2023). https:\/\/doi.org\/10.1007\/s10586-022-03951-2","journal-title":"Clust. Comput."},{"key":"4679_CR11","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, \u0141., Polosukhin, I.: Attention is all you need. Adv. Neural Inf. Process Syst. 30 (2017)"},{"key":"4679_CR12","doi-asserted-by":"publisher","unstructured":"Valanarasu, J.M.J., Oza, P., Hacihaliloglu, I., Patel, V.M.: Medical transformer: gated axial-attention for medical image segmentation, in: Medical image computing and computer assisted intervention\u2014MICCAI 2021: 24th international conference, Strasbourg, France, Sept 27\u2013Oct 1, 2021, proceedings, Part I 24, pp. 36\u201346. Springer (2021). https:\/\/doi.org\/10.1007\/978-3-030-87193-2_4","DOI":"10.1007\/978-3-030-87193-2_4"},{"key":"4679_CR13","doi-asserted-by":"publisher","unstructured":"H. Cao, Y. Wang, J. Chen, D. Jiang, X. Zhang, Q. Tian, M. Wang, Swin-UNet: UNet-like pure transformer for medical image segmentation, in: European conference on computer vision, pp. 205\u2013218. Springer, (2022). https:\/\/doi.org\/10.1007\/978-3-031-25066-8_9","DOI":"10.1007\/978-3-031-25066-8_9"},{"key":"4679_CR14","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2022.3178991","volume":"71","author":"A Lin","year":"2022","unstructured":"Lin, A., Chen, B., Xu, J., Zhang, Z., Lu, G., Zhang, D.: Ds-transunet: Dual swin transformer u-net for medical image segmentation. IEEE Trans. Instrum. Meas. 71, 1\u201315 (2022). https:\/\/doi.org\/10.1109\/TIM.2022.3178991","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"1","key":"4679_CR15","doi-asserted-by":"publisher","first-page":"87","DOI":"10.1109\/TPAMI.2022.3152247","volume":"45","author":"K Han","year":"2022","unstructured":"Han, K., Wang, Y., Chen, H., Chen, X., Guo, J., Liu, Z., Tang, Y., Xiao, A., Xu, C., Xu, Y., et al.: A survey on vision transformer. IEEE Trans. Pattern Anal. Mach. Intell. 45(1), 87\u2013110 (2022). https:\/\/doi.org\/10.1109\/TPAMI.2022.3152247","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"4679_CR16","doi-asserted-by":"publisher","unstructured":"Gao, Y., Zhou, M., Metaxas, D.N.: UTNet: a hybrid transformer architecture for medical image segmentation, in: Medical image computing and computer assisted intervention\u2014MICCAI 2021: 24th international conference, Strasbourg, France, Sept 27\u2013Oct 1, 2021, proceedings, Part III 24, pp. 61\u201371. Springer (2021). https:\/\/doi.org\/10.1007\/978-3-030-87199-4_6","DOI":"10.1007\/978-3-030-87199-4_6"},{"key":"4679_CR17","doi-asserted-by":"publisher","unstructured":"Petit, O., Thome, N., Rambour, C., Themyr, L., Collins, T., Soler, L.: Unet transformer: self and cross attention for medical image segmentation, in: Machine learning in medical imaging: 12th international workshop, MLMI 2021, held in conjunction with MICCAI 2021, Strasbourg, France, Sept 27, proceedings 12, pp. 267\u2013276. Springer (2021). https:\/\/doi.org\/10.1007\/978-3-030-87589-3_28","DOI":"10.1007\/978-3-030-87589-3_28"},{"key":"4679_CR18","doi-asserted-by":"publisher","DOI":"10.1007\/s10586-024-04532-1","author":"I Pacal","year":"2024","unstructured":"Pacal, I., Celik, O., Bayram, B., Cunha, A.: Enhancing efficientnetv2 with global and efficient channel attention mechanisms for accurate MRI-based brain tumor classification. Clust. Comput. (2024). https:\/\/doi.org\/10.1007\/s10586-024-04532-1","journal-title":"Clust. Comput."},{"key":"4679_CR19","doi-asserted-by":"crossref","unstructured":"Kirillov, A., Mintun, E., Ravi, N., Mao, H., Rolland, C., Gustafson, L., Xiao, T., Whitehead, S., Berg, A.C., Lo, W.-Y. et al.: Segment anything, in: Proceedings of the IEEE\/CVF international conference on computer vision, pp. 4015\u20134026 (2023)","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"4679_CR20","doi-asserted-by":"publisher","first-page":"108238","DOI":"10.1016\/j.compbiomed.2024.108238","volume":"171","author":"Y Zhang","year":"2024","unstructured":"Zhang, Y., Shen, Z., Jiao, R.: Segment anything model for medical image segmentation: current applications and future directions. Comput. Biol. Med. 171, 108238 (2024). https:\/\/doi.org\/10.1016\/j.compbiomed.2024.108238","journal-title":"Comput. Biol. Med."},{"key":"4679_CR21","doi-asserted-by":"publisher","first-page":"103061","DOI":"10.1016\/j.media.2023.103061","volume":"92","author":"Y Huang","year":"2024","unstructured":"Huang, Y., Yang, X., Liu, L., Zhou, H., Chang, A., Zhou, X., Chen, R., Yu, J., Chen, J., Chen, C., et al.: Segment anything model for medical images? Med. Image Anal. 92, 103061 (2024). https:\/\/doi.org\/10.1016\/j.media.2023.103061","journal-title":"Med. Image Anal."},{"issue":"1","key":"4679_CR22","doi-asserted-by":"publisher","first-page":"654","DOI":"10.1038\/s41467-024-44824-z","volume":"15","author":"J Ma","year":"2024","unstructured":"Ma, J., He, Y., Li, F., Han, L., You, C., Wang, B.: Segment anything in medical images. Nat. Commun. 15(1), 654 (2024). https:\/\/doi.org\/10.1038\/s41467-024-44824-z","journal-title":"Nat. Commun."},{"key":"4679_CR23","unstructured":"Hu, E.J., Shen, Y., Wallis, P., Allen-Zhu, Z., Li, Y., Wang, S., Wang, L., Chen, W.: Lora: low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685 (2021)"},{"issue":"2","key":"4679_CR24","doi-asserted-by":"publisher","first-page":"915","DOI":"10.1118\/1.3528204","volume":"38","author":"SG Armato III","year":"2011","unstructured":"Armato, S.G., III., McLennan, G., Bidaut, L., McNitt-Gray, M.F., Meyer, C.R., Reeves, A.P., Zhao, B., Aberle, D.R., Henschke, C.I., Hoffman, E.A., et al.: The lung image database consortium (LIDC) and image database resource initiative (IDRI): a completed reference database of lung nodules on CT scans. Med. Phys. 38(2), 915\u2013931 (2011). https:\/\/doi.org\/10.1118\/1.3528204","journal-title":"Med. Phys."},{"key":"4679_CR25","unstructured":"Codella, N., Rotemberg, V., Tschandl, P., Celebi, M.E., Dusza, S., Gutman, D., Helba, B., Kalloo, A., Liopyris, K., Marchetti, M. et al.: Skin lesion analysis toward melanoma detection 2018: a challenge hosted by the international skin imaging collaboration (ISIC), arXiv preprint arXiv:1902.03368 (2019)"},{"issue":"10","key":"4679_CR26","doi-asserted-by":"publisher","first-page":"1993","DOI":"10.1109\/TMI.2014.2377694","volume":"34","author":"BH Menze","year":"2014","unstructured":"Menze, B.H., Jakab, A., Bauer, S., Kalpathy-Cramer, J., Farahani, K., Kirby, J., Burren, Y., Porz, N., Slotboom, J., Wiest, R., et al.: The multimodal brain tumor image segmentation benchmark (BRATS). IEEE Trans. Med. Imaging 34(10), 1993\u20132024 (2014). https:\/\/doi.org\/10.1109\/TMI.2014.2377694","journal-title":"IEEE Trans. Med. Imaging"},{"issue":"1","key":"4679_CR27","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1038\/sdata.2017.117","volume":"4","author":"S Bakas","year":"2017","unstructured":"Bakas, S., Akbari, H., Sotiras, A., Bilello, M., Rozycki, M., Kirby, J.S., Freymann, J.B., Farahani, K., Davatzikos, C.: Advancing the cancer genome atlas glioma MRI collections with expert segmentation labels and radiomic features. Scientific data 4(1), 1\u201313 (2017). https:\/\/doi.org\/10.1038\/sdata.2017.117","journal-title":"Scientific data"},{"key":"4679_CR28","unstructured":"Heller, N., Sathianathen, N., Kalapara, A., Walczak, E., Moore, K., Kaluzniak, H., Rosenberg, J., Blake, P., Rengel, Z., Oestreich, M. et al.: The kits19 challenge data: 300 kidney tumor cases with clinical context, CT semantic segmentations, and surgical outcomes. arXiv preprint arXiv:1904.00445 (2019)"},{"key":"4679_CR29","doi-asserted-by":"publisher","unstructured":"Zhang, L., Deng, X., Lu, Y.: Segment anything model (SAM) for medical image segmentation: a preliminary review, in: 2023 IEEE international conference on bioinformatics and biomedicine (BIBM), pp. 4187\u20134194. IEEE (2023) https:\/\/doi.org\/10.1109\/BIBM58861.2023.10386032","DOI":"10.1109\/BIBM58861.2023.10386032"},{"key":"4679_CR30","doi-asserted-by":"publisher","first-page":"102918","DOI":"10.1016\/j.media.2023.102918","volume":"89","author":"MA Mazurowski","year":"2023","unstructured":"Mazurowski, M.A., Dong, H., Gu, H., Yang, J., Konz, N., Zhang, Y.: Segment anything model for medical image analysis: an experimental study. Med. Image Anal. 89, 102918 (2023). https:\/\/doi.org\/10.1016\/j.media.2023.102918","journal-title":"Med. Image Anal."},{"key":"4679_CR31","doi-asserted-by":"crossref","unstructured":"Zhang, K., Liu, D.: Customized segment anything model for medical image segmentation. arXiv preprint arXiv:2304.13785 (2023)","DOI":"10.2139\/ssrn.4495221"},{"key":"4679_CR32","doi-asserted-by":"crossref","unstructured":"Wang, L., Ye, X., Zhu, L., Wu, W., Zhang, J., Xing, H., Hu, C.: When SAM meets sonar images. arXiv preprint arXiv:2306.14109 (2023)","DOI":"10.1109\/LGRS.2024.3387712"},{"key":"4679_CR33","doi-asserted-by":"crossref","unstructured":"Fazekas, B., Morano, J., Lachinov, D., Aresta, G., Bogunovi\u0107, H.: Samedoct: adapting segment anything model (SAM) for retinal Oct. arXiv preprint arXiv:2308.09331 (2023)","DOI":"10.1007\/978-3-031-44013-7_10"},{"key":"4679_CR34","unstructured":"Dettmers, T., Pagnoni, A., Holtzman, A., Zettlemoyer, L.: Qlora: efficient finetuning of quantized LLMS. Adv. Neural Inf. Process. Syst 36 (2024)"},{"key":"4679_CR35","doi-asserted-by":"publisher","first-page":"355","DOI":"10.1146\/annurev-statistics-031017-100325","volume":"6","author":"GJ McLachlan","year":"2019","unstructured":"McLachlan, G.J., Lee, S.X., Rathnayake, S.I.: Finite mixture models. Ann. Rev. Stat. Appl. 6, 355\u2013378 (2019). https:\/\/doi.org\/10.1146\/annurev-statistics-031017-100325","journal-title":"Ann. Rev. Stat. Appl."},{"issue":"7","key":"4679_CR36","doi-asserted-by":"publisher","first-page":"1219","DOI":"10.3390\/rs12071219","volume":"12","author":"X Shi","year":"2020","unstructured":"Shi, X., Li, Y., Zhao, Q.: Flexible hierarchical Gaussian mixture model for high-resolution remote sensing image segmentation. Remote Sens. 12(7), 1219 (2020). https:\/\/doi.org\/10.3390\/rs12071219","journal-title":"Remote Sens."},{"key":"4679_CR37","doi-asserted-by":"publisher","first-page":"77323","DOI":"10.1109\/ACCESS.2022.3192605","volume":"10","author":"D Saire","year":"2022","unstructured":"Saire, D., Rivera, A.R.: Global and local features through Gaussian mixture models on image semantic segmentation. IEEE Access 10, 77323\u201377336 (2022). https:\/\/doi.org\/10.1109\/ACCESS.2022.3192605","journal-title":"IEEE Access"},{"key":"4679_CR38","doi-asserted-by":"publisher","first-page":"179","DOI":"10.1016\/j.cmpb.2019.04.026","volume":"175","author":"M Hassan","year":"2019","unstructured":"Hassan, M., Murtza, I., Hira, A., Ali, S., Kifayat, K.: Robust spatial fuzzy GMM based MRI segmentation and carotid artery plaque detection in ultrasound images. Comput. Methods Programs Biomed. 175, 179\u2013192 (2019). https:\/\/doi.org\/10.1016\/j.cmpb.2019.04.026","journal-title":"Comput. Methods Programs Biomed."},{"key":"4679_CR39","doi-asserted-by":"publisher","first-page":"41","DOI":"10.1016\/j.ins.2021.06.034","volume":"575","author":"Y Chen","year":"2021","unstructured":"Chen, Y., Cheng, N., Cai, M., Cao, C., Yang, J., Zhang, Z.: A spatially constrained asymmetric Gaussian mixture model for image segmentation. Inf. Sci. 575, 41\u201365 (2021). https:\/\/doi.org\/10.1016\/j.ins.2021.06.034","journal-title":"Inf. Sci."},{"issue":"1","key":"4679_CR40","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1109\/TNNLS.2021.3119071","volume":"33","author":"A Chaddad","year":"2021","unstructured":"Chaddad, A., Hassan, L., Desrosiers, C.: Deep radiomic analysis for predicting coronavirus disease 2019 in computerized tomography and X-ray images. IEEE Trans. Neural Networks Learn. Syst. 33(1), 3\u201311 (2021). https:\/\/doi.org\/10.1109\/TNNLS.2021.3119071","journal-title":"IEEE Trans. Neural Networks Learn. Syst."},{"issue":"9","key":"4679_CR41","doi-asserted-by":"publisher","first-page":"1089","DOI":"10.1109\/TPAMI.2003.1227985","volume":"25","author":"F Forbes","year":"2003","unstructured":"Forbes, F., Peyrard, N.: Hidden Markov random field model selection criteria based on mean field-like approximations. IEEE Trans. Pattern Anal. Mach. Intell. 25(9), 1089\u20131101 (2003). https:\/\/doi.org\/10.1109\/TPAMI.2003.1227985","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"4679_CR42","doi-asserted-by":"publisher","first-page":"107990","DOI":"10.1016\/j.sigpro.2021.107990","volume":"183","author":"J Sun","year":"2021","unstructured":"Sun, J., Zhao, Y., Wang, S., Wei, J.: Image compression based on Gaussian mixture model constrained using Markov random field. Signal Process. 183, 107990 (2021). https:\/\/doi.org\/10.1016\/j.sigpro.2021.107990","journal-title":"Signal Process."},{"issue":"8","key":"4679_CR43","doi-asserted-by":"publisher","first-page":"4551","DOI":"10.1007\/s11760-023-02690-0","volume":"17","author":"T Hou","year":"2023","unstructured":"Hou, T., Zhu, H., Yang, S.: BM-GMM: belief function-based Gaussian Markov model for image segmentation. SIViP 17(8), 4551\u20134560 (2023). https:\/\/doi.org\/10.1007\/s11760-023-02690-0","journal-title":"SIViP"},{"key":"4679_CR44","doi-asserted-by":"crossref","unstructured":"He, K., Chen, X., Xie, S., Li, Y., Doll\u00e1r, P., Girshick, R.: Masked autoencoders are scalable vision learners, in: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 16000\u201316009 (2022)","DOI":"10.1109\/CVPR52688.2022.01553"},{"key":"4679_CR45","unstructured":"Gower, R.M., Loizou, M., Qian, X., Sailanbayev, A., Shulgin, E., Richt\u00e1rik, P.: Sgd: general analysis and improved rates, in: International conference on machine learning, pp. 5200\u20135209. PMLR (2019)"},{"issue":"10","key":"4679_CR46","doi-asserted-by":"publisher","first-page":"2281","DOI":"10.1109\/TMI.2019.2903562","volume":"38","author":"Z Gu","year":"2019","unstructured":"Gu, Z., Cheng, J., Fu, H., Zhou, K., Hao, H., Zhao, Y., Zhang, T., Gao, S., Liu, J.: Ce-net: context encoder network for 2d medical image segmentation. IEEE Trans. Med. Imaging 38(10), 2281\u20132292 (2019). https:\/\/doi.org\/10.1109\/TMI.2019.2903562","journal-title":"IEEE Trans. Med. Imaging"},{"key":"4679_CR47","doi-asserted-by":"publisher","unstructured":"Oktay, O., Schlemper, J., Folgoc, L.L., Lee, M., Heinrich, M., Misawa, K., Mori, K., McDonagh, S., Hammerla, N.Y., Kainz, B. et al.: Attention u-net: learning where to look for the pancreas, arXiv preprint arXiv:1804.03999 (2018). https:\/\/doi.org\/10.48550\/arXiv.1804.03999","DOI":"10.48550\/arXiv.1804.03999"}],"container-title":["Cluster Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10586-024-04679-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10586-024-04679-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10586-024-04679-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,9,27]],"date-time":"2024-09-27T22:07:15Z","timestamp":1727474835000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10586-024-04679-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,7,17]]},"references-count":47,"journal-issue":{"issue":"10","published-print":{"date-parts":[[2024,12]]}},"alternative-id":["4679"],"URL":"https:\/\/doi.org\/10.1007\/s10586-024-04679-x","relation":{},"ISSN":["1386-7857","1573-7543"],"issn-type":[{"value":"1386-7857","type":"print"},{"value":"1573-7543","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,7,17]]},"assertion":[{"value":"28 May 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"25 June 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 July 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 July 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interest"}}]}}