{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,29]],"date-time":"2025-09-29T00:04:36Z","timestamp":1759104276809,"version":"3.44.0"},"publisher-location":"Cham","reference-count":38,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783032061171","type":"print"},{"value":"9783032061188","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,9,29]],"date-time":"2025-09-29T00:00:00Z","timestamp":1759104000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,9,29]],"date-time":"2025-09-29T00:00:00Z","timestamp":1759104000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-3-032-06118-8_17","type":"book-chapter","created":{"date-parts":[[2025,9,28]],"date-time":"2025-09-28T11:23:06Z","timestamp":1759058586000},"page":"285-303","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["WoundAmbit: Bridging State-of-the-Art Semantic Segmentation and\u00a0Real-World Wound Care"],"prefix":"10.1007","author":[{"given":"Vanessa","family":"Borst","sequence":"first","affiliation":[]},{"given":"Timo","family":"Dittus","sequence":"additional","affiliation":[]},{"given":"Tassilo","family":"Dege","sequence":"additional","affiliation":[]},{"given":"Astrid","family":"Schmieder","sequence":"additional","affiliation":[]},{"given":"Samuel","family":"Kounev","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,9,29]]},"reference":[{"key":"17_CR1","doi-asserted-by":"crossref","unstructured":"Awais, M., et al.: Foundation models defining a new era in vision: a survey and outlook. IEEE Trans. Pattern Anal. Mach. Intell. 47(4) (2025)","DOI":"10.1109\/TPAMI.2024.3506283"},{"key":"17_CR2","doi-asserted-by":"crossref","unstructured":"Azad, R., et al.: Advances in medical image analysis with vision transformers: a comprehensive review. Med. Image Anal. 91 (2024)","DOI":"10.1016\/j.media.2023.103000"},{"key":"17_CR3","unstructured":"Borst, V., et al.: WoundAIssist: a patient-centered mobile app for AI-assisted wound care with physicians in the loop. arXiv (2025). https:\/\/arxiv.org\/abs\/2506.06104"},{"key":"17_CR4","doi-asserted-by":"crossref","unstructured":"Chairat, S., Chaichulee, S., Dissaneewate, T., Wangkulangkul, P., Kongpanichakul, L.: AI-assisted assessment of wound tissue with automatic color and measurement calibration on images taken with a smartphone. Healthcare 11(2) (2023)","DOI":"10.3390\/healthcare11020273"},{"key":"17_CR5","doi-asserted-by":"crossref","unstructured":"Chino, D.Y., Scabora, L.C., Cazzolato, M.T., Jorge, A.E., Traina-Jr, C., Traina, A.J.: Segmenting skin ulcers and measuring the wound area using deep convolutional networks. Comput. Methods Programs Biomed. 191 (2020)","DOI":"10.1016\/j.cmpb.2020.105376"},{"key":"17_CR6","doi-asserted-by":"crossref","unstructured":"Dhar, M.K., Zhang, T., Patel, Y., Gopalakrishnan, S., Yu, Z.: FUSegNet: a deep convolutional neural network for foot ulcer segmentation. Biomed. Sig. Process. Control 92 (2024)","DOI":"10.1016\/j.bspc.2024.106057"},{"key":"17_CR7","doi-asserted-by":"crossref","unstructured":"Foltynski, P., Ladyzynski, P.: Internet service for wound area measurement using digital planimetry with adaptive calibration and image segmentation with deep convolutional neural networks. Biocybernetics Biomed. Eng. 43(1) (2023)","DOI":"10.1016\/j.bbe.2022.11.004"},{"key":"17_CR8","doi-asserted-by":"crossref","unstructured":"Goyal, M., Yap, M.H., Reeves, N.D., Rajbhandari, S., Spragg, J.: Fully convolutional networks for diabetic foot ulcer segmentation. In: IEEE International Conference on Systems, Man, and Cybernetics (2017)","DOI":"10.1109\/SMC.2017.8122675"},{"key":"17_CR9","unstructured":"Guo, M.H., Lu, C.Z., Hou, Q., Liu, Z.N., Cheng, M.M., Hu, S.M.: SegNeXt: rethinking convolutional attention design for semantic segmentation. In: NeurIPS (2022)"},{"key":"17_CR10","doi-asserted-by":"crossref","unstructured":"Heidari, M., et al.: HiFormer: hierarchical multi-scale representations using transformers for medical image segmentation. In: WACV (2023)","DOI":"10.1109\/WACV56688.2023.00614"},{"key":"17_CR11","doi-asserted-by":"crossref","unstructured":"Huang, X., Deng, Z., Li, D., Yuan, X., Fu, Y.: MISSFormer: an effective transformer for 2D medical image segmentation. IEEE Trans. Med. Imaging 42(5) (2022)","DOI":"10.1109\/TMI.2022.3230943"},{"key":"17_CR12","unstructured":"imito AG: imitoWound (2025). https:\/\/imito.io\/en\/imitowound"},{"key":"17_CR13","doi-asserted-by":"crossref","unstructured":"J\u00f8rgensen, L.B., S\u00f8rensen, J.A., Jemec, G.B., Yderstr\u00e6de, K.B.: Methods to assess area and volume of wounds \u2013 a systematic review. Int. Wound J. 13(4) (2016)","DOI":"10.1111\/iwj.12472"},{"key":"17_CR14","doi-asserted-by":"crossref","unstructured":"Kendrick, C., et al.: Translating clinical delineation of diabetic foot ulcers into machine interpretable segmentation. In: Diabetic Foot Ulcers Grand Challenge (2025)","DOI":"10.1007\/978-3-031-80871-5_1"},{"key":"17_CR15","doi-asserted-by":"crossref","unstructured":"Kirillov, A., et\u00a0al.: Segment anything. In: ICCV (2023)","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"17_CR16","doi-asserted-by":"crossref","unstructured":"Langemo, D., Anderson, J., Hanson, D., Hunter, S., Thompson, P.: Measuring wound length, width, and area: which technique? Adv. Skin Wound Care 21(1) (2008)","DOI":"10.1097\/01.ASW.0000284967.69863.2f"},{"key":"17_CR17","doi-asserted-by":"crossref","unstructured":"Liao, T.Y., Yang, C.H., Lo, Y.W., Lai, K.Y., Shen, P.H., Lin, Y.L.: HarDNet-DFUS: enhancing backbone and decoder of HarDNet-MSEG for diabetic foot ulcer image segmentation. In: Diabetic Foot Ulcers Grand Challenge (2023)","DOI":"10.1007\/978-3-031-26354-5_2"},{"key":"17_CR18","doi-asserted-by":"crossref","unstructured":"Liu, X., Wang, C., Li, F., Zhao, X., Zhu, E., Peng, Y.: A framework of wound segmentation based on deep convolutional networks. In: International Congress on Image and Signal Processing, Biomedical Engineering and Informatics (2017)","DOI":"10.1109\/CISP-BMEI.2017.8302184"},{"key":"17_CR19","doi-asserted-by":"crossref","unstructured":"Ma, J., He, Y., Li, F., Han, L., You, C., Wang, B.: Segment anything in medical images. Nat. Commun. 15(1) (2024)","DOI":"10.1038\/s41467-024-44824-z"},{"key":"17_CR20","doi-asserted-by":"crossref","unstructured":"Minaee, S., Boykov, Y., Porikli, F., Plaza, A., Kehtarnavaz, N., Terzopoulos, D.: Image segmentation using deep learning: a survey. IEEE Trans. Pattern Anal. Mach. Intell. 44(7) (2022)","DOI":"10.1109\/TPAMI.2021.3059968"},{"key":"17_CR21","doi-asserted-by":"crossref","unstructured":"Oota, S.R., Rowtula, V., Mohammed, S., Liu, M., Gupta, M.: WSNet: towards an effective method for wound image segmentation. In: WACV (2023)","DOI":"10.1109\/WACV56688.2023.00325"},{"key":"17_CR22","doi-asserted-by":"crossref","unstructured":"Ramachandram, D., et\u00a0al.: Fully automated wound tissue segmentation using deep learning on mobile devices: cohort study. JMIR mHealth and uHealth 10(4) (2022)","DOI":"10.2196\/36977"},{"key":"17_CR23","doi-asserted-by":"crossref","unstructured":"Rayed, M.E., Islam, S.S., Niha, S.I., Jim, J.R., Kabir, M.M., Mridha, M.: Deep learning for medical image segmentation: state-of-the-art advancements and challenges. Inf. Med. Unlocked (2024)","DOI":"10.1016\/j.imu.2024.101504"},{"key":"17_CR24","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"234","DOI":"10.1007\/978-3-319-24574-4_28","volume-title":"Medical Image Computing and Computer-Assisted Intervention \u2013 MICCAI 2015","author":"O Ronneberger","year":"2015","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234\u2013241. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28"},{"key":"17_CR25","doi-asserted-by":"crossref","unstructured":"Sanderson, E., Matuszewski, B.J.: FCN-transformer feature fusion for polyp segmentation. In: Medical Image Understanding and Analysis (2022)","DOI":"10.1007\/978-3-031-12053-4_65"},{"key":"17_CR26","doi-asserted-by":"crossref","unstructured":"Shaw, J., Bell, P.M.: Wound measurement in diabetic foot ulceration. In: Global Perspective on Diabetic Foot Ulcerations, chap.\u00a05. IntechOpen (2011)","DOI":"10.5772\/29730"},{"key":"17_CR27","doi-asserted-by":"crossref","unstructured":"Shi, D.: TransNeXt: robust foveal visual perception for vision transformers. In: CVPR (2024)","DOI":"10.1109\/CVPR52733.2024.01683"},{"key":"17_CR28","doi-asserted-by":"crossref","unstructured":"Siddique, N., Paheding, S., Elkin, C.P., Devabhaktuni, V.: U-Net and its variants for medical image segmentation: a review of theory and applications. IEEE Access 9 (2021)","DOI":"10.1109\/ACCESS.2021.3086020"},{"key":"17_CR29","doi-asserted-by":"crossref","unstructured":"Thisanke, H., Deshan, C., Chamith, K., Seneviratne, S., Vidanaarachchi, R., Herath, D.: Semantic segmentation using vision transformers: a survey. Eng. Appl. Artif. Intell. 126 (2023)","DOI":"10.1016\/j.engappai.2023.106669"},{"key":"17_CR30","doi-asserted-by":"crossref","unstructured":"Wang, C., et al.: Fully automatic wound segmentation with deep convolutional neural networks. Sci. Rep. 10(1) (2020)","DOI":"10.1038\/s41598-020-78799-w"},{"key":"17_CR31","doi-asserted-by":"crossref","unstructured":"Wang, C., et al.: FUSeg: the foot ulcer segmentation challenge. Information 15(3) (2024)","DOI":"10.3390\/info15030140"},{"key":"17_CR32","doi-asserted-by":"crossref","unstructured":"Wang, L., Pedersen, P.C., Agu, E., Strong, D.M., Tulu, B.: Area determination of diabetic foot ulcer images using a cascaded two-stage SVM-based classification. IEEE Trans. Biomed. Eng. 64(9) (2017)","DOI":"10.1109\/TBME.2016.2632522"},{"key":"17_CR33","doi-asserted-by":"crossref","unstructured":"Wang, W., et\u00a0al.: InternImage: exploring large-scale vision foundation models with deformable convolutions. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.01385"},{"key":"17_CR34","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"432","DOI":"10.1007\/978-3-030-01228-1_26","volume-title":"Computer Vision \u2013 ECCV 2018","author":"T Xiao","year":"2018","unstructured":"Xiao, T., Liu, Y., Zhou, B., Jiang, Y., Sun, J.: Unified perceptual parsing for scene understanding. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11209, pp. 432\u2013448. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01228-1_26"},{"key":"17_CR35","unstructured":"Xie, E., Wang, W., Yu, Z., Anandkumar, A., Alvarez, J.M., Luo, P.: SegFormer: simple and efficient design for semantic segmentation with transformers. In: NeurIPS (2021)"},{"key":"17_CR36","unstructured":"Yan, H., Wu, M., Zhang, C.: Multi-scale representations by varying window attention for semantic segmentation. In: ICLR (2024)"},{"key":"17_CR37","doi-asserted-by":"crossref","unstructured":"Zhang, P., et al.: Interactive skin wound segmentation based on feature augment networks. IEEE J. Biomed. Health Inf. 27(7) (2023)","DOI":"10.1109\/JBHI.2023.3270711"},{"key":"17_CR38","doi-asserted-by":"crossref","unstructured":"Zhang, R., Tian, D., Xu, D., Qian, W., Yao, Y.: A survey of wound image analysis using deep learning: classification, detection, and segmentation. IEEE Access 10 (2022)","DOI":"10.1109\/ACCESS.2022.3194529"}],"container-title":["Lecture Notes in Computer Science","Machine Learning and Knowledge Discovery in Databases. Applied Data Science Track"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-032-06118-8_17","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,28]],"date-time":"2025-09-28T11:23:20Z","timestamp":1759058600000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-032-06118-8_17"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9,29]]},"ISBN":["9783032061171","9783032061188"],"references-count":38,"URL":"https:\/\/doi.org\/10.1007\/978-3-032-06118-8_17","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,9,29]]},"assertion":[{"value":"29 September 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"The authors declare no conflicts of interest. Ethical approval for this study was waived by the Ethics Committee of the Medical Faculty of the University of W\u00fcrzburg. All patients provided informed consent for the publication of their photographs.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics"}},{"value":"ECML PKDD","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Joint European Conference on Machine Learning and Knowledge Discovery in Databases","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Porto","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Portugal","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 September 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"19 September 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ecml2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/ecmlpkdd.org\/2025\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}