{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,11]],"date-time":"2026-02-11T21:47:11Z","timestamp":1770846431874,"version":"3.50.1"},"reference-count":45,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,12,15]],"date-time":"2025-12-15T00:00:00Z","timestamp":1765756800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,12,15]],"date-time":"2025-12-15T00:00:00Z","timestamp":1765756800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001667","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62401097"],"award-info":[{"award-number":["62401097"]}],"id":[{"id":"10.13039\/501100001667","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100005047","name":"Natural Science Foundation of Liaoning Province","doi-asserted-by":"publisher","award":["2024-BS-028"],"award-info":[{"award-number":["2024-BS-028"]}],"id":[{"id":"10.13039\/501100005047","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012226","name":"Fundamental Research Funds for Central Universities","doi-asserted-by":"publisher","award":["0854-53"],"award-info":[{"award-number":["0854-53"]}],"id":[{"id":"10.13039\/501100012226","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100019032","name":"Dalian Minzu University","doi-asserted-by":"publisher","award":["0854-53"],"award-info":[{"award-number":["0854-53"]}],"id":[{"id":"10.13039\/501100019032","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,12,15]]},"DOI":"10.1109\/bibm66473.2025.11356495","type":"proceedings-article","created":{"date-parts":[[2026,1,29]],"date-time":"2026-01-29T21:19:40Z","timestamp":1769721580000},"page":"6631-6638","source":"Crossref","is-referenced-by-count":0,"title":["Seg4Fusion: A Tumor-Aware Framework for 3D Medical Image Fusion"],"prefix":"10.1109","author":[{"given":"Haiyu","family":"Song","sequence":"first","affiliation":[{"name":"College of Computer Science and Engineering, Dalian Minzu University,Dalian,China"}]},{"given":"Jiawei","family":"Feng","sequence":"additional","affiliation":[{"name":"College of Computer Science and Engineering, Dalian Minzu University,Dalian,China"}]},{"given":"Maoyu","family":"Wang","sequence":"additional","affiliation":[{"name":"College of Computer Science and Engineering, Dalian Minzu University,Dalian,China"}]},{"given":"Wenqi","family":"Wang","sequence":"additional","affiliation":[{"name":"College of Computer Science and Engineering, Dalian Minzu University,Dalian,China"}]},{"given":"Aohua","family":"Ma","sequence":"additional","affiliation":[{"name":"College of Computer Science and Engineering, Dalian Minzu University,Dalian,China"}]},{"given":"Zeyu","family":"Wang","sequence":"additional","affiliation":[{"name":"College of Computer Science and Engineering, Dalian Minzu University,Dalian,China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2020.2977573"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2022.3193288"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2021.108434"},{"key":"ref4","article-title":"Adafuse: Adaptive medical image fusion based on spatial-frequential cross attention","volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","author":"Zhang"},{"key":"ref5","volume-title":"Fuse4seg: Imagelevel fusion based multi-modality medical image segmentation","author":"Wang","year":"2024"},{"key":"ref6","volume-title":"Drifa-net: Multimodal fusion learning with dual attention for medical imaging","author":"Chen","year":"2024"},{"key":"ref7","doi-asserted-by":"crossref","DOI":"10.1007\/978-3-031-18814-5","volume-title":"An attention-based multiscale feature learning network for multimodal medical image fusion","author":"Li","year":"2022"},{"issue":"11","key":"ref8","doi-asserted-by":"crossref","DOI":"10.3390\/electronics14112266","article-title":"Multimodal medical image fusion using a progressive parallel strategy based on deep learning","volume":"14","author":"Peng","year":"2025","journal-title":"Electronics"},{"issue":"11","key":"ref9","first-page":"2266","article-title":"Ppmf-net: Progressive parallel network for pet-mri fusion","volume":"14","author":"Guo","year":"2024","journal-title":"Electronics"},{"key":"ref10","volume-title":"Multi-modal evidential fusion network for trustworthy pet\/ct tumor segmentation","author":"Qi","year":"2024"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2014.05.004"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.3906\/elk-2105-170"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1080\/24725854.2024.2376650"},{"key":"ref14","first-page":"45934602","article-title":"Ffnet: Feature fusion network for infrared and visible image fusion","volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","author":"Tang"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1016\/j.patrec.2020.08.002"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-021-01501-8"},{"issue":"07","key":"ref17","first-page":"13094","article-title":"Cross-modal and hierarchical modeling for image-text matching","volume-title":"Proceedings of the AAAI Conference on Artificial Intelligence","volume":"34","author":"Zhou"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i7.28536"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1007\/978-981-99-8141-0_41"},{"key":"ref20","article-title":"Correlation-driven dual-branch feature decomposition fusion for multi-modality image fusion","author":"Zhao","year":"2023","journal-title":"CVPR 2023"},{"key":"ref21","article-title":"A new multimodal medical image fusion based on laplacian autoencoder with channel attention","author":"Wankhede","year":"2023","journal-title":"arXiv preprint"},{"key":"ref22","article-title":"Ecfusion: Edge-augmented cross-scale transformer fusion for medical image fusion","author":"List","year":"2025","journal-title":"Scientific Reports"},{"key":"ref23","article-title":"Ppmf-net: Progressive parallel network for pet-mri fusion","author":"Guo","year":"2024","journal-title":"Electronics"},{"key":"ref24","article-title":"Cmaf-net: Cross-modal attention fusion network for incomplete multimodal brain tumor segmentation","author":"C.","year":"2023","journal-title":"Comput Biol Med"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/wacv57701.2024.00132"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1038\/s41598-025-90339-y"},{"key":"ref27","article-title":"Multi-modal brain tumor segmentation via 3d multi-scale self-attention and cross-attention (tma-transbts)","author":"H.","year":"2025","journal-title":"arXiv preprint"},{"key":"ref28","article-title":"Smaformer: Synergistic multi-attention transformer for medical image segmentation","author":"Z.","year":"2024","journal-title":"arXiv preprint"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/WACV.2018.00066"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-08999-2_22"},{"key":"ref31","volume-title":"Monai: An open-source framework for deep learning in healthcare","author":"Cardoso","year":"2022"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/TMI.2014.2377694"},{"key":"ref33","volume-title":"Adam: A method for stochastic optimization","author":"Kingma","year":"2017"},{"key":"ref34","article-title":"Identifying the best machine learning algorithms for brain tumor segmentation, progression assessment, and overall survival prediction in the brats challenge","author":"Bakas","year":"2020","journal-title":"arXiv preprint"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1155\/2015\/813696"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1049\/el:20020212"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2013.2244222"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1049\/el:20000267"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2003.819861"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.02448"},{"key":"ref41","article-title":"Image fusion via vision-language model","volume-title":"Proceedings of the 41st International Conference on Machine Learning, ser. ICML\u201924. JMLR.org","author":"Zhao"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-025-02427-1"},{"key":"ref43","volume-title":"Dcevo: Discriminative cross-dimensional evolutionary learning for infrared and visible image fusion","author":"Liu","year":"2025"},{"key":"ref44","volume-title":"One model for all: Low-level task interaction is a key to task-agnostic image fusion","author":"Cheng","year":"2025"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-91721-9_4"}],"event":{"name":"2025 IEEE International Conference on Bioinformatics and Biomedicine (BIBM)","location":"Wuhan, China","start":{"date-parts":[[2025,12,15]]},"end":{"date-parts":[[2025,12,18]]}},"container-title":["2025 IEEE International Conference on Bioinformatics and Biomedicine (BIBM)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11355913\/11355975\/11356495.pdf?arnumber=11356495","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,11]],"date-time":"2026-02-11T20:53:26Z","timestamp":1770843206000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11356495\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,15]]},"references-count":45,"URL":"https:\/\/doi.org\/10.1109\/bibm66473.2025.11356495","relation":{},"subject":[],"published":{"date-parts":[[2025,12,15]]}}}