{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T04:01:28Z","timestamp":1743048088567,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":29,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819985579"},{"type":"electronic","value":"9789819985586"}],"license":[{"start":{"date-parts":[[2023,12,26]],"date-time":"2023-12-26T00:00:00Z","timestamp":1703548800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,12,26]],"date-time":"2023-12-26T00:00:00Z","timestamp":1703548800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-99-8558-6_36","type":"book-chapter","created":{"date-parts":[[2023,12,25]],"date-time":"2023-12-25T20:02:02Z","timestamp":1703534522000},"page":"436-447","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["CDBIFusion: A Cross-Domain Bidirectional Interaction Fusion Network for PET and MRI Images"],"prefix":"10.1007","author":[{"given":"Jie","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Bicao","family":"Li","sequence":"additional","affiliation":[]},{"given":"Bei","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Zhuhong","family":"Shao","sequence":"additional","affiliation":[]},{"given":"Jie","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Jiaxi","family":"Lu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,12,26]]},"reference":[{"key":"36_CR1","doi-asserted-by":"crossref","unstructured":"Zhao, Z., et al.: Cddfuse: correlation-driven dual-branch feature decomposition for multi-modality image fusion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5906\u20135916 (2023)","DOI":"10.1109\/CVPR52729.2023.00572"},{"key":"36_CR2","unstructured":"Zhao, Z., et al.: Equivariant Multi-Modality Image Fusion. ArXiv abs\/2305.11443 (2023)"},{"key":"36_CR3","doi-asserted-by":"publisher","first-page":"1076","DOI":"10.1109\/TIP.2016.2633863","volume":"26","author":"M Al-Mualla","year":"2017","unstructured":"Al-Mualla, M.: Ebrahim, bull, david, hill, paul: perceptual image fusion using wavelets. IEEE Trans. Image Process. 26, 1076\u20131088 (2017)","journal-title":"IEEE Trans. Image Process."},{"key":"36_CR4","doi-asserted-by":"publisher","first-page":"326","DOI":"10.1016\/j.neucom.2016.02.047","volume":"194","author":"J Du","year":"2016","unstructured":"Du, J., Li, W., Xiao, B., Nawaz, Q.: Union Laplacian pyramid with multiple features for medical image fusion. Neurocomputing 194, 326\u2013339 (2016)","journal-title":"Neurocomputing"},{"key":"36_CR5","doi-asserted-by":"crossref","unstructured":"Singh, S., Gupta, D., Anand, R.S., Kumar, V.: Nonsubsampled shearlet based CT and MR medical image fusion using biologically inspired spiking neural network. Biomed. Signal Process. Control (2015)","DOI":"10.1016\/j.bspc.2014.11.009"},{"key":"36_CR6","first-page":"1014","volume":"9","author":"G Bhatnagar","year":"2014","unstructured":"Bhatnagar, G., Wu, Q., Zheng, L.: Directive contrast based multimodal medical image fusion in NSCT domain. IEEE Trans. Multimedia 9, 1014\u20131024 (2014)","journal-title":"IEEE Trans. Multimedia"},{"key":"36_CR7","doi-asserted-by":"publisher","first-page":"347","DOI":"10.1049\/iet-ipr.2014.0311","volume":"9","author":"Y Liu","year":"2014","unstructured":"Liu, Y., Wang, Z.: Simultaneous image fusion and denoising with adaptive sparse representation. Image Process. Iet 9, 347\u2013357 (2014)","journal-title":"Image Process. Iet"},{"key":"36_CR8","unstructured":"Yu, L., Xun, C., Cheng, J., Hu, P.: A medical image fusion method based on convolutional neural networks. In: 2017 20th International Conference on Information Fusion (2017)"},{"key":"36_CR9","doi-asserted-by":"publisher","first-page":"2169","DOI":"10.3390\/s20082169","volume":"20","author":"K Wang","year":"2020","unstructured":"Wang, K., Zheng, M., Wei, H., Qi, G., Li, Y.: Multi-modality medical image fusion using convolutional neural network and contrast pyramid. Sensors 20, 2169 (2020)","journal-title":"Sensors"},{"key":"36_CR10","doi-asserted-by":"crossref","unstructured":"Lahoud, F., S\u00fcsstrunk, S.: Zero-learning fast medical image fusion. In: 2019 22th International Conference on Information Fusion (2019)","DOI":"10.23919\/FUSION43075.2019.9011178"},{"key":"36_CR11","doi-asserted-by":"crossref","unstructured":"Song, S., Wang, J., Wang, Z., Su, J., Ding, X., Dang, K.: Bilateral-Fuser: A Novel Multi-cue Fusion Architecture with Anatomical-aware Tokens for Fovea Localization. arXiv preprint arXiv:2302.06961 (2023)","DOI":"10.1109\/JBHI.2024.3445112"},{"key":"36_CR12","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","volume":"48","author":"J Ma","year":"2019","unstructured":"Ma, J., Yu, W., Liang, P., Li, C., Jiang, J.: FusionGAN: a generative adversarial network for infrared and visible image fusion. Inform. Fus. 48, 11\u201326 (2019)","journal-title":"Inform. Fus."},{"key":"36_CR13","doi-asserted-by":"crossref","unstructured":"Ma, J., Xu, H., Jiang, J., Mei, X., Zhang, X.P.: DDcGAN: a dual-discriminator conditional generative adversarial network for multi-resolution image fusion. IEEE Trans. Image Process. (2020)","DOI":"10.1109\/TIP.2020.2977573"},{"key":"36_CR14","doi-asserted-by":"crossref","unstructured":"Guo, X., Nie, R., Cao, J., Zhou, D., Mei, L., He, K.: FuseGAN: learning to fuse multi-focus image via conditional generative adversarial network. IEEE Trans. Multimedia 1982\u20131996 (2019)","DOI":"10.1109\/TMM.2019.2895292"},{"key":"36_CR15","doi-asserted-by":"crossref","unstructured":"Ma, J., Zhang, H., Shao, Z., Liang, P., Xu, H.: GANMcC: a generative adversarial network with multiclassification constraints for infrared and visible image fusion. IEEE Trans. Instrument. Measure. 1 (2020)","DOI":"10.1109\/TIM.2020.3038013"},{"key":"36_CR16","doi-asserted-by":"publisher","first-page":"40","DOI":"10.1016\/j.inffus.2020.08.022","volume":"66","author":"ZA Hao","year":"2021","unstructured":"Hao, Z.A., Zl, A., Zs, B., Han, X.A., Jm, A.: MFF-GAN: an unsupervised generative adversarial network with adaptive and gradient joint constraints for multi-focus image fusion. Inform. Fusion 66, 40\u201353 (2021)","journal-title":"Inform. Fusion"},{"key":"36_CR17","doi-asserted-by":"crossref","unstructured":"Wang, Z., Shao, W., Chen, Y., Xu, J., Zhang, X.: Infrared and Visible Image Fusion via Interactive Compensatory Attention Adversarial Learning (2022)","DOI":"10.1109\/TMM.2022.3228685"},{"key":"36_CR18","doi-asserted-by":"crossref","unstructured":"Ma, T., Li, B., Liu, W., Hua, M., Dong, J., Tan, T.: CFFT-GAN: cross-domain Feature Fusion Transformer for Exemplar-based Image Translation. ArXiv abs\/2302.01608, (2023)","DOI":"10.1609\/aaai.v37i2.25279"},{"key":"36_CR19","unstructured":"Hu, Q., Guo, X.: Trash or Treasure? An Interactive Dual-Stream Strategy for Single Image Reflection Separation (2021)"},{"key":"36_CR20","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"36_CR21","unstructured":"Kumar, B.S.: Image fusion based on pixel significance using cross bilateral filter. Signal, image and video processing (2015)"},{"key":"36_CR22","doi-asserted-by":"crossref","unstructured":"Dhuli, Ravindra, Bavirisetti, Prasad, D.: Two-scale image fusion of visible and infrared images using saliency detection. Infrared Phys. Technol. (2016)","DOI":"10.1016\/j.infrared.2016.01.009"},{"key":"36_CR23","doi-asserted-by":"crossref","unstructured":"Bavirisetti, D.P.: Multi-sensor image fusion based on fourth order partial differential equations. In: 20th International Conference on Information Fusion (2017)","DOI":"10.23919\/ICIF.2017.8009719"},{"key":"36_CR24","doi-asserted-by":"crossref","unstructured":"Li, H., Wu, X.J.: Infrared and visible image fusion using Latent Low-Rank Representation (2018)","DOI":"10.1109\/ICPR.2018.8546006"},{"key":"36_CR25","unstructured":"Xu, H., Ma, J., Jiang, J., Guo, X., Ling, H.: U2Fusion: a unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 1\u20131 (2020)"},{"key":"36_CR26","doi-asserted-by":"publisher","first-page":"177","DOI":"10.1016\/j.inffus.2021.06.001","volume":"76","author":"H Xu","year":"2021","unstructured":"Xu, H., Ma, J.: EMFusion: an unsupervised enhanced medical image fusion network. Inform. Fusion 76, 177\u2013186 (2021)","journal-title":"Inform. Fusion"},{"key":"36_CR27","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1016\/j.inffus.2021.12.004","volume":"82","author":"L Tang","year":"2022","unstructured":"Tang, L., Yuan, J., Ma, J.: Image fusion in the loop of high-level vision tasks: a semantic-aware real-time infrared and visible image fusion network. Inform. Fusion 82, 28\u201342 (2022)","journal-title":"Inform. Fusion"},{"key":"36_CR28","doi-asserted-by":"crossref","unstructured":"Xu, H., Gong, M., Tian, X., Huang, J., Ma, J.: CUFD: an encoder\u2013decoder network for visible and infrared image fusion based on common and unique feature decomposition. Comput. Vis. Image Understand. 218 (2022)","DOI":"10.1016\/j.cviu.2022.103407"},{"key":"36_CR29","doi-asserted-by":"publisher","DOI":"10.1016\/j.compbiomed.2021.105005","volume":"141","author":"W Li","year":"2022","unstructured":"Li, W., Peng, X., Fu, J., Wang, G., Huang, Y., Chao, F.: A multiscale double-branch residual attention network for anatomical\u2013functional medical image fusion. Comput. Biol. Med. 141, 105005 (2022)","journal-title":"Comput. Biol. Med."}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-99-8558-6_36","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,6]],"date-time":"2024-11-06T20:52:47Z","timestamp":1730926367000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-99-8558-6_36"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,26]]},"ISBN":["9789819985579","9789819985586"],"references-count":29,"URL":"https:\/\/doi.org\/10.1007\/978-981-99-8558-6_36","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023,12,26]]},"assertion":[{"value":"26 December 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Xiamen","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"13 October 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 October 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/prcv2023.xmu.edu.cn\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1420","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"532","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"37% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3,78","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3,69","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}