Abstract:In order to reduce the spatial detail loss of image fusion in the feedforward process of deep network, a medical image fusion method based on conditional reversible neural network (CINN) is proposed. Through the application of reversible analysis-synthesis architecture, the multi-modal fusion of spatial details and key semantic complementarity is realized. In the forward analysis phase, multi-resolution features are embedded into CINN as conditional vectors to realize multi-modal representation learning. In the reverse synthesis stage, a wavelet-based conditional fusion (WCF) network is used to guide CINN to complete the reverse reconstruction. In feature fusion, the correlation activation template (RAM) is applied to focus on the consistency information of key structural regions and texture details in multi-modal medical images. The combination of forward analysis and reverse reconstruction is constructed to optimize network parameters efficiently to obtain high quality fusion images. CT-MRI and MRI-PET scenes were tested. Compared with the existing fusion baseline, the proposed method improved the performance of objective fusion indicators such as SCD and VIFF by 15.16% and 46.53%, respectively, achieved superior results in subjective visual quality.