[edit]
LightFusionNet: Lightweight Dual-Stream Network with Predictive Context Attention for Efficient Medical Image Fusion
Proceedings of The Second AAAI Bridge Program on AI for Medicine and Healthcare, PMLR 317:210-218, 2026.
Abstract
Multimodal image fusion aims to integrate complementary information from multiple imaging modalities into a single, informative representation, which is crucial for applications in medical imaging and microscopy. Existing methods often face trade-offs between structural fidelity, edge preservation, and computational efficiency. In this work, we propose Light-FusionNet, a lightweight dual-stream network designed to efficiently fuse multimodal images while retaining key structural, textural, and intensity features. The network leverages depthwise separable convolutions to reduce model complexity and incorporates a Predictive Context Attention (PCA) mechanism to selectively emphasize informative regions in the feature maps. Extensive experiments on benchmark medical imaging datasets, including PET-MRI, SPECT-MRI, and CT-MRI, demonstrate that our approach achieves comparable qualitative and quantitative performance compared to state-of-the-art fusion methods, while maintaining low computational cost. The proposed method provides an effective and efficient solution for multimodal image fusion, suitable for both clinical and research applications.