Multimodal medical image fusion network based on target information enhancement
Glioma is a kind of brain disease with high incidence, high recurrence rate, high mortality, and low cure rate. To obtain accurate diagnosis results of brain glioma, doctors need to manually compare the imaging results of different modalities many times, which will increase the diagnosis time and re...
Saved in:
Published in: | IEEE access Vol. 12; p. 1 |
---|---|
Main Authors: | , , , |
Format: | Journal Article |
Language: | English |
Published: |
Piscataway
IEEE
01-01-2024
The Institute of Electrical and Electronics Engineers, Inc. (IEEE) |
Subjects: | |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | Glioma is a kind of brain disease with high incidence, high recurrence rate, high mortality, and low cure rate. To obtain accurate diagnosis results of brain glioma, doctors need to manually compare the imaging results of different modalities many times, which will increase the diagnosis time and reduce the diagnostic efficiency. Image fusion technology has been widely used in recent years to obtain information on multimodal medical images. This paper proposes a novel image fusion framework, target information enhanced image fusion network (TIEF), using cross-modal learning and information enhancement techniques. The framework consists of a multi-sequence feature extraction block, a feature selection block, and a fusion block. The multi-sequence feature extraction block consists of multiple sobel dense conv leaky ReLu block (SDCL-block). SDCL-block mainly realizes the extraction of edge features, shallow features, and deep features. The feature selection block identifies the feature channels with rich texture information and strong discrimination ability through the effective combination of global information entropy criterion and feature jump connection. The feature fusion block mainly comprises multi-head and spatial attention mechanisms, which can realize the fusion of intra-modality and inter-modality features. On this basis, considering the influence of tumor spatial location and structure information on the fusion results, a loss function is designed, which is a weighted combination of texture loss, structure loss, and saliency loss so that texture information from multimodal magnetic resonance imaging (MMRI) and saliency information from different anatomical structures of the brain can be fused at the same time to improve the expression ability of features. In this paper, the TIEF algorithm is trained and validated on the MMRI and (Single-Photon Emission Computed Tomography-MRI) SPECT-MRI datasets of glioma and generalized on the (Computed Tomography-MRI) CT-MRI dataset of meningioma to verify the performance of the TIEF algorithm. In the image fusion task, quantitative results showed that TIEF exhibited optimal or suboptimal performance in information entropy, spatial frequency, and average gradient metrics. Qualitative results indicate that the fused images can highlight tumor and edematous features. A downstream image segmentation task was used for evaluation to further verify TIEF's effectiveness. TIEF achieved the best results in both (Dice similarity coefficient) Dice and (Hausdorff distance 95%) HD95 segmentation metrics. In the generalization task, quantitative results indicated that TIEF obtained more information in the meningioma dataset. In conclusion, TIEF can effectively achieve cross-domain information acquisition and fusion and has robustness and generalization ability. |
---|---|
ISSN: | 2169-3536 2169-3536 |
DOI: | 10.1109/ACCESS.2024.3402965 |