This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
1. Introduction
With the development of imaging equipment and technology, there are different modality medical images which may reflect different organ or tissue information. For example, the computed tomography (CT) image can precisely exhibit dense structures such as bones and implants, while the magnetic resonance (MR) images detect enough soft tissue information with high-resolution anatomical details but are less sensitive to the diagnosis of fractures than CT. To obtain sufficient image information for accurate diagnosis, a doctor often needs to sequentially analyze different modality images, but this one by one manner way still brings inconvenience in many cases. The aim of the medical image fusion method is to generate a single comprehensive image contained in multiple medical images with different modalities, which are more suitable for doctor diagnosis. In fact, the fusion method can not only provide diagnosis information for doctors but also provide auxiliary treatment information [1, 2].
Over the last few years, a variety of medical image fusion methods have been proposed in various clinical applications. According to a recent survey [3], there are three categories based on multiscale decomposition (MSD) methods, based on learning representation methods, and based on combination of different methods. Classical MSD-based fusion methods [4–7] assume that the salient information of the source images is contained in the decomposition coefficients. Obviously, the selection of transform methods and decomposition levels are very important. Li et al. [8] proposed a comparative study of different MSD-based methods, where they found that the fusion method based on NSCT can generally obtain the best fusion effects. Based on learning representation methods include spares representation (SR) [9, 10], parameter-adaptive pulse-coupled neural network (PAPCNN) [11], convolutional sparse representation (CSR) [12], convolutional neural network (CNN) [13], convolutional sparsity-based morphological component analysis (CSMCA) [14], and deep learning (DL) [15, 16]. These methods represent the information of the source image by learning dictionary or learning model. Compared with the MSD-based fusion methods, the DL-based methods can achieve better results. The methods based on combination of different methods overcome the shortcomings of the singe method. For an example, the fusion method based on MSD (the high-pass bands are merged with the “max-absolute” rule, while the low-pass bands are fused using the “averaging” rule) has two main drawbacks: the loss of contrast and the difficulty of selecting level.
Yang et al. [9] first introduced the SR theory into image fusion field. The early fusion methods based on SR use the standard sparse coding which is applied local patches. After this practice, there are many image fusion methods based on SR. These fusion methods generally try to improve fusion performance by adding constraints [17, 18] and designing the effective strategies of learning dictionary [19]. The standard SR methods often have three defects: the details tend to be smoothed, spatial inconsistency, and the low computational efficiency. To solve these problems, there are many effective improved algorithms which aim at learning a compact and efficient dictionary. Qiu et al. [20] learnt a discriminative dictionary using the mutual information rule. In order to improve the localization and recognition of multiple objects, Siyahjani and Doretto [21] proposed a context aware dictionary. Qi et al. [22] learnt an integrated dictionary which used an entropy-based algorithm for informative block selection. They used the online dictionary learning algorithm to extract discriminative features from high-frequency components, which enhances the accuracy and efficiency of the fusion result.
Convolutional sparsity representation (CSR) model, unlike the standard SR model [7], is based on single image and overlapping patches in the original spatial domain for sparse coding; it is a global SR model of the source image. In addition, the sparse coding of the CSR model is performed over the entire image rather than on overlapping patches. Therefore, the fusion method based on CSR achieves better representations.
The representation of different components of the image has become a hot topic in recent research. Any image
A fast cartoon-texture decomposition (FCTD) [24] applied a pair of low-high pass filters; therefore, it is fast and simple. However, it blurs strong edges and retains certain textures in the cartoon part. One of the reasons for these results is that the edge maps computed are used as a local gradient computation which utilizes a few pixels around the central pixel. The local gradient operator is inaccurate for the noise image. In order to improve the stability of gradient, we use the global sparse gradient (GSG) [25] instead of local operators to improve FCTD. GSG uses more information around the central pixel. It is more stable for noise. Figure 1 shows an example of various gradient operators and the GSG on a noisy image
[figures omitted; refer to PDF]
Figure 2 shows an example of cartoon + texture decomposition for medical images using the FCTD method and the IFCTD method, respectively. Figures 2(c)–2(f) show the decomposition results of FCTD, and 2(g)–2(j) give the decomposition results of IFCTD, respectively. 2(c) and 2(g) are cartoon parts of CT images; 2(e) and 2(i) are texture parts of CT images; 2(d) and 2(h) are cartoon parts of MR images; and 2(f) and 2(j) are texture parts of MR images. From the experimental results, the result of IFCTD can extract details better than the result of FCTD. In the amplification texture parts of the MRI, the result of the IFCTD (Figure 2(j)) contains more texture details than the result of the FCTD (Figure 2(f)).
[figures omitted; refer to PDF]
In addition to using the IFCTD tool instead of the prelearned dictionaries, in order to better protect the energy of the cartoon part in the fusion process, we use the fusion rule of energy protection (WEL and WSEML) in the cartoon part and use the fusion rule of the texture part of CSR.
The main contribution of this paper is to introduce the IFCTD into the cartoon + texture decomposition for medical images and combine the energy protection method and CSR method to improve the fusion effects.
The rest of this paper is organized as follows. In Section 2, the CSR model is briefly introduced. Section 3 describes the proposed method in detail. Section 4 presents experiments and discussion. Finally, the conclusions are reported in Section 5.
2. CSR Model
SR-based image fusion method was first introduced by Yang and Li [9]. In this model, source images were divided into a large number of overlapping patches using the sliding window technique, and then the “max L1-norm” of the sparse coefficient vector was selected as the activity level measurement. SR has been widely used in image fusion. These methods have achieved great success. However, it is worthwhile to notice that these methods have some defects, such as (1) the SR-based methods are only shift-invariant when the stride of patches is one pixel in both vertical and horizontal directions, (2) the fine details in source images like textures and edges tend to be smoothed, (3) the “max-L1” rule may cause spatial inconsistency in the fused results for different modality images, and (4) the computational efficiency is low because the sliding window’s step length should be small enough in the sparse coding technique. In a word, the reasons for these defects are caused by patch-based coding which is performed on overlapping patches to achieve better representations. In order to solve these problems, CSR mode is introduced by Liu et al. [12]. The sparse coding is a global sparse representation, which is performed over the entire image.
CSR model can be seen as an alternative representation to SR using the convolutional form. This model can be formulated as the sum over a set of
3. Proposed Fusion Method
Suppose that there are two pre-registered source images denoted as
(1)
Use the low-pass filter
where
The local total variation (LTV) is obtained by convolution with the gradient norm of
Set
(2)
We obtain the cartoon image
In these two parts, one of the key elements is how to choose the appropriate fusion rules to improve the fusion effect.
3.1. Fusion Ruler of Cartoon Parts
In IFCTD,
Finally, the strategy of cartoon part fusion is
3.2. Fusion Ruler of Textural Parts
Suppose a set of dictionary filters
Let
Then, the “choose-max” rule is applied to achieve the fused coefficient maps:
The fused texture part can be expressed as
3.3. Final Fusion Results
In the process of cartoon and texture decomposition, the texture part is obtained by subtracting the cartoon part from the source image, so the final fusion result is still obtained by simply stacking the fused cartoon part and texture part.
Figure 4 shows the fusion flowchart of the proposed method.
[figure omitted; refer to PDF]4. Experimental Results and Analysis
4.1. Testing Images
In our experiments, the eight pairs of medical images are used as test images which are collected from Yu Liu’s personal homepage (http://home.ustc.edu.cn/∼liuyu1/) and the website web page: http://www.imagefusion.org/, Figure 5. The first line is the CT images, and the second line is the corresponding MR images. We assumed that each pair of source images is pre-registered.
4.2. Objective Evaluation Metrics of Image Fusion Effect
To measure the performance of the algorithm, five popular objective metrics are applied to evaluate the fusion results from different aspects. They are the entropy (EN), standard deviation (SD), normalized mutual information (MI) [28], gradient-based fusion metric
[figures omitted; refer to PDF]
4.3. Experimental Discussion
Because the proposed method, namely, CTCSR, is mainly aimed at the improvement of the SR-based methods, the comparison algorithm chooses the fusion method based on SR, such as standard SR [9], a hybrid cartoon texture sparse representation method (BSR) [10], PAPCNN [11], CSR [12], and CSMCA [14]. All the parameters are set to the recommended values as reported in [9–12, 14]. Apart from PAPCNN, the dictionaries of other SR-based methods have 256 atoms and are learned by the K-SVD method from natural image patches. For the CTCSR method, the spatial size of each dictionary filter is set to
All of the fusion methods are implemented on the platform of the HP-Z600 workshop (Four Core 2.4 GHz CPU and 8G RAM), Matlab R2017b programming environment in the Windows 7 operating systems.
Figures 6–9 show four examples with different fusion methods for CT and MR images. We magnify the marked red rectangle part of the experimental results and show them the marked green rectangle part.
[figures omitted; refer to PDF]
[figures omitted; refer to PDF]
[figures omitted; refer to PDF]
[figures omitted; refer to PDF]
In Figure 6, it can be seen that the performances of the SR and BSR methods suffer from obvious undesirable visual artifacts. The effects of the PAPCNN, CSR, and CTCSR methods are basically the same; these methods enhance the anatomical details (in the enlarged part). The CSMAC method is relatively low contrast.
Figure 7 shows a set of C1 and C2 image fusion results. Because structural details are mainly contained in the MR image, almost all of these methods can extract details well. But, from the partially enlarged image, the details are seriously blurred in the SR method. The BSR, PAPCNN, and CSR methods lose a lot of information of
Figure 8 shows a set of E1 and E2 image fusion results. The BSR, PAPCNN, and CTCSR methods have almost the same visual effect. There are artificial defects in the result of CSR. The CSMCA method reduces the contrast of the fusion image. The details are seriously blurred in the SR method.
Figure 9 gives a set of H1 and H2 image fusion results. The BSR and CTCSR methods not only keep the brightness of the bone but also contain rich soft tissue information. They have good visual effect. There are artificial defects in the result of CSR. The CSMCA and CSR methods reduce the contrast of the fusion image. The result of the SR based-method lost a lot of details.
Table 1 lists the average objective metrics of different fusion methods on eight sets of CT and MR images. For each metric, the biggest value given in bold indicates the best results among all the methods. Overall, the proposed method shows the best performances on SD, MI,QG and QP. These metric values reflect the high robustness of the proposed method. It is further confirmed that the proposed method can achieve better fusion effect.
Table 1
The average objective metrics of SR, BSR, PAPCNN, CSR, CSMCA, and CTCSR methods on pairs of medical images.
SR | BSR | PAPCNN | CSR | CSMCA | CTCSR | |
EN | 5.7277 | 5.7636 | 6.1723 | 6.1707 | 5.5939 | 5.9027 |
SD | 65.3614 | 66.6342 | 73.3334 | 71.6315 | 62.1341 | 75.5059 |
MI | 0.7729 | 0.7720 | 0.6135 | 0.6523 | 0.6439 | 0.7947 |
0.5596 | 0.5767 | 0.5478 | 0.5928 | 0.5945 | 0.5987 | |
0.3732 | 0.40376 | 0.3849 | 0.4938 | 0.5301 | 0.5421 |
5. Conclusion
In this paper, a fusion method based on the cartoon-texture decomposition method and convolution sparse representation theory is proposed for medical images. The fusion rules of energy protection and feature extraction are used in the cartoon part, while the fusion method of convolution sparse representation is used in the texture part. Different fusion rules are selected in different feature parts, which can better represent image information and achieve better fusion effect. The experimental results show that the proposed algorithm is effective in terms of visual quality and objective metric values.
Acknowledgments
The authors thank first-class discipline construction in Ningxia Institutions of Higher Learning (Pedagogy) (Grant NXYLXK2017B11), the National Natural Science Foundation of China (Grants 61772389, 61972264, and 61971005), General Projects in Guyuan (Grant 2019GKGY041), and Key Research and Development Projects of Ningxia Autonomous Region (Talent Introduction Program) (2019BEB04021) for supporting our research work.
[1] A. Mauro, "A generalised porous medium approach to study thermo-fluid dynamics in human eyes," Medical & Biological Engineering & Computing Journal of the International Federation for Medical & Biological Engineering, vol. 34, 2018.
[2] A. Mauro, V. Romano, P. Nithiarasu, "Suprachoroidal shunts for treatment of glaucoma," International Journal of Numerical Methods for Heat & Fluid Flow, vol. 28 no. 2, pp. 297-314, DOI: 10.1108/hff-12-2016-0508, 2018.
[3] Y. Liu, X. Chen, R. K. Wang, X. Wang, "Deep learning for pixel-level image fusion: recent advances and future prospects," Information Fusion, vol. 42, pp. 158-173, DOI: 10.1016/j.inffus.2017.10, 2018.
[4] Z. Liu, K. Tsukada, K. Hanasaki, Y. P. Ho, "Image fusion by using steerable pyramid," Pattern Recognition Letters, vol. 22 no. 9, pp. 929-939, DOI: 10.1016/s0167-8655(01)00047-2, 2001.
[5] L. J. Chipman, T. M. Orr, L. N. Graham, "Wavelets and image fusion," Proceedings of the International Conference on Image Processing, .
[6] J. J. Lewis, R. J. O’Callaghan, S. G. Nikolov, D. R. Bull, "Pixel- and region-based image fusion with complex wavelets," Information Fusion, vol. 8 no. 2, pp. 119-130, DOI: 10.1016/j.inffus.2005.09.006, 2007.
[7] Q. Canagarajah, B.-L. Guo, "Multifocus image fusion using the nonsubsampled contourlet transform," Signal Processing, vol. 89 no. 7, pp. 1334-1346, DOI: 10.1016/j.sigpro.2009.01.012, 2009.
[8] S. Li, B. Yang, "Hybrid multiresolution method for multisensor multimodal image fusion," Sensors Journal, IEEE, vol. 10, pp. 1519-1526, 2010.
[9] B. Yang, S. Li, "Multifocus image fusion and restoration with sparse representation," IEEE Transactions on Instrumentation and Measurement, vol. 59 no. 4, pp. 884-892, 2010.
[10] L. Chang, X. Feng, X. Zhu, "CT and MRI image fusion based on multiscale decomposition method and hybrid approach," IET Image Processing, vol. 13 no. 1, pp. 83-88, 2018.
[11] Y. Ming, L. Xiaoning, L. Yu, "Medical image fusion with parameter adaptive pulse coupled-neural network in nonsubsampled shearlet transform domain," IEEE Transactions on Instrumentation and Measurement, vol. 34, 2018.
[12] Y. Liu, X. Chen, R. K. Ward, Z. J. Wang, "Image fusion with convolutional sparse representation," IEEE Signal Processing Letters, vol. 23 no. 12, pp. 1882-1886, 2016.
[13] Y. Liu, X. Chen, J. Cheng, H. Peng, "A medical image fusion method based on convolutional neural networks," Proceedings of 20th International Conference on Information Fusion, vol. 35, 2017.
[14] Y. Liu, X. Chen, R. K. Ward, "Medical image fusion via convolutional sparsity based morphological component analysis," IEEE Signal Processing Letters, vol. 56, 2019.
[15] A. Azarang, H. G. hassemian, "A new pansharpening method using multiresolution analysis framework and deep neural networks," Proceedings of 3rd International Conference on Pattern Recognition and Image Analysis, vol. 56, 2017.
[16] Y. Liu, X. Chen, H. Peng, Z. Wang, "Multi-focus image fusion with a deep convolutional neural network," Information Fusion, vol. 36, pp. 191-207, DOI: 10.1016/j.inffus.2016.12.001, 2017.
[17] Y. Liu, S. Liu, Z. Wang, "A general framework for image fusion based on multi-scale transform and sparse representation," Information Fusion, vol. 24, pp. 147-164, DOI: 10.1016/j.inffus.2014.09.004, 2015.
[18] H. Li, X. He, D. Tao, Y. Tang, R. Wang, "Joint medical image fusion, denoising and enhancement via discriminative low-rank sparse dictionaries learning," Pattern Recognition, vol. 79, pp. 130-146, DOI: 10.1016/j.patcog.2018.02.005, 2018.
[19] M. Kim, D. K. Han, H. Ko, "Joint patch clustering-based dictionary learning for multimodal image fusion," Information Fusion, vol. 27, pp. 198-214, DOI: 10.1016/j.inffus.2015.03.003, 2016.
[20] Q. Qiu, Z. Jiang, R. Chellappa, "Joint patch clustering-based dictionary learning for multimodal image fusion," Proceedings of the 2011 IEEE International Conference on Computer Vision (ICCV), .
[21] F. Siyahjani, G. Doretto, "Learning a context aware dictionary for sparse representation," Lecture Notes in Computer Science Book Series, pp. 228-241, 2013.
[22] G. Qi, J. Wang, Q. Zhang, F. Zeng, Z. Zhu, "An integrated dictionary-learning entropy-based medical image fusion framework," Future Internet, vol. 9 no. 4,DOI: 10.3390/fi9040061, 2017.
[23] Z. Zhu, H. Yin, Y. Chai, Y. Li, G. Qi, "A novel multi-modality image fusion method based on image decomposition and sparse representation," Information Sciences, vol. 432, pp. 516-529, 2018.
[24] A. Buades, J.-L. Lisani, "Directional filters for cartoon + texture image decomposition," Image Processing on Line, vol. 5, pp. 75-88, DOI: 10.5201/ipol.2016.165, 2016.
[25] R. Zhang, X. Feng, S. Wang, L. Chang, "A sparse gradients field based image denoising algorithm via non-local means," Acta Automatica Sinica, vol. 14 no. 9, pp. 1542-1548, 2015.
[26] B. Wohlberg, "Efficient algorithms for convolutional sparse representationsficient algorithms for convolutional sparse representation," IEEE Transactions on Image Processing, vol. 25 no. 1, pp. 301-315, DOI: 10.1109/tip.2015.2495260, 2016.
[27] M. Morup, M. Schmidt, "Transformation invariant sparse coding," .
[28] M. Hossny, S. Nahavandi, D. Creighton, "Comments on 'Information measure for performance of image fusion," Electronics Letters, vol. 44 no. 18, pp. 1066-1067, DOI: 10.1049/el:20081754, 2008.
[29] C. S. Xydeas, V. Petrović, "Objective image fusion performance measure," Electronics Letters, vol. 36 no. 4, pp. 308-309, DOI: 10.1049/el:20000267, 2000.
[30] Z. Liu, E. Blasch, Z. Xue, J. Zhao, R. au, W. Wu, "assessment of multiresolution image fusion algorithms for context enhancement in night vision: a comparative study," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 34 no. 1, pp. 94-109, DOI: 10.1109/tpami.2011.109, 2012.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Copyright © 2020 Lihong Chang et al. This is an open access article distributed under the Creative Commons Attribution License (the “License”), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. http://creativecommons.org/licenses/by/4.0/
Abstract
A fusion method based on the cartoon+texture decomposition method and convolution sparse representation theory is proposed for medical images. It can be divided into three steps: firstly, the cartoon and texture parts are obtained using the improved cartoon-texture decomposition method. Secondly, the fusion rules of energy protection and feature extraction are used in the cartoon part, while the fusion method of convolution sparse representation is used in the texture part. Finally, the fused image is obtained using superimposing the fused cartoon and texture parts. Experiments show that the proposed algorithm is effective.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer