American Journal of Biomedical Engineering

p-ISSN: 2163-1050    e-ISSN: 2163-1077

2016;  6(2): 53-58

doi:10.5923/j.ajbe.20160602.02

 

The Image Registration Techniques for Medical Imaging (MRI-CT)

Hiba A. Mohammed

Department of Medical Engineering, University of Science and Technology, Omdurman, Sudan

Correspondence to: Hiba A. Mohammed , Department of Medical Engineering, University of Science and Technology, Omdurman, Sudan.

Email:

Copyright © 2016 Scientific & Academic Publishing. All Rights Reserved.

This work is licensed under the Creative Commons Attribution International License (CC BY).
http://creativecommons.org/licenses/by/4.0/

Abstract

Image registration is the process of combining two or more images for providing more information. Medical image fusion refers to the fusion of medical images obtained from different modalities. Medical image fusion helps in medical diagnosis by way of improving the quality of the images. In diagnosis, image obtained from a single modality like MRI, CT etc, maynot be able to provide all the required information. It is needed to combine information obtained from other modalities also to improve the information acquired. For example combination of information from MRI and CT modalities gives more information than the individual modalities separately. The aim is to provide a method for fusing the images from the individual modalities in such a way that the fusion results is an image that gives more information without any loss of the input information and without any redundancy or artefacts. In the fusion of medical images obtained from different modalities the images might be in different coordinate systems and have to be aligned properly for efficient fusion. The aligning of the input images before proceeding with the fusion is called image registration. The intensity based registration is carried out before decomposing the images. The two imaging modalities CT and MRI are considered for this thesis. The results on the CT and MRI images display the performance of the fusion algorithms in comparison with the registration.

Keywords: 2D/2D images, Wavelet, Image registration, Rigid images, Medical imaging

Cite this paper: Hiba A. Mohammed , The Image Registration Techniques for Medical Imaging (MRI-CT), American Journal of Biomedical Engineering, Vol. 6 No. 2, 2016, pp. 53-58. doi: 10.5923/j.ajbe.20160602.02.

1. Introduction

Medical images are increasingly being used in treatment planning and diagnosis. Medical devices such as Computerized Tomography (CT), Magnetic Resonance Imaging (MRI) are used to provide information about the anatomical structure of the organs while other devices Positron Emission Tomography (PET), Single Photon Emission Computerized Tomography (SPECT) provide the functional information. Each of these provides information that is useful in the treatment planning. In medical diagnosis and treatment, there are a wide interest in integrating information from multiple images using different modalities. This process is known as image registration. Image registration, also known as image fusion, matching or warping, can be defined as the process of aligning two or more images. The goal of an image registration method is to find the optimal transformation that best aligns the structures of interest in the input images. Image registration is a crucial step for image analysis in which valuable information is conveyed in more than one image; i.e., images acquired at different times, from distinct viewpoints or by different sensors can be complementary. Therefore, accurate integration (or fusion) of the useful information from two or more images is very important.
Image fusion means the combining of two images into a single image that has the maximum information content without producing details that are non-existent in the given images. With rapid advancements in technology, it is now possible to obtain information from multi source images to produce a high quality fused image with spatial and spectral information. Image Fusion is a mechanism to improve the quality of information from a set of images. The objective of this paper is to develop a method for rigid image registration of CT and MRI with wavelet image fusion.

2. Methodology

The proposed method was developed using the MATLAB tool box, for the simulation of the image registration and image fusion algorithms.
The data of DICOM images of CT and MRI are used for the implementation of the code were downloaded from the Math-works website. These multimodal images were processed prior to the application of the fusion algorithm. The program was performed in two stages: the first stage registration of DICOM images (CT-MRI) using intensity based registration, the reason of choosing this method is the simplest and least complicated, And the second stage the implementation of the fusion algorithm on the registered images using wavelet based image fusion, Also reason for choosing this method to achieve the best results possible by combing both techniques of registration and image fusion.

2.1. Image Registration

Intensity-based automatic image registration is an iterative process. It requires that you specify a pair of images, a metric, an optimizer, and a transformation type. In this case the pair of images is the MRI image (called the reference or fixed image) which is of size 512*512 and the CT image (called the moving or target image) which is of size 256*256. The metric is used to define the image similarity metric for evaluating the accuracy of the registration. This image similarity metric takes the two images with all the intensity values and returns a scalar value that describes how similar the images are. The optimizer then defines the methodology for minimizing or maximizing the similarity metric. The transformation type that used is rigid transformation (2-Dimension) that work translation and rotation for target image (brings the misaligned target image into alignment image with the reference image). Before the registration process can begin the two images (CT and MRI) need to be pre processed to get the best alignment results. After the pre processing phase the images are ready for alignment. The first step in the registration process was specifying the transform type with an internally determined transformation matrix. Together, they determine the specific image transformation that is applied to the moving image. Next, the metric compares the transformed moving image to the fixed image and a metric value is computed. Finally, the optimizer checks for a stop condition. In this case, the stop condition is the specified maximum number of iterations. If there is no stop condition, the optimizer adjusts the transformation matrix to begin the next iteration. And display the results of it in part of result.

2.2. Image Fusion

The next stage after the registration process is the wavelet based image fusion. Wavelets are finite duration oscillatory functions with a zero average value. They can be described by two functions the scaling (also known as the father) function, and the wavelet (also known as the mother) function. A number of basic functions can be used as the mother wavelet for Wavelet Transformations. The mother wavelet through translation and scaling produces various wavelet families which are used in the transformations.
The wavelets are chosen based on their shape and their ability to analyze the signal in a particular application. The Discrete Wavelet Transform has the property that the spatial resolution is small in low-frequency bands but large in high-frequency bands. This is because the scaling function is treated as a low pass filter and the wavelet function as a high pass filter in DWT implementation. The wavelet transform decomposes the image into low-high, high-low, high-high spatial frequency bands at different scales and the low-low band at the coarsest scale. The low-low image has the smallest spatial resolution and represents the approximation information of the original image. The other sub images show the detailed information of the original image. Different levels of decomposition can be applied as shown in the figure below.
Figure 1. Wavelet decomposition
Figure 2. Image fusing scheme using wavelet transforms
2.2.1. Implementation of Images Fusion
After the registration process was inclusive the second stage was implemented using the wavelet toolbox main menu. To open the wavelet toolbox main menu write wave menu in the Matlab command window and the wavelet toolbox will appear directly. Once the window is open there are several selections, here the specialized tools 2-D was used, which is also subdivided into four selections, the one used for the implementation is the image fusion. After the image fusion is selected it opens another separate window. The two images (registered image and reference image) from both image data sets were then imported into this window in order to be fused one data set at a time. On the left side of the window are the wavelet families and level of decomposition selections, once these selections were defined the images were decomposed. In this thesis the eighth order Daubechies (db) wavelet family was used and the third level decomposition was applied to obtain the coefficients which are called the approximation coefficients, horizontal detail coefficients, vertical detail & diagonal detail coefficients that represent the four frequency bands LL, LH, HL and HH respectively once these selections were defined the decompose button was tapped and the images were decomposed. These coefficients of both the images are then fused using different fusion methods (rules), on the image fusion window below the decompose button there is a fusion selection method option here are where the fusion rules applied to the images are selected, certain rules that can be used are Maximum value, Minimum value, Mean Value and Random value. These techniques respectively merge the two approximations or details structures obtained from the images to be fused, element wise by taking the maximum, minimum, mean or a randomly chosen element. Here the rules applied to the images are the mean-mean, max-max, min-min. Finally the inverse wavelet decomposition was applied which results in the final fused image. And display the results of the fusion process it in part of result.
2.2.2. Performance Metrics to Evaluate Image Fusion Techniques
The image fusion process should preserve all valid and useful information from the input images also it should not introduce undesired art facts. Various metrics are used in order to evaluate the performance of Image Fusion techniques. Examples of these metrics are: entropy, mean square error (MSE) and peak signal to noise ratio (PSNR).
A) Entropy
The Entropy (H) is the measure of information content in an image. The amaximum value of entropy can be produced when each gray level of the whole range has the same frequency. If the entropy of the fused image is higher than reference image then it indicates that the fused image contains more information.
(1)
B) Correlation Coefficient (CC)
The correlation coefficient is the measure the closeness or similarity in small size structures between the original and the fused images. It can vary between −1 and +l. Values closer to +1 indicate that the reference and fused images are highly similar while the values closer to −1 indicate that the images are highly dissimilar.
(2)
Where
(3)
(4)
(5)
C) Mean square error (MSE)
Is the squared of the difference divided by the number of elements in the image.
(6)
D) Peak signal to noise ratio (PSNR)
Is the ratio maximum possible power of a signal and the power of corrupting noise that affects the fidelity of the signal.
Here in this thesis the correlation coefficient (CC) was used as an image fusion evaluation criterion for the quantitative analysis of the image fusion results. The “corr2” MATLAB function was used to obtain the values of the correlation coefficient for the image fusion rules, the results of it in part of result.
(7)
(8)
(9)

3. Results and Discussion

3.1. Result of the Registration Phase

The DICOM CT and MRI images used to implement the code were downloaded from the math-works website. In this thesis two image data sets of the CT and MRI images were used for the implementation. These multimodal images were processed prior to the application of the fusion algorithm. The results of the preprocessing are shown below along with the original images before the preprocessing stage.

3.2. Pre processing Phase

The preprocessing phase generally means resizing the images, removing noise and improving or altering the image quality to suit a certain purpose. The first step in the preprocessing phase after reading and displaying the images into Mat lab was to resize the MRI (reference) image (a) to match the size of the CT (moving) image (b), and then several functions were used to complete the preprocessing task these step. Below are the results of the preprocessing along with the original images before preprocessing. From the images it can be seen that there is a difference between the original images (a), (b) and the images after the preprocessing (c), (d).
Figure 3. (a) Original MRI image, (b) Original CT image, (c) MRI image after pre processing, (d) CT image after pre processing

3.3. Implementation Result of the Images Registration

After the images were preprocessed they were ready for the registration, here the registration method applied was the automated intensity based registration which was discussed in the methodology. The similarity between the images was found using the correlation function and the value was displayed in percentage along with the elapsed time as mentioned before. The similarity of the images before and after the registration and the elapsed time were 71.7448%, 82.7396%, 6.3648s for the image data sets. Below are the results of the images before and after the registration. The results before the registration are shown in figures (a), the image in green represents the reference image and the image in violet represents the moving image. Figures (b) the results of the images after the registration.
Figure 4. (a) CT-MRI images before registration, (b) CT-MRI images after registration

3.4. Result of the Images Fusion

After the registration phase was done successfully the image fusion phase was implemented using the wavelet toolbox main menu.
3.4.1. Implementation Results of the Fusion
After the registration stage was complete the registered image and the reference image for both image data sets were imported into the 2-D image fusion window in order to be fused. The images were decomposed using the eighth order Daubechies wavelet transform and third level decomposition was applied. The decomposed images were then fused using different fusion rules and finally the inverse wavelet decomposition was applied to obtain the final fused image. The results of the different fusion methods are shown in the figures below. The figures (a), (b), (c) are the results of the image fusion rules for the image data sets. The performance of the image fusion rules applied to the image data sets was evaluated using the Correlation Coefficient (CC). The values of this evaluation are listed in the tables below. Table 1. shows the image fusion rules of the image data sets. The results of the evaluation using the CC have shown very good performance.
Figure 5. (a) Fused image CT-MRI using mean-mean fusion rule, (b) Fused image CT-MRI using max-max fusion rule , (c) Fused image CT-MRI using min-min fusion rule
Table 1. Shows the image fusion rules and corresponding Correlation Coefficient (CC) values for the image data set
     

4. Conclusions

The rigid image registration using wavelet based image fusion was performed using CT and MRI images. The quality of the images obtained by this technique has been verified using the Correlation Coefficient. The original input images and their corresponding registration and fusion results using the proposed technique are depicted in detail. The intensity based registration method was applied to the two input images, the quality of the registration was measured using the Correlation function to measure the similarity between the images before and after the registration the values of the correlation were displayed in percentage and the elapsed time taken for the registration process was also displayed, then the wavelet based image fusion algorithm was applied to the registered and reference images, the images were then decomposed and certain fusion rules were applied, finally the inverse decomposition was done to obtain the final fused images that contain information from both images. The quality of the fused images using the different fusion rules was done by measuring the Correlation Coefficient for each image fusion rule, and the resultant images are shown in the part of the result.

References

[1]  http://www.fda.gov/Radiation-EmittingProducts/RadiationEmittingProductsandProcedures/ MedicalImaging/ucm2005914.htm. Accessed on April, 2015.
[2]  http://www.medicalimaging.org/about-mita/medical-imaging-primer/.Accessed on April, 2015.
[3]  MATLAB tool for nuclear medicine imaging, Maria Lyra, Agapi Ploussi and Antonios Georgantzoglou.
[4]  Francisco P. M. Oliveira and Joao Manuel R. S. Tavares, “Medical image registration: a Review.” Computer methods in Biomechanics and Biomedical engineering, September 2012.
[5]  J. B. Antoine Maintz1 and Max A. Viergever. “An Overview of Medical Image Registration Methods”.
[6]  Pandav, Reena S., and S R. Khot. "Registration of CT And MR Images in Image Fusion." International Journal of Electrical, Electronics and Computer Systems 3.2 (2015).
[7]  Deepak Kumar Sahu, and M P. Parsai. "Different Image Fusion Techniques – A Critical Review." International Journal of Modern Engineering Research 2.5 (2012).
[8]  ParekhParmit, Nehal Patel, Robinson Macwan, Priteshkumar Prajapati, and SaritaVisavalia. "Comparative Study and Analysis of Medical Image Fusion Techniques." International Journal of Computer Applications (0975 – 8887) 90.19 (2014).Web. 2015.
[9]  Techniques for Accelerating Intensity-based Rigid Image Registration
[10]  J. P.W. Pluim, J.B.A. Maintz and M.A. Viergever “Image registration by maximization of combined mutual information and gradient information,” IEEE Trans. on Medical Imaging, Vol.19, No.8, Aug. 2000.
[11]  L. Ding, A. Goshtasby, and M. Satter, “Volume image registration by template matching” Image and Vision Computing, Vol. 19, No. 12, 2001.
[12]  B. Zitova and J. Flusser “Image registration methods: a survey,” Image and Vision Computing, Vol 21, 2003.
[13]  JAUN, DU, SONGYUAN TANG, TIANZI JIANG, and ZHENSU LU. "Intensity-based robust similarity for multimodal image registration." International Journal of Computer Mathematics 83.No.1 (2006).
[14]  Lifeng, Shang, Jian Cheng Lv,, and Zhang Yi. "Rigid medical image registration using PCA neural network." Neurocomputing 69 (2006).
[15]  Demetrios Gerogiannis, Christophoros Nikou, and Aristidis Likas. "Rigid Image Registration based on Pixel Grouping." 14th International Conference on Image Analysis and Processing (2007).
[16]  R. Suganya, Dr. S. Rajaram and K. Priyadharsini “Intensity Based Image Registration by Maximization of Mutual Information,” International Journal of Computer Applications. Vol 1, No. 20, 2010.
[17]  Nagham E. Mekky, F. E.-Z. Abou-Chadi, and S. Kishk, Wavelet-based image registration techniques: a study of performance, International Journal of Computer Science and Network Security IJCSNS, vol. 11, no. 2, February 2011.
[18]  J.Hu, Y. Yang and Z. Su, “A Novel Hierarchical Medical Image Registration Method Based on Multi-scale and Contour line”, Proceeding of Int. Conf. on Systems and Informatics, Yantai, 2012.
[19]  Ayyachamy, Swarnambiga, and Vasuki Mannivannan. "Medical Image Registration Based Retrieval." Applied Medical Informatics 32.No.1.
[20]  Sangeeta Sahu and Umesh Chandra Pati “Rigid medical image registration for monomodal images,” Proceedings of Int. Conf. on Mathematics and Eng. Sciences (ICMES), Kalka, Vol. 2, No. 1, March 2014.
[21]  Paulson, Christopher, Soundararajan Ezekiel, and Dapeng Wu. “Wavelet-Based Image Registration”.
[22]  Maddaiah, PN, PN Pournami, and VK Govindan. "Optimization of Image Registration for Medical Image Analysis." International Journal of Computer Science and Information Technologies 5.3 (2014).
[23]  Mrs. Disha S. Bhosle, and Mrs. Kanchan S. Gorde. "Image Registration and Wavelet Based Hybrid Image Fusion." IOSR Journal of VLSI and Signal Processing 4.2 (2014).
[24]  France Lalibert, Langis Gagnon ,” Registration And Fusion of Retinal images-An evaluation study,”IEEE Transaction on Medical Imaging, Vol. 22, No. 5, May 2003.
[25]  Myungjin Choi,” A New Intensity-Hue-Saturation Fusion Approach Image Fusion With a Tradeoff Parameter,” IEEE Transaction on Geoscience and Remote Sensing, Vol. 44, No. 6, June 2006.
[26]  Krista Amolins, Yun Zhang, Peter Dare, “Wavelet based Image fusion techniques – An introduction review and comparison,” ISPRS Journal of Photogrammetry & Remote Sensing 62, 2007.
[27]  V.P.S Naidu and J.R Raol “Pixel Leevel Image Fusion using wavelet nad Prinicipal Component Analysis,” Defence Science Journal, Vol. 58, No. 3, May 2008.
[28]  Susmitha Vekkot, and Pancham Shukla, “A Novel Architecture of Wavelet Based Image Fusion, World Academy of Science, Engineering and Technology 57, 2009.
[29]  A. Umamaheshvari, K. Thanushkodi, “Image Fusion Techniques,” IJRRAS 4(1), July 2010.
[30]  N. Indhumani, G. Padmavathi, “Enhanced Image Fusion Algorithm Using Laplace Pyramids and Spatial Frequency Based Wavelet Algorithm,” International Journal of Soft Computing and Engineering, Vol. 1, Issue 5, Nov. 2011.
[31]  Firouz Abdullah Al-Wassai, N.V. Kalyankar, Ali A. Ali-Zuky; “The IHS Transformation Based Image fusion,” Computer Vision and Pattern Recognition, July 2011.
[32]  M.A. Mohamed and B.M. El Den 2, “Implementation of image fusion Techniques for Multifocus Images Using FPGA,” National Radio Science Conference, April 2011
[33]  Hamid Reza Shahdoosti, Hassan Ghassemian, “Spatial PCA as A New Method for Image Fusion,” The 16th CSI International Symposium on Artificial Intelligence and Signal Processing (AISP), 2012.
[34]  Deepak Kumar Sahu, M.P. Parsai, “Different Image Fusion Techniques-A Critical Review,” International Journal of Modern Engineering Research, Vol.2,Issue 5Sept.-Oct. 2012.
[35]  Xiao Xiang Zhu, “A Sparse Image Fusion Algorithm with Application Of Pan sharpening,” IEEE Transactions on Geoscience And Remote Sensing, Vol. 51, No. 5, May 2013.
[36]  Kusum Rani, Reecha Sharma, “Study of Different Image Fusion Algorithm,” International Journal of Emerging Technology and Advanced Engineering (IJETAE), Vol. 3, Issue 5, May 2013.
[37]  Simrandeep Singh, Narwant Singh Grewal, Harbinder Singh, “Multiresolution Representation of Multifocus Image fusion Using Gaussian and Laplacian Pyramids,” International Journal of Advanced Research in Computer Science and Software Engineering (IJARCSSE), Vol. 3, Issue 11, Nov. 2013.
[38]  Nisha Gawari, Dr. Lalitha. Y.S., “Comparitive Analysis of PCA, DCT & DWT based Image Fusion Techniques,” Interanational Journal of Emerging Research in Mangement And Technology, Vol. 3, Issue 5, May 2014.
[39]  Deepali Sale, Varsha Patil, Dr. Madhuri A. Joshi, “Effective Image Enhancement using Hybrid Multi- resolution Image Fusion, IEEE global Conference on Wireless Computing and Networking (GCWCN), 2014.