Abstract
Tracking microsctructural changes in the developing brain relies on accurate inter-subject image registration. However, most methods rely on either structural or diffusion data to learn the spatial correspondences between two or more images, without taking into account the complementary information provided by using both. Here we propose a deep learning registration framework which combines the structural information provided by \(T_2\)-weighted (\(T_2\)w) images with the rich microstructural information offered by diffusion tensor imaging (DTI) scans. This allows our trained network to register pairs of images in a single pass. We perform a leave-one-out cross-validation study where we compare the performance of our multi-modality registration model with a baseline model trained on structural data only, in terms of Dice scores and differences in fractional anisotropy (FA) maps. Our results show that in terms of average Dice scores our model performs better in subcortical regions when compared to using structural data only. Moreover, average sum-of-squared differences between warped and fixed FA maps show that our proposed model performs better at aligning the diffusion data.
You have full access to this open access chapter, Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
Medical image registration is a vital component of a large number of clinical applications. For example, image registration is used to track longitudinal changes occurring in the brain. However, most applications in this field rely on a single modality, without taking into account the rich information provided by other modalities. Although \(T_2\)w magnetic resonance imaging (MRI) scans provide good contrast between different brain tissues, they do not have knowledge of the extent or location of white matter tracts. Moreover, during early life, the brain undergoes dramatic changes, such as cortical folding and myelination, processes which affect not only the brain’s shape, but also the MRI tissue contrast.
In order to establish correspondences between images acquired at different gestational ages, we propose a deep learning image registration framework which combines both \(T_2\)w and DTI scans. More specifically, we build a neural network starting from the popular diffeomorphic VoxelMorph framework [2], on which we add layers capable of dealing with diffusion tensor (DT) images. The key novelties in our proposed deep learning registration framework are:
-
The network is capable of dealing with higher-order data, such as DT images, by accounting for the change in orientation of diffusion tensors induced by the predicted deformation field.
-
During inference, our trained network can register pairs of \(T_2\)w images without the need to provide the extra microstructural information. This is helpful when higher-order data is missing in the test dataset.
Throughout this work we use 3-D MRI brain scans acquired as part of the developing Human Connectome ProjectFootnote 1 (dHCP). We showcase the capabilities of our proposed framework on images of infants born and scanned at different gestational ages and we compare the results against the baseline network trained on only \(T_2\)w images. Our results show that by using both modalities to drive the learning process we achieve superior alignment in subcortical regions and a better alignment of the white matter tracts.
2 Method
Let \(F, \, M\) represent the fixed (target) and the moving (source) magnetic resonance (MR) volumes, respectively, defined over the 3-D spatial domain \(\varOmega \), and let \(\phi \) be the deformation field. In this paper we focus on \(T_2\)w images (\(F^{T2w}\) and \(M^{T2w}\) which are single channel data) and DT images (\(F^{DTI}\) and \(M^{DTI}\) which are 6 channels data) acquired from the same subjects. Our aim is to align pairs of \(T_2\)w volumes using similarity metrics defined on both the \(T_2\)w and DTI data, while only using the structural data as input to the network.
In order to achieve this, we model a function \(g_{\theta }(F^{T2w}, M^{T2w}) = v\) a velocity field (with learnable parameters \(\theta \)) using a convolutional neural network (CNN) architecture based on VoxelMorph [2]. In addition to the baseline architecture, we construct layers capable of dealing with the higher-order data represented by our DT images. Throughout this work we use \(T_2\)w and DTI scans that have been affinely aligned to a common 40 weeks gestational age atlas space [14], prior to being used by the network.
Figure 1 shows the general architecture of the proposed network. During training, our model uses pairs of \(T_2\)w images to learn a velocity field v, while the squaring and scaling layers [2] transform it into a topology-preserving deformation field \(\phi \). The moving images M are warped by the deformation field using a SpatialTransform layer [5] which outputs the moved (linearly resampled) \(T_2\)w and DT images. The DT images are further processed to obtain the final moved and reoriented image.
The model is trained using stochastic gradient descent to find the optimal parameters \(\hat{\theta }\) that minimize a sum of three loss functions, representing the tensor similarity measure, the scalar-data similarity measure and a regulariser applied on the predicted deformation field. The DTI data is not used as input to our CNN, but only used to drive the learning process through calculating the similarity measure. During inference, our model uses only \(T_2\)w images to predict the deformation field, without the need for a second modality. In the following subsections, we describe our model in further detail.
Network Architecture. The baseline architecture of our network is a 3-D UNet [12] based on VoxelMorph [2]. The encoding branch is made up of four 3D convolutions of 16, 32, 32, and 32 filters, respectively, with a kernel size of \(3 \times 3 \times 3\), followed by Leaky ReLU (\(\alpha = 0.2\)) activations [18]. The decoding branch contains four transverse 3D convolutions of 32 filters each, with the same kernel size and activation function. Skip connections are used to concatenate the encoding branch’s information to the decoder branch. Two more convolutional layers, one with 16 filters and a second one with 3 filters, are added at the end, both with the same kernel size and activation function as before.
A pair of \(T_2\)w images are concatenated on the channel axis and become a \(96\,\times \,96\,\times \,64\,\times \,2\) input for the CNN network. The output is a three channel velocity field of the same size as the input images. The velocity field is smoothed with a \(3 \times 3 \times 3\) Gaussian kernel (with \(\sigma \,=\,1.2\) mm), and passed onto seven squaring and scaling layers [2], which transform it into a topology-preserving deformation field. The SpatialTransform layer [5] receives as input the predicted field \(\phi \) and the moving scalar-valued \(T_2\)w image, and outputs the warped and resampled image. A similar process is necessary to warp the moving DT image, with a few extra steps which are explained in the next subsection.
Tensor Reorientation. Registration of DT images is not as straightforward to perform as scalar-valued data. When transforming the latter, the intensities in the moving image are interpolated at the new locations determined by the deformation field \(\phi \) and copied to the corresponding location in the target image space. However, after interpolating DT images, the diffusion tensors need to be reoriented to remain anatomically correct [1]. In this work we use the finite strain (FS) strategy [1].
When the transformation is non-linear, such as in our case, the reorientation matrix can be computed at each point in the deformation field \(\phi \) through a polar decomposition of the local Jacobian matrix. This factorisation transforms the non-singular matrix J into a unitary matrix R (the pure rotation) and a positive-semidefinite Hermitian matrix P, such that \(J = RP\) [15]. The rotation matrices R are then used to reorient the tensors without changing the local microstructure.
Loss Function. We train our model using a loss function composed of three parts. First, the structural loss \(\mathcal {L}_{struct}\) (applied on the \(T_2\)w data only) is a popular similarity measure used in medical image registration, called normalised cross correlation (NCC). We define it as:
where \(\overline{F}\) is the mean voxel value in the fixed image F and \(\overline{M}\) is the mean voxel value in the transformed moving image \(M(\phi )\).
Second, to encourage a good alignment between the DT images, we set \(\mathcal {L}_{tensor}\) to be one of the most commonly used diffusion tensor similarity measures, known as the Euclidean distance squared. We define it as:
where the euclidean distance between two pairs of tensors \(\mathbf {D_1}\) and \(\mathbf {D_2}\) is defined as \(||\mathbf {D_1} - \mathbf {D_2} ||_C = \sqrt{ Tr ( ( \mathbf {D_1} - \mathbf {D_2})^2 ) }\) [19].
Finally, to ensure a smooth deformation field \(\phi \) we use a regularisation penalty \(\mathcal {L}_{reg}\) in the form of bending energy [13]:
Thus, the final loss function is:
We compare our network with a baseline trained on \(T_2\)w data only. For the latter case the loss function becomes: \(\mathcal {L} (F, M(\phi )) = \beta \, NCC(F^{T2w},M^{T2w}(\phi )) + \lambda \, BE(\phi )\). In all of our experiments we set the weights to \(\alpha = 1.0\), \(\beta = 1.0\) and \(\lambda = 0.001\) when using both DTI and \(T_2\)w images, and to \(\beta = 1.0\) and \(\lambda = 0.001\) when using \(T_2\)w data only. These hyper-parameters were found to be optimal on our validation set.
3 Experiments
Dataset. The image dataset used in this work is part of the developing Human Connectome Project. Both the \(T_2\)w images and the diffusion weighted (DW) images were acquired using a 3T Philips Achieva scanner and a 32-channels neonatal head coil [6]. The structural data was acquired using a turbo spin echo (TSE) sequence in two stacks of 2D slices (sagittal and axial planes), with parameters: \(T_R = 12\) s, \(T_E = 156\) ms, and SENSE factors of 2.11 for the axial plane and 2.58 for the sagittal plane. The data was subsequently corrected for motion [4, 8] and resampled to an isotropic voxel size of 0.5 mm.
The DW images were acquired using a monopolar spin echo echo-planar imaging (SE-EPI) Stejksal-Tanner sequence [7]. A multiband factor of 4 and a total of 64 interleaved overlapping slices (1.5 mm in-plane resolution, 3 mm thickness, 1.5 mm overlap) were used to acquire a single volume, with parameters \(T_R = 3800\) ms, \(T_E = 90\) ms. This data underwent outlier removal, motion correction and it was subsequently super-resolved to a 1.5 mm isotropic voxel resolution [3]. All resulting images were checked for abnormalities by a paediatric neuroradiologist.
For this study, we use a total of 368 \(T_2\)w and DT volumes of neonates born between 23–42 weeks gestational age (GA) and scanned at term-equivalent age (37–45 weeks GA). The age distribution in our dataset is found in Fig. 2, where GA at birth is shown in blue, and post-menstrual age (PMA) at scan is shown in orange. In order to use both the \(T_2\)w and DT volumes in our registration network, we first resampled the \(T_2\)w data into the DW space of 1.5 mm voxel resolution. Then, we affinely registered all of our data to a common 40 weeks gestational age atlas space [14] available in the MIRTKFootnote 2 software toolbox [13] and obtained the DT images using the dwi2tensor [17] command available in the MRTRIXFootnote 3 toolbox. Finally, we performed skull-stripping using the available dHCP brain masks [3] and we cropped the resulting images to a \(96\,\times \,96\,\times \,64\) volume size.
Training. We trained our models using the rectified Adam (RAdam) optimiser [9] with a cyclical learning rate [16] varying from \(10^{-9}\) to \(10^{-4}\), for 90, 000 iterations. Out of the 368 subjects in our entire dataset, 318 were used for training, 25 for validation and 25 for test. The subjects in each category were chosen such that their GA at birth and PMA at scan were distributed across the entire range. The validation set was used to help us choose the best hyperparameters for our network and the best performing models. The results reported in the next section are on the test set.
Final Model Results. In both our \(T_2\)w-only and \(T_2\)w+DTI cases we performed a leave-one-out cross-validation, where we aligned 24 of the test subjects to a single subject, and repeated until all the subjects were used as target. Each of the 25 subjects had tissue label segmentations (obtained using the Draw-EM pipeline for automatic brain MRI segmentation of the developing neonatal brain [10]) which were propagated using NiftyRegFootnote 4 [11] and the predicted deformation fields. The average resulting Dice scores are summarised in Fig. 3, where the initial pre-alignment is shown in pink, the \(T_2\)w-only results are shown in light blue and the \(T_2\)w+DTI are shown in purple. Our proposed model performs better than the baseline model for all subcortical structures (cerebellum, deep gray matter, brainstem and hippocampi and amygdala), while performing similarly well in white matter structures. In contrast, cortical gray matter regions were better aligned when using the \(T_2\)w-only model, as structural data has higher contrast than DTI in these areas.
We also computed the FA maps for all the initial affinely aligned and all the warped subjects in the cross-validation study and calculated the sum-of-squared differences (SSD) between the moved FA maps and the fixed FA maps. The resulting average values are summarised in Table 1, which shows that our proposed model achieved better alignment in terms of FA maps.
Finally, Fig. 4 shows two example registrations. The target images are from two term-born infants with GA = 40.86 weeks and PMA = 41.43 weeks, and GA = 40.57w and PMA = 41w, respectively, while the moving images are from infants with GA = 40.57 weeks and PMA = 41 weeks, and GA = 37.14w and PMA = 37.28w, respectively. The figure shows both \(T_2\)w and FA maps of axial slices of the fixed (first column), the moving (second column) and the warped images by our proposed method (third column) and the baseline method (fourth column), respectively. The moved FA maps show that by using DTI data to drive the learning process of a deep learning registration framework, we were able to achieve good alignment not only on the structural data, but also on the diffusion data as well.
4 Discussion and Future Work
In this work we showed for the first time a deep learning registration framework capable of aligning both structural (T2w) and microstructural (DTI) data, while using only \(T_2\)w data at inference time. A key result from our study is that our proposed \(T_2\)w+DTI model performed better in terms of aligning subcortical structures, even though the labels for these regions were obtained from structural data only. For future work we plan to focus on improving the registration in the cortical regions, and to compare our deep learning model with classic registration algorithms.
References
Alexander, D.C., Pierpaoli, C., Basser, P.J., Gee, J.C.: Spatial transformations of diffusion tensor magnetic resonance images. IEEE Trans. Med. Imaging 20(11), 1131–1139 (2001)
Balakrishnan, G., Zhao, A., Sabuncu, M.R., Guttag, J., Dalca, A.V.: Voxelmorph: a learning framework for deformable medical image registration. IEEE Trans. Med. Imaging 38(8), 1788–1800 (2019)
Christiaens, D., et al.: Scattered slice shard reconstruction for motion correction in multi-shell diffusion MRI of the neonatal brain. arXiv preprint arXiv:1905.02996 (2019)
Cordero-Grande, L., Hughes, E.J., Hutter, J., Price, A.N., Hajnal, J.V.: Three-dimensional motion corrected sensitivity encoding reconstruction for multi-shot multi-slice MRI: application to neonatal brain imaging. Magnet. Reson. Med. 79(3), 1365–1376 (2018)
Dalca, A.V., Balakrishnan, G., Guttag, J., Sabuncu, M.R.: Unsupervised learning for fast probabilistic diffeomorphic registration. In: Frangi, A.F., Schnabel, J.A., Davatzikos, C., Alberola-López, C., Fichtinger, G. (eds.) MICCAI 2018. LNCS, vol. 11070, pp. 729–738. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-00928-1_82
Hughes, E.J., et al.: A dedicated neonatal brain imaging system. Magnet. Reson. Med. 78(2), 794–804 (2017)
Hutter, J., et al.: Time-efficient and flexible design of optimized multishell hardi diffusion. Magnet. Reson. Med. 79(3), 1276–1292 (2018)
Kuklisova-Murgasova, M., Quaghebeur, G., Rutherford, M.A., Hajnal, J.V., Schnabel, J.A.: Reconstruction of fetal brain MRI with intensity matching and complete outlier removal. Med. Image Anal. 16(8), 1550–1564 (2012)
Liu, L., et al.: On the variance of the adaptive learning rate and beyond (2019)
Makropoulos, A., et al.: Automatic whole brain MRI segmentation of the developing neonatal brain. IEEE Trans. Med. Imaging 33(9), 1818–1831 (2014)
Modat, M., et al.: Fast free-form deformation using graphics processing units. Comput. Methods Programs Biomed. 98(3), 278–284 (2010)
Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234–241. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-24574-4_28
Rueckert, D., Sonoda, L.I., Hayes, C., Hill, D.L.G., Leach, M.O., Hawkes, D.J.: Nonrigid registration using free-form deformations: application to breast MR images. IEEE Trans. Med. Imaging 18(8), 712–721 (1999)
Schuh, A., et al.: Unbiased construction of a temporally consistent morphological atlas of neonatal brain development. bioRxiv (2018)
Shoemake, K., Duff, T.: Matrix animation and polar decomposition. In: Proceedings of the Conference on Graphics Interface, vol. 92, pp. 258–264. Citeseer (1992)
Smith, L.N.: Cyclical learning rates for training neural networks (2015)
Veraart, J., Sijbers, J., Sunaert, S., Leemans, A., Jeurissen, B.: Weighted linear least squares estimation of diffusion MRI parameters: strengths, limitations, and pitfalls. NeuroImage 81, 335–346 (2013)
Xu, B., Wang, N., Chen, T., Li, M.: Empirical evaluation of rectified activations in convolutional network (2015)
Zhang, H., Yushkevich, P.A., Alexander, D.C., Gee, J.C.: Deformable registration of diffusion tensor mr images with explicit orientation optimization. Med. Image Anal. 10(5), 764–785 (2006)
Acknowledgments
This work was supported by the Academy of Medical Sciences Springboard Award (SBF004\(\backslash \)1040), European Research Council under the European Union’s Seventh Framework Programme (FP7/ 20072013)/ERC grant agreement no. 319456 dHCP project, the Wellcome/EPSRC Centre for Medical Engineering at King’s College London (WT 203148/Z/16/Z), the NIHR Clinical Research Facility (CRF) at Guy’s and St Thomas’ and by the National Institute for Health Research Biomedical Research Centre based at Guy’s and St Thomas’ NHS Foundation Trust and King’s College London. The views expressed are those of the authors and not necessarily those of the NHS, the NIHR or the Department of Health.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Grigorescu, I. et al. (2020). Diffusion Tensor Driven Image Registration: A Deep Learning Approach. In: Špiclin, Ž., McClelland, J., Kybic, J., Goksel, O. (eds) Biomedical Image Registration. WBIR 2020. Lecture Notes in Computer Science(), vol 12120. Springer, Cham. https://doi.org/10.1007/978-3-030-50120-4_13
Download citation
DOI: https://doi.org/10.1007/978-3-030-50120-4_13
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-50119-8
Online ISBN: 978-3-030-50120-4
eBook Packages: Computer ScienceComputer Science (R0)