This is a blog post for the paper ‘Data Augmentation Using Learned Transformations for One-shot Medical Image Segmentation’.
Written by Amy Zhao, Guha Balakrishnan, Fredo Durand, John V. Guttag, Adrian V. Dalca.

Introduction and Problem Statement

Semantic segmentation is an important task in medical image analysis since it may be used in various application settings, including disease diagnosis and treatment planning [6]. A wide range of methods is used to improve the segmentation accuracy including supervised [1], semi-supervised [2], and unsupervised [3] learning approaches. However, the current state-of-the-art results are achieved with supervised learning models that require a big chunk of manually annotated data. Collecting and labeling medical images is not only a laborious task, but it is also complicated by the discrepancies in image intensities, resolution, and subject positioning, caused by variation in individual subjects and the equipment used [4].

Data Augmentation

Data augmentation techniques can be used in order to increase the size of the training dataset. It is common practice to use simple transformations to the available data, such as rotations, translations, or random crops [9]. However, these augmentation schemes rely on hand-tuned parameters which makes data augmentation an implicit form of parameter tuning. Also, as biomedical images differ significantly in anatomy, contrast, and texture, simple transformations are insufficient to model such complex variations.

Transformation Models

Recently it has become popular to learn transformations that can be later on used to synthesize new data from the available labeled data. In [10] the authors propose to learn the transformations that align a pair of images - spatial transformations, for the MNIST dataset, and use them for data augmentation [Figure 1]. In addition to spatial transformations, many works have been known to use the appearance transformations that are accountable for the intensities variations [7]. 

Figure 1: Example of learned spatial transformations. In the top row there is original data and in the second row the images with the random transformation from the model applied [10].

In this paper, the authors drive their ideas from the combination of spatial and appearance transformation models to propose a novel data augmentation technique, which works even in a more challenging scenario of a few-shot semantic segmentation, where just a few annotated instances may be used to perform semantic segmentation. The high-level idea of the proposed method is to learn the transformation from the data and use them to generate a naturally looking dataset that can capture the complexity of biomedical image real variation

Methodology

In order to synthesize realistically-looking labeled scans the proposed approach requires only one labeled image and a number of unlabeled instances. Two separate transformation models are used to learn and compute a set of spatial and appearance transformations between each pair of the labeled and unlabeled samples. Later on, these transformations can be randomly picked from the set and applied to the labeled instance in order to produce a new synthetic example [Figure 2].

Figure 2: The overview of the proposed method [11].

Spatial Transformation Model

Spatial transformation model is used to align two given images and find anatomical correspondences from scan to scan. In this paper, the results from an unsupervised learning model VoxelMorph [5], that performs spatial transformation for atlas-based segmentation, are used. Having a labeled reference volume $(x,l_x)$  - an atlas, and an unlabeled set of biomedical image volumes $\{y^{(i)}\}$, the transformation to align $x$ to $\{y^{(i)}\}$ is modeled as a voxel-wise displacement field $u$ [Figure 4]. The deformation function is defined as $\phi = id + u$, where $id$ is an identity function. To account for the natural variability in the data, the deformation that wraps atlas $x$ to each volume $y^{(i)}$ is modeled with $\phi^{(i)}=g_{\theta_s}(x,y^{(i)})$, where $g_{\theta_s}(x,y^{(i)})$ is a parametric function that is learned by a convolutional neural network based on the U-Net [13] architecture [Figure 3].

Figure 3: Overview of the spatial transformation model [5].

Figure 4: An example of displacement field from aligning the moving image to the fixed one. The displacement field shows with vectors where each of the voxels should be moved [12].

The loss function for this network consists of two parts:

\mathcal{L}(x,y^{(i)},\phi^{(i)})=\mathcal{L}_{sim}(y^{(i)}, x\circ\phi^{(i)})+\lambda \mathcal{L}_{smooth}(\phi^{(i)})

  1. Similarity loss \mathcal{L}_{sim}(y^{(i)}, x\circ\phi^{(i)}) that is used to measure image similarity. It is based on cross-correlation, which makes it robust to intensity variations [8].
    \mathcal{L}_{sim}(y^{(i)}, x\circ\phi^{(i)}) = -CC(y^{(i)}, x\circ\phi^{(i)})
    CC(y, x\circ\phi)=\sum_{p \in \Omega}{\frac{\Big(\sum_{p_i}\big(y(p_i)-\hat{y}(p)\big)\big([x\circ\phi](p_i)-[\hat{x}\circ\phi](p)\big)\Big)^2}{\Big(\sum_{p_i}\big(y(p_i)-\hat{y}(p)\big)^2\Big)\Big(\sum_{p_i}\big([x\circ\phi](p_i)-[\hat{x}\circ\phi](p)\big)^2\Big)}}, where \hat{f}(p) = \frac{1}{n^3}\sum_{p_i}f(p_i) - local mean intensity
  2. Smoothness loss \mathcal{L}_{smooth}(\phi^{(i)})  that makes the deformation field smooth and is modeled with a spatial gradient of the deformation field
    \mathcal{L}_{smooth}(\phi)=\sum_{p \in \Omega}||\nabla u(p)||^2


Appearance Transformation Model

The appearance transformation model is responsible for modeling variations in intensities. It is modeled as per-voxel addition in intensities.

$\tau_a^{(i)}(x)=x+\psi^{(i)}$

It should be noted that we are trying to find such transformation that can be applied directly to our atlas in order to synthesize new images. Because of that, we have to register our unlabeled instance to the atlas space first, using an inverse spatial transformation \phi^{-1^{(i)}}  . The inverse spatial transformation is implemented as a separate convolutional neural network, identical to the one described in the previous section. 


The naive solution to estimate $\psi$  would be to simply subtract the intensities between the atlas and unlabeled instances in the atlas space. However, we would capture all of the errors associated with the incorrect image registrations. In order to avoid that, similarly to spatial transformation, we model $\psi$  as a parametric function $h_{\theta_a}(x, y^{(i)}\circ\phi^{-1^{(i)}})$, that can be learned by a convolutional neural network based on the U-Net [13] architecture [Figure 5].

Figure 5: Overview of the appearance transformation model [6].

Similarly to spatial transformation model the loss function consists of a similarity and a smoothness term: 

$\mathcal{L}(x,y^{(i)},\phi^{(i)},\phi^{-1^{(i)}},\psi^{(i)},c_x)=\mathcal{L}_{sim}((x+\psi^{(i)})\circ\phi^{(i)},y^{(i)})+\lambda_a \mathcal{L}_{smooth}(c_x,\psi^{(i)})$

  1. Similarity loss \mathcal{L}_{sim}((x+\psi^{(i)})\circ\phi^{(i)},y^{(i)}) is computed as a mean squared error $\mathcal{L}_{sim}(\hat{y},y)=||\hat{y}-y||^2$
  2. Smoothness term $\mathcal{L}_{smooth}(c_x, \psi)=(1-c_x)\nabla\psi$  uses segmentation labels $l_x$  to compute binary image of anatomical boundaries c_x, which makes the smoothness term prevent any big intensity changes within the same anatomical structure.

Synthesizing new examples

For generating a new labeled instance, the system samples two unlabeled images for which it extracts previously saved spatial and appearance transformation. Then the sampled intensity transformation is overlayed on the atlas. On top of that, the spatial transformation is applied to make the voxels structurally match an unlabeled scan. In the end, the labels are mapped to the new structures, by application of the same spatial transformation to them.

Figure 6: Overview of the process for synthesizing new training samples [6].

Results and Discussion

Experimental Setup

For the experiments,  the task of brain MRI scans segmentation with 30 anatomical labels was chosen. For that, the same dataset of T1-weighted MRI brain scans as in VoxelMorph [5] was taken and the ground-truth segmentation labels were generated using FreeSurfer [14]. The authors selected a state-of-the-art segmentation method [15], for which the training dataset is augmented using the proposed technique. They used 101,  50, 100 brain volumes for the training, validation and testing set accordingly. For training, all of the methods are allowed to use labels only for one sample and leverage the other 100 instances without labels. 

Segmentation Baselines

  1. Single-atlas segmentation (SAS) - uses VoxelMorph atlas-based segmentation [5] to warp the atlas labels with the same spatial transformation as in this work 
  2. Data augmentation using single-atlas segmentation (SAS-aug) - use SAS segmentation as ground-truth labels for the 100 unlabeled training samples and include them as training examples for supervised segmentation model
  3. Hand-tuned random data augmentation (rand-aug) - hand-tuned data augmentation that synthesizes a new random brain in each training iteration using random smooth deformation fields [13] and global intensity multiplicative factor [16]
  4. Supervised Segmentation - segmentation network that uses 101 ground truth labels in its training dataset. Serves as an upper-bound to all other methods

Method Variants

  1. Independent sampling (ours-indep) - spatial and appearance transformations are sampled independently, resulting in 10000 new training instances.
  2. Coupled sampling (ours-coupled) - spatial and appearance transformations are sampled from the same unlabeled image resulting in only 100 new training instances
  3. Ours-indep + rand-aug - both ours-indep and rand-aug methods are used in alternation to synthesize new examples

Segmentation results

Mean segmentation performance

For testing the segmentation accuracy the Dice score, which measures the overlapping area between two segmentations, was selected. All of the versions of the proposed method outperformed other methods both in mean Dice score and in mean pairwise Dice improvement [Figure 7], which corresponds to the mean value of the improvement over SAS baseline on each testing sample.

Figure 7: Segmentation performance in terms of dice score [6]

The pairwise improvement for the proposed method is also more consistent than random augmentations [Figure 8] and is higher than other methods for almost every sample in the testing dataset [Figure 9].

Figure 8:  Pairwise Dice improvement over the SAS baseline [6].

Figure 9: Pairwise Dice improvement over the SAS baseline for each test sample [6].

Segmentation accuracy for different structures

The authors also obtained interesting results when comparing the segmentation accuracy across different segmented structures [Figure 10].

Figure 10: Segmentation accuracy across different brain structures with the following abbreviations of the labels: white matter (WM), cortex (CX), ventricle (vent), and cerebrospinal fluid (CSF). The volume taken by each structure is shown in parenthesis [6].

While for the large brain structures all of the methods were comparably accurate, the proposed approach outperformed all other approaches on smaller structures, such as the hippocampus [Figure 11].

Figure 11: Examples of two test instances (rows) hippocampus segmentations [6].

Conclusions

In this paper, a novel approach for data augmentation that can be applied to a one-shot segmentation scenario was introduced. The method requires only a single labeled instance and a bunch of unlabeled samples, from which it is able to learn and emulate real variation of medical images. The proposed approach has proven to outperform all other segmentation methods and produce more accurate segmentation labels overall. As this approach is very general and can be applied to different medical problems the next steps for this method are trying it out on various anatomies and imaging modalities. The other improvement for this work is to try and make the transformations space continuous by using the interpolation or composition of transformations

Own Review

Overall, I think the paper is well structured and has very nice accompanying tables and pictures, that really enhance the understanding of this work. Also, the source code for all models has been published and other people can use it to test the method on different problems. I liked that the approach presented here is quite broad and one can see many other problems that can benefit from this technique. 

However, the experimental setup and the results presented by the authors, in my opinion, are insufficient to claim the superiority of this approach. A big disadvantage of this work is that no manual labels were used for the segmentation and all of the labels were automatically generated by third-party software. Also, all of the baseline methods are either created or modified by the same research group. The authors do not mention in the paper any other state-of-the-art augmentation techniques that they can compare their work to. 

Another point is that I think not all of the experimental setup has been explained in detail. For example, the authors only mention the data for training the supervised segmentation model. However, they omit whether the transformation models have been trained on the same small subsample of the data, or the full dataset was used for it. According to my research on VoxelMorph [5], from which the spatial transformation model was taken, it was trained on a far bigger dataset than 100 images. Also, there is no time discussion required for training or inference provided in the paper. 

To sum up, I think this work has a good idea for data augmentation, but it would be important to see how well it works in other experiments first. In particular, it seems interesting to check whether naturally looking images with unhealthy anatomies (e.g. tumors) can be synthesized this way.

References

[1] P. Moeskops, M. A. Viergever, A. M. Mendrik, L. S. de Vries, M. J. Benders, and I. Isgum. Automatic segmentation of mr brain images with a convolutional neural network. IEEE transactions on medical imaging, 35(5):1252– 1261, 2016

[2] P.-A. Ganaye, M. Sdika, and H. Benoit-Cattin. Semisupervised learning for segmentation under semantic constraint. In International Conference on Medical Image Computing and Computer-Assisted Intervention, pages 595–602. Springer, 2018

[3] A. V. Dalca, J. Guttag, and M. R. Sabuncu. Anatomical priors in convolutional networks for unsupervised biomedical segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 9290– 9299, 2018

[4] K. K. Leung, M. J. Clarkson, J. W. Bartlett, S. Clegg, C. R. Jack Jr, M. W. Weiner, N. C. Fox, S. Ourselin, A. D. N. Initiative, et al. Robust atrophy rate measurement in alzheimer’s disease using multi-site serial mri: tissue-specific intensity normalization and parameter selection. Neuroimage, 50(2):516–523, 2010

[5] G. Balakrishnan, A. Zhao, M. R. Sabuncu, J. Guttag, and A. V. Dalca. Voxelmorph: a learning framework for deformable medical image registration. IEEE transactions on medical imaging, 2019.

[6] Zhao, A., Balakrishnan, G., Durand, F., Guttag, J.V., Dalca, A.V., 2019. Data augmentation using learned transformations for one-shot medical image segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8543–8553

[7] G. Vincent, G. Guillard, and M. Bowes. Fully automatic segmentation of the prostate using active appearance models. MICCAI Grand Challenge: Prostate MR Image Segmentation, 2012, 2012B. B. Avants, C. L.

[8] B. B. Avants, C. L. Epstein, M. Grossman, and J. C. Gee, “Symmetric diffeomorphic image registration with cross-correlation: evaluating automated labeling of elderly and neurodegenerative brain,” Medical image analysis, vol. 12, no. 1, pp. 26–41, 2008

[9] Z. Akkus, A. Galimzianova, A. Hoogi, D. L. Rubin, and B. J. Erickson. Deep learning for brain mri segmentation: state of the art and future directions. Journal of digital imaging, 30(4):449–459, 2017.

[10] S. Hauberg, O. Freifeld, A. B. L. Larsen, J. Fisher, and L. Hansen. Dreaming more data: Class-dependent distributions over diffeomorphisms for learned data augmentation. In Artificial Intelligence and Statistics, pages 342–350, 2016

[11] https://www.rsipvision.com/CVPR2019-Thursday/4/

[12] https://www.kaggle.com/adalca/learn2reg

[13] O. Ronneberger, P. Fischer, and T. Brox. U-net: Convolutional networks for biomedical image segmentation. In International Conference on Medical image computing and computer-assisted intervention, pages 234–241. Springer, 2015

[14] B. Fischl. Freesurfer. Neuroimage, 62(2):774–781, 2012

[15] A. G. Roy, S. Conjeti, D. Sheet, A. Katouzian, N. Navab, and C. Wachinger. Error corrective boosting for learning fully convolutional networks with limited data. In International Conference on Medical Image Computing and Computer-Assisted Intervention, pages 231–239. Springer, 2017

[16] K. Kamnitsas, C. Ledig, V. F. Newcombe, J. P. Simpson, A. D. Kane, D. K. Menon, D. Rueckert, and B. Glocker. Efficient multi-scale 3d cnn with fully connected crf for accurate brain lesion segmentation. Medical image analysis, 36:61– 78, 2017

  • Keine Stichwörter