2017 NI Carass Longitudinal Multiple Sclerosis Lesion Segmentation Resource and Challenge
2017 NI Carass Longitudinal Multiple Sclerosis Lesion Segmentation Resource and Challenge
NeuroImage
journal homepage: www.elsevier.com/locate/neuroimage
A R T I C L E I N F O A BS T RAC T
Keywords: In conjunction with the ISBI 2015 conference, we organized a longitudinal lesion segmentation challenge
Magnetic resonance imaging providing training and test data to registered participants. The training data consisted of five subjects with a
Multiple sclerosis mean of 4.4 time-points, and test data of fourteen subjects with a mean of 4.4 time-points. All 82 data sets had
the white matter lesions associated with multiple sclerosis delineated by two human expert raters. Eleven teams
submitted results using state-of-the-art lesion segmentation algorithms to the challenge, with ten teams
presenting their results at the conference. We present a quantitative evaluation comparing the consistency of the
two raters as well as exploring the performance of the eleven submitted results in addition to three other lesion
⁎
Correspondence to: Department of Electrical and Computer Engineering, The Johns Hopkins University, 105 Barton Hall, 3400 N. Charles St., Baltimore, MD 21218, USA.
E-mail address: [email protected] (A. Carass).
1
These authors co-organized the challenge, all others contributed results.
http://dx.doi.org/10.1016/j.neuroimage.2016.12.064
Received 19 August 2016; Accepted 19 December 2016
Available online 11 January 2017
1053-8119/ © 2017 Elsevier Inc. All rights reserved.
A. Carass et al. NeuroImage 148 (2017) 77–102
segmentation algorithms. The challenge presented three unique opportunities: (1) the sharing of a rich data set;
(2) collaboration and comparison of the various avenues of research being pursued in the community; and (3) a
review and refinement of the evaluation metrics currently in use. We report on the performance of the challenge
participants, as well as the construction and evaluation of a consensus delineation. The image data and manual
delineations will continue to be available for download, through an evaluation website2 as a resource for future
researchers in the area. This data resource provides a platform to compare existing methods in a fair and
consistent manner to each other and multiple manual raters.
78
A. Carass et al. NeuroImage 148 (2017) 77–102
topology consistency constraint (Shiee et al., 2010); and adaptive Public databases have played a transformative role in medical
dictionary learning (Deshpande et al., 2015; Roy et al., 2014a, imaging, an early example of this is the now ubiquitous BrainWeb
2015b); along with many other techniques. (Collins et al., 1998) computational phantom (see also Cocosco et al.,
The majority of these methods operate in an unsupervised manner 1997 and Kwan et al., 1999). With over one hundred citations per year
using statistical notions about distributions to identify lesions. There for the last decade, it is almost inconceivable to write an MR-based
has also been significant work done to develop supervised methods, brain segmentation paper without including an evaluation on the
which use training data to identify lesions within new subjects. One BrainWeb phantom. These public databases have served to standardize
such approach included an anatomical template-based registration to comparisons and evaluation criteria. In recent years there has been a
help modulate a k-NN classification scheme (Warfield et al., 2000), shift in the community to launch these data sets as a challenge
which used features from the images as well as distances to the associated with a workshop or conference (Styner et al., 2008;
template following the registration. Sweeney et al. (2013b) presented Schaap et al., 2009; Heimann et al., 2009; Menze et al., 2015;
a logistic regression model that assigned voxel-level probabilities of Mendrik et al., 2015; Maier et al., 2017). In particular, the 2008
lesion presence. Roy et al. (2014b) demonstrated a patch-based lesion MICCAI MS Lesion challenge (Styner et al., 2008) was a significant step
segmentation that used examples from an atlas to match patches in the forward in the sharing of clinically relevant data. These benchmark
input images using a sparse dictionary approach. Variants of this data sets allow for a direct comparison between competing methods
supervised machine learning solution include: generic machine learn- without any unique data issues, and just as importantly, these bench-
ing (Xie and Tao, 2011); dictionary learning and sparse-coding (Roy marks remove the barrier of data that limits the number of researchers
et al., 2014a, 2015b; Weiss et al., 2013); and random forest (RF) work working in a particular area. An important feature of benchmarks is the
by Mitra et al. (2014), variations of the RF approach include Geremia retention of the test data set labels from the public domain avoiding the
et al. (2010, 2011) using multi-channel MR intensities, long-range “unintentional overtraining of the method being tested” and preserving
spatial context, and asymmetry features to identify lesions; Jog et al. “the method's segmentation performance in practice” (Menze et al.,
(2015) producing overlapping lesion masks from the RF that were 2015).
averaged to create a probabilistic segmentation, and Maier et al. (2015) In this paper, we present details of the Longitudinal White Matter
used extra tree forests (Geurts et al., 2006) which are robust to noise Lesion Segmentation of Multiple Sclerosis Challenge (hereafter the
and uncertain training data. Challenge) that was conducted during the 2015 International
There has been less work on automated methods for serial lesion Symposium on Biomedical Imaging (ISBI). The Challenge data will
segmentation (segmentation of lesions for the same subject over serve as an ongoing resource with future submissions for evaluation
different time-points). The earliest reported approach (Rey et al., possible through the Challenge Website (see Footnote 2). In Section 2,
1999, 2002) performed an optical flow registration between successive we outline the data provided to the Challenge participants, the set-up of
rigidly registered time-points, then used the Jacobian of the deforma- the Challenge, and the evaluation metrics used in comparing the
tion field to identify the lesions. Published at about the same time, submitted results from each team. Section 2 also includes a description
Kikinis et al. (1999) used 4D connected component analysis for of our Consensus Delineation, which avoids the biases of depending on
longitudinal lesion segmentation. Prima et al. (2002) introduced voxel a single rater. Section 3 provides an overview of the methods involved
wise statistical testing to identify regions with significantly increased in the Challenge with complete descriptions of each algorithm included
intensity over time, treating the appearance of WMLs as a change-point in Appendix B. Section 4 includes the comparison between the manual
problem. Welti et al. (2001) created a feature vector of radial intensity- delineations, algorithms, and the Consensus Delineation. We conclude
based descriptors of lesions from four contrast images at multiple time- the main body of the manuscript with a discussion of the impact of this
points. The course of these descriptors is then analyzed with a principal Challenge and future directions in WML segmentation in Section 5.
component analysis (PCA) to build a model of spatio-temporal lesion Appendix A includes a complete description of the protocol used for the
evolution. Projection of candidate lesions into the PCA space was used manual delineation. Appendix C includes the results from the
to identify lesions, with the maximal temporal gradient of a FLAIR Challenge at ISBI.
image being used to identify the onset of the lesion. Bosc et al. (2003)
used a pipeline comprised of iterative affine registration, deformable 2. Materials and metrics
registration, image resampling, and intensity normalization, followed
by a temporal change point detection scheme. Their change point Teams registered for the Challenge, and received access to a
detection used a generalized likelihood ratio test (GLRT) (Hsu et al., Training Set of images in February of 2015. Followed one month later
1984) that computes the probabilities of the two hypotheses (no change by the first evaluation data set (Test Set A), with the Teams having one
vs. significant change). We note that the initial steps of Bosc et al. month to return their results for evaluation. One week before the
(2003), up to change detection, are now considered standard prepro- Challenge event at ISBI 2015, Teams were provided with a second
cessing for time-series data and is similar to the preprocessing that was evaluation data set (Test Set B). Teams were told that the time between
performed on the data in our challenge. As previously mentioned Aït- downloading Test Set B and the return of their results would be timed
Ali et al. (2005) extended the outlier detection approach (Van Leemput for comparison. Participants were informed of the criteria for the
et al., 2001) to the entire time series using TLE followed by refinement Challenge prizes, which were furnished by the National MS Society.
steps. Roy et al. (2015a) extended their 3D example patch-based lesion Details of the data, preprocessing, and the Challenge metrics are
segmentation algorithm to 4D by considering a time series of patches provided below. The results of the Challenge are provided in
from available training data. Other work evaluated WML changes over Appendix C.
time (Battaglini et al., 2014; Elliott et al., 2010; Ganiler et al., 2014;
Roura et al., 2015; Sweeney et al., 2013a) with the focus being on the 2.1. Challenge data
appearance/disappearance of lesions by subtraction of the intensity
images of consecutive time-points. As there clearly has been a relative The Challenge participants were given three tranches of data: (1)
dearth of work on the automated segmentation of time-series WMLs, Training Set; (2) Test Set A; and (3) Test Set B. The Training Set
and as there is no approach that has gained widespread acceptance, a consisted of five subjects, four of which had four time-points, while the
main purpose of this paper is to provide a public database to reignite fifth subject had five time-points. Test Set A included ten subjects, eight
work in this area. of which had four time-points, one had five time-points, and one had
79
A. Carass et al. NeuroImage 148 (2017) 77–102
Table 1 size; a double spin echo (DSE) which produces the PD-w and T2-w
Demographic details for the training data and both test data sets. The top line is the images with TR=4177 ms, TE1=12.31 ms, TE2=80 ms, and
information of the entire data set, while subsequent lines within a section are specific to
0.82×0.82×2.2 mm3 voxel size; and a T2-w fluid attenuated inversion
the patient diagnoses. The codes are RR for relapsing remitting MS, PP for primary
progressive MS, and SP for secondary progressive MS. N (M/F) denotes the number of recovery (FLAIR) with TI=835 ms, TE=68 ms, and
patients and the male/female ratio, respectively. Time-points is the mean (and standard 0.82×0.82×2.2 mm3 voxel size. The imaging protocols were approved
deviation) of the number of time-points provided to participants. Age is the mean age by the local institutional review board. Each subject underwent the
(and standard deviation), in years, at baseline. Follow-up is the mean (and standard following preprocessing: the baseline (first time-point) MPRAGE was
deviation), in years, of the time between follow-up scans.
inhomogeneity-corrected using N4 (Tustison et al., 2010), skull-
Data Set N (M/F) Time-Points Age Follow-Up stripped (Carass et al., 2007, 2010), dura stripped (Shiee et al.,
2014), followed by a second N4 inhomogeneity correction, and rigid
Mean (SD) Mean (SD) Mean (SD) registration to a 1 mm isotropic MNI template. We have found that
running N4 a second time after skull and dura stripping is more
Training 5 (1/4) 4.4 ( ± 0.55) 43.5 ( ± 10.3) 1.0 ( ± 0.13)
RR 4 (1/3) 4.5 ( ± 0.50) 40.0 ( ± 7.55) 1.0 ( ± 0.14) effective (relative to a single correction) at reducing any inhomogeneity
PP 1 (0/1) 4.0 57.9 1.0 ( ± 0.04) within the images (see Fig. 1 for an example image set after
preprocessing). Once the baseline MPRAGE is in MNI space, it is used
Test A 10 (2/8) 4.3 ( ± 0.68) 37.8 ( ± 9.18) 1.1 ( ± 0.28)
as a target for the remaining images. The remaining images include the
RR 9 (2/7) 4.3 ( ± 0.71) 37.4 ( ± 9.63) 1.1 ( ± 0.29)
SP 1 (0/1) 4.0 41.7 1.0 ( ± 0.05) baseline T2-w, PD-w, and FLAIR, as well as the scans from each of the
follow-up time-points. These images are N4 corrected and then rigidly
Test B 4 (1/3) 4.5 ( ± 0.58) 43.3 ( ± 7.64) 1.0 ( ± 0.05) registered to the 1 mm isotropic baseline MPRAGE in MNI space. Our
RR 3 (1/2) 4.7 ( ± 0.58) 44.8 ( ± 8.65) 1.0 ( ± 0.05) registration steps are inverse consistent and thus any registration
PP 1 (0/1) 4.0 39.0 1.0 ( ± 0.04)
based biases are avoided (Reuter and Fischl, 2011). The skull and dura
stripped mask from the baseline MPRAGE is applied to all the
six time-points. Test Set B had four subjects–three with four time- subsequent images, which are then N4 corrected again. All the images
points and two with five time-points. Two consecutive time-points are in the Training Set, Test Set A, and Test Set B, had their lesions
separated by approximately one year for all subjects. Table 1 includes a manually delineated by two raters in the MNI space. Rater #1 has four
demographic breakdown for the training and test data sets. Challenge years of experience delineating lesions, while Rater #2 has 10 years
participants did not know the MS status of the subjects of each data set. experience with manual lesion segmentation and 17 years experience
Each scan was imaged and preprocessed in the same manner, with in structural MRI analysis. We note that the raters were blinded to the
data acquired on a 3.0 Tesla MRI scanner (Philips Medical Systems, temporal ordering of the data. The protocol for the manual delineation
Best, The Netherlands) using the following sequences: a T1-weighted followed by both raters is in Appendix A. The preprocessing steps were
(T1-w) magnetization prepared rapid gradient echo (MPRAGE) with performed using JIST (Version 3.2) (Lucas et al., 2010).
TR=10.3 ms, TE=6 ms, flip angle =8°, and 0.82×0.82×1.17 mm3 voxel For each time-point of every subject's scans in the Training Set, Test
(a) (b ) (c )
(d ) ( e) (f )
Fig. 1. Shown are the preprocessed (a) MPRAGE, (b) FLAIR, (c) T2-w, and (d) PD-w images for a single time-point from one of the provided Training Set subjects. The corresponding
manual delineations by our two raters are shown in (e) for Rater #1 and (f) for Rater #2.
80
A. Carass et al. NeuroImage 148 (2017) 77–102
Set A, and Test Set B, the participants were provided the following Table 2
data: the original scan images consisting of T1-w MPRAGE, T2-w, PD- Inter-rater comparison averaged across the 82 images from the training and test data set.
The first table displays the symmetric metrics: Dice, average symmetric surface distance
w, and FLAIR, as well as the preprocessed images (in MNI space) for
(ASSD), and longitudinal correlation. The second table shows the asymmetric metrics:
each of the scan modalities. The Training Set also included manual positive predictive value (PPV), true positive rate (TPR), lesion false positive rate, lesion
delineations by two experts identifying and segmenting WMLs on MR true positive rate, and absolute volume difference (AVD). R1 refers to Rater #1, R2 to
images: details about the delineation protocol and lesion inclusion Rater #2, and “R1 vs. R2” denotes that R1 was regarded as the truth within the
criteria are in Appendix A. comparison.
As teams registered for the Challenge, they were provided with the Symmetric metrics
Training Data. One month prior to the scheduled Challenge, Test Set A
was made available to participants. The results for Test Set A could be Dice 0.6340
returned to the organizers at any time prior to the Challenge event, ASSD 3.5290
Longitudinal correlation −0.0053
though a preferred return date was given. The third data set, Test Set B,
was provided to participants one week before the Challenge event with Asymmetric metrics R1 vs. R2 R2 vs. R1
the caveat that teams would be timed. The times used were based on PPV 0.7828 0.5688
the initial download time for each team and the time at which they TPR 0.5029 0.8224
Lesion FPR 0.1380 0.5630
returned their results to the Challenge organizers. In Appendix C we
Lesion TPR 0.4370 0.8620
include a comparison of the ten Challenge participants on both Test Set AVD 0.3726 0.6117
A and B, and in C.1 we report the time it took participants to process
and return Test Set B.
where 3 cR is the 18-connected components of 4 cR .
2.2. Challenge metrics Absolute volume difference (AVD) is the absolute difference in
volumes divided by the true volume,
To compare the results from the participants with our two manual
raters and Consensus Delineation, we used the following metrics: Dice Max( 4 R , 4 A ) − Min( 4 R , 4 A )
AVD(4 R , 4 A) = .
overlap (Dice, 1945), positive predictive value, true positive rate, lesion 4R
true positive rate, lesion false positive rate, absolute volume difference,
average symmetric surface distance, volume correlation, and long- Average symmetric surface distance (ASSD) is the average of the
itudinal volume correlation. The Dice overlap is a commonly used distance (in millimeters) from the lesions in 4 R to the nearest lesion
volume metric for comparing the quality of two binary label masks. It is identified in 4 A plus the distance from the lesions in 4 A to the nearest
defined as the ratio of twice the number of overlapping voxels to the lesion identified in 4 R .
total number of voxels in each mask. If 4 R is the mask of one of the
∑r ∈ 3 d (r , 3A) + ∑a ∈ 3 d (a, 3R)
human raters and 4 A is the mask generated by a particular algorithm, ASSD(4 R , 4 A) = R A
,
then the Dice overlap is computed as 2
81
A. Carass et al. NeuroImage 148 (2017) 77–102
Fig. 2. Delineations are shown for a sample slice from the preprocessed MPRAGE, FLAIR, and T2-w images for a time-point of a test data set, followed by our Consensus Delineation
and the results for the top eight delineations as ranked by their Dice Score with the Consensus. For ease of reference, a grid has been added underneath the delineations. The bottom
eight delineations, as ranked by their Dice Score with the Consensus, can be see in Fig. 3.
2.4. Consensus delineation combination of the input segmentations, the weights for which are
determined by the estimated performance level of the individual
To avoid the biases of depending on either rater, we choose to segmentations. The resultant Consensus Delineation, from the
construct a Consensus Delineation for each of the 61 images included STAPLE combination of the 14 algorithms and 2 manual raters, is
in Test Set A and B. To achieve such a delineation, we employ the regarded as the “ground truth” for the comparisons within Section 4.
simultaneous truth and performance level estimation (STAPLE) algo- The Consensus Delineation provides the opportunity to simultaneously
rithm (Warfield et al., 2004). STAPLE is an expectation-maximization compare the human raters and the Challenge participants across all of
algorithm for the statistical fusion of binary segmentations. The our metrics; this—to our knowledge—is something that has not been
algorithm considers several segmentations and computes a probabil- reported in any previous Challenge (Styner et al., 2008; Schaap et al.,
istic estimate of the true segmentation—as well as other quantities. 2009; Heimann et al., 2009; Menze et al., 2015; Mendrik et al., 2015;
Given that we have only two manual delineations for each patient Maier et al., 2017).
image, we have taken the Challenge Delineations provide by each team
(see Section 3 and Appendix B for details) and included them with our 3. Methods overview
two manual delineations in construction of the Consensus Delineation.
In brief, STAPLE estimates the true segmentation from an optimal We present a brief overview of each of the methods used in this
82
A. Carass et al. NeuroImage 148 (2017) 77–102
Fig. 3. Delineations are shown for a sample slice from the preprocessed MPRAGE, FLAIR, and T2-w images for a time-point of a test data set, followed by our Consensus Delineation
and the results for the bottom eight delineations as ranked by their Dice Score with the Consensus. For ease of reference, a grid has been added underneath the delineations. The top
eight delineations, as ranked by their Dice Score with the Consensus, can be see in Fig. 2.
paper, complete details of each approach are available in Appendix B. Rater #1 and #2 with colored squares with the letter M to denote
Figs. 2 and 3 show results of each algorithm on a typical slice from one manual delineations.
time-point of one of our data sets, as well as the corresponding
MPRAGE, FLAIR, and T2-w images. Ten teams originally submitted
results for the Challenge data sets and were able to participate in the 3.1. Challenge participants
Challenge event (see Section 2.1 for a complete description of the data).
In addition to these methods, we received results for two methods from .
teams that did not participate in the Challenge event. To provide some Multi-Contrast PatchMatch Algorithm for Multiple Sclerosis
context with the 2008 MICCAI MS Lesion challenge (Styner et al., Lesion Detection.
2008), we also include the methods that finished first and third in that (F. Prados, M. J. Cardoso, N. Cawley, O. Ciccarelli, C. A. M.
challenge. Where we present descriptions or results of the methods, we Wheeler-Kingshott, and S. Ourselin).
use a colored square to help identify methods and within that square Team CMIC used the PatchMatch (Barnes et al., 2010) algorithm
we denote methods that are unsupervised with the letter U and those for MS lesion detection. The main contribution of this work is the
that require some training data (supervised methods) with the letter S. generalization of the optimized PatchMatch algorithm to the context of
When considering the Consensus Delineation in Section 4, we identify MS lesion detection and its extension to multimodal data.
.
83
A. Carass et al. NeuroImage 148 (2017) 77–102
Automatic Graph Cut Segmentation of Multiple Sclerosis Lesions. (M. Ghafoorian and B. Platel).
(L. Catanese, O. Commowick, and C. Barillot). Team DIAG utilizes a deep CNN with five layers in a sliding window
Team VISAGES GCEM used a robust Expectation-Maximization fashion to create a voxel-based classifier.
(EM) algorithm to initialize a graph, followed by a min-cut of the graph .
to detect lesions, and an estimate of the WM to help remove false Model Selection Propagation for Application on Longitudinal MS
positives. GCEM stands for G raph-c ut with E xpectation-M aximisa- Lesion Segmentation.
tion. (C. H. Sudre, M. J. Cardoso, and S. Ourselin).
. Based on the assumption that the structural anatomy of the brain
Sparse Representations and Dictionary Learning Based should be temporally consistent for a given patient, Team TIG proposes
Longitudinal Segmentation of Multiple Sclerosis Lesions. a lesion segmentation method that first derives a GMM separating
(H. Deshpande, P. Maurel, and C. Barillot). healthy tissues from pathological and unexpected ones on a multi-time-
Team VISAGES DL used sparse representation and a dictionary point intra-subject group-wise image. This average patient-specific
learning paradigm to automatically segment MS lesions within the GMM is then used as an initialization for a final time-point specific
longitudinal MR data. Dictionaries are learned for the lesion and GMM from which final lesion segmentations are obtained. Team TIG
healthy brain tissue classes, and a reconstruction error-based classifi- submitted new results after the completion of the Challenge to address
cation approach for prediction. a bug in their code, the second submitted results are denoted TIG BF.
. Both sets of results are included in Appendix C; however, the
Model of Population and Subject (MOPS) Segmentation. Consensus Delineation was only compared to the bug fixed results
(X. Tomas-Fernandez and S. K. Warfield). (TIG BF).
Inspired by the ability of experts to detect lesions based on their
local signal intensity characteristics, Team CRL proposes an algorithm
that achieves lesion and brain tissue segmentation through simulta- 3.2. Other included methods
neous estimation of a spatially global within-the-subject intensity
distribution and a spatially local intensity distribution derived from a These methods did not participate in the Challenge, however they
healthy reference population. are included to add to the richness and variety of the methods
. presented. MORF and Lesion-TOADS represent methods that finished
Longitudinal Multiple Sclerosis Lesion Segmentation using 3D first and third in the 2008 MICCAI MS Lesion challenge (Styner et al.,
Convolutional Neural Networks. 2008), respectively, and as such offer the opportunity to provide a
(S. Vaidya, A. Chunduru, R. Muthuganapathy, and G. reference between the two challenges. In particular, the two algorithms
Krishnamurthi). offer different perspectives on the problem (supervised versus unsu-
Team IIT Madras modeled a voxel-wise classifier using multi- pervised, respectively) while also testing the ongoing viability of these
channel 3D patches of MRI volumes as input. For each ground truth, two methods within the field. Our third included method (MV-CNN)—
a convolutional neural network (CNN) is trained and the final based on deep-learning—is a state-of-the-art approach; the authors of
segmentation is obtained by combining the probability outputs of these MV-CNN submitted their results to the Challenge Website while this
CNNs. Efficient training is achieved by using sub-sampling methods manuscript was in preparation. As a deep-learning method, MV-CNN
and sparse convolutions. represents a key direction in which the medical imaging community is
. moving. While the fourth included method, BAUMIP, submitted results
Hierarchical MRF and Random Forest Segmentation of MS for both Challenge data sets but was unable to participate at the
Lesions and Healthy Tissues in Brain MRI. Challenge event.
(A. Jesson and T. Arbel). .
Team PVG One built a hierarchical framework for the segmentation Automatic White Matter Hyperintensity Segmentation using
of a variety of healthy tissues and lesions. At the voxel level, lesion and FLAIR MRI.
tissue labels are estimated through a MRF segmentation framework (L. O. Iheme and D. Unay).
that leverages spatial prior probabilities for nine healthy tissues BAUMIP is a method based on intensity thresholding and 3D voxel
through multi-atlas label fusion (MALF). A random forest (RF) connectivity analysis. A simple model is trained that is optimized by
classifier then provides region level lesion refinement. searching for the maximum obtainable Dice overlap.
. .
MS-Lesion Segmentation in MRI with Random Forests. Multi-View Convolutional Neural Networks.
(O. Maier and H. Handels). (A. Birenbaum and H. Greenspan).
Team IMI trained a RF with supervised learning to infer the MV-CNN is a method based on a Longitudinal Multi-View CNN
classification function underlying the training data. The classification (Roth et al., 2014). The classifier is modeled as a CNN, whose input for
of brain lesions in MRI is a complex task with high levels of noise, every evaluated voxel are patches from axial, coronal, and sagittal views
hence a total of 200 trees are trained without any growth-restriction. of the T1-w, T2-w, PD-w, and FLAIR images of the current and previous
Contrary to reported observations, no overfitting occurred. time-points. That is multiple contrasts, multiple views, and multiple
. time-points. MV-CNN consists of three phases: Preprocessing the
Automatic Longitudinal Multiple Sclerosis Lesion Segmentation. Challenge data, Candidate Extraction, and CNN Prediction. The
(S. Jain, D. M. Sima, and D. Smeets). Challenge data is preprocessed by intensity clamping the top and
MSmetrix (Jain et al., 2015) is presented, which performs lesion bottom 1% and the intensity values are scaled to the range [0, 1].
segmentation while segmenting brain tissue into CSF, GM, and WM, .
with lesions identified based on a spatial prior and hyperintense Multi-Output Random Forests for Lesion Segmentation in Multiple
appearance in FLAIR. Sclerosis.
. (A. Jog, A. Carass, D. L. Pham, and J. L. Prince).
Convolution Neural Networks for MS Lesion Segmentation. MORF is an automated algorithm to segment WML in MR images
using multi-output random forests. The work is similar to Geremia
84
A. Carass et al. NeuroImage 148 (2017) 77–102
et al. (2011) in that it uses binary decision trees that are learned from 5. Discussion and conclusions
intensity and context features. However, instead of predicting a single
voxel, an entire neighborhood or patch is predicted for a given input 5.1. Inter-rater comparison
feature vector. The multi-output decision trees implementation has
similarities to output kernel trees (Geurts et al., 2007). Predicting As organizers, we felt that the overall performance of our two raters
entire neighborhoods gives further context information such as the relative to each other was disappointing (see Table 2). For example, the
presence of lesions predominantly inside WM, which has been shown inter-rater Dice overlap of 0.6340 was below that of other inter-rater
to improve patch based methods (Jog et al., 2017). This approach was
originally presented in Jog et al. (2015). Geremia et al. (2011) finished Table 3
first at the 2008 MICCAI MS Lesion challenge (Styner et al., 2008), and Mean, standard deviation (SD), and range of Dice overlap scores for the Segmentations
thus this should represent a good proxy for that work. against the Consensus Delineation. The Segmentations are ranked by their mean Dice
. overlap.
A topology-preserving approach to the segmentation of brain
Dice
images with multiple sclerosis lesions.
(N. Shiee, P.-L. Bazin, A. Ozturk, D. S. Reich, P. A. Calabresi, and D. Method Mean (SD) Range
L. Pham).
0.670 ( ± 0.178) [0.246, 0.843]
Lesion-TOADS (Shiee et al., 2010) is an atlas-based segmentation
technique employing topological and statistical atlases. The method
builds upon previous work (Bazin and Pham, 2008) by handling lesions
as topological outliers that can be addressed in a topology-preserving 0.658 ( ± 0.149) [0.218, 0.852]
85
A. Carass et al. NeuroImage 148 (2017) 77–102
Fig. 5. Each subplot shows the range of values for the true positive rate (TPR) and the
positive predictive value (PPV) between eight of the Segmentations and the Consensus
Delineation. The top plot shows the top eight Segmentations as ranked by the Dice
overlap, and the bottom plot shows the remaining eight Segmentations. The desirable
point on each of the subplots is the upper right hand corner, where TPR is 1 and PPV is 1.
Fig. 4. The plot shows a least squares linear regression fit between the lesion load
estimated by each of the Segmentations and that from the Consensus Delineation. The
dashed line represents a line of unit slope. All volumes are in mm3. #2 is closest to the line of unit slope suggesting that it may be a proxy
for the lesion load as represented by the Consensus Delineation. We do
note that the Consensus Delineation, as generated by STAPLE, may be
overly inclusive—which would explain the grouping together of all the
studies of MS lesions: Zijdenbos et al. (1994) reports a mean inter-rater other Segmentations in Fig. 4. Fig. 5 shows a cross-hairs plot of the
Dice overlap of 0.700, they refer to Dice overlap as similarity. They also range of true positive rate (TPR) versus the positive predictive value
note that when restricted to the same scanner their inter-rater Dice (PPV). The desired operating point for any segmentation in this plot is
overlap rose to 0.732—as reported earlier all of our data was acquired the upper right hand corner (TPR=1, PPV=1). Rater #2 is the closest to
on the same scanner. However, the 2008 MICCAI MS Lesion challenge this desired operating point, with Rater #1 second, and Team PVG One
(Styner et al., 2008) had two raters repeat ten of the scans and their third. A two-sided Wilcoxon Signed-Rank Paired Test comparing the
inter-rater mean Dice overlap was 0.2498. We therefore believe that distance from the operating points to the desired operating point
our inter-rater performance is acceptable, especially considering our between either Rater #2 or Rater #1 and Team PVG One had p-values
raters worked on 82 data sets—61 in Test Set A and B, and another 21 of 0.0058 and 0.0458, respectively. Again, suggesting that the level of
in the Training Set. expertise is critical in achieving the best results. A similar analysis of
Fig. 6 shows that MV-CNN operates closest to the desired optimal point
5.2. Consensus delineation (in this case the lower right hand corner), with Rater #1 second, and
Lesion-TOADS third. Moreover, the two-sided Wilcoxon Signed-Rank
The Consensus Delineation afforded us the opportunity to directly Paired Test has a p-value of <0.0001 between MV-CNN and Rater #1.
compare the quality of our two manual raters with the submitted This suggests that MV-CNN may be better than manual delineators
results. When performing statistical comparisons we use an α level of when it comes to LFPR and LTPR. Tables 4 and 5 show other metrics
0.001. If the Dice overlap (Table 3) is considered the definitive metric that are generally not reported in the lesion segmentation literature.
for rating lesion segmentation then the expert human raters are still However, both of which suggest advantages to the use of algorithms
better than algorithms. However, the level of expertise is important, we over manual raters. For the average symmetric surface distance
note that Rater #2 has a decade more delineation experience than (ASSD); of the algorithms that rank above Rater #2, only Team PVG
Rater #1. A two-sided Wilcoxon Signed-Rank Paired Test (Wilcoxon, One is statistically significantly different with a p-value of <0.0001.
1945) between the highest ranking algorithm (Team PVG One) and the However, with respect to longitudinal correlation (see Table 5) none of
highest ranking manual delineation (Rater #2) reaches significance the algorithms are statistically significantly better than the highest
with a a p-value of 0.0093. Whereas, the same test, between Team PVG rated manual delineation, which comes from Rater #2. Based on the
One and Rater #1 does not reach significance (p-value of 0.1076). Of comparison to the Consensus Delineation, there is not clear evidence to
course Dice overlap is a crude metric with volumetric insensitivities. suggest that any of the automated algorithms is better than the manual
From Fig. 4, we can see that the least squares linear regression of Rater
86
A. Carass et al. NeuroImage 148 (2017) 77–102
Table 4
Mean, standard deviation (SD), and range of the average symmetric surface distance
(ASSD) for the Segmentations against the Consensus Delineation. The Segmentations are
ranked by their mean ASSD.
ASSD
87
A. Carass et al. NeuroImage 148 (2017) 77–102
Table 5 set available and providing an automated site for method comparisons,
Mean, standard deviation (SD), and range of the longitudinal correlation (LongCorr) for the Challenge data will foster new efforts and developments to further
the Segmentations against the Consensus Delineation. The Segmentations are ranked by
improve algorithms and increase detection accuracy.
their mean LongCorr.
The results of the Consensus Delineation suggest that there is still
LongCorr work to be done before we can stop depending on manual delineations
to identify WMLs. This is a disappointing state, considering the amount
Method Mean (SD) Range of research that has been done in this area in the last two decades. The
0.657 ( ± 0.483) [−0.583, 0.997] situation is made worse when considering the shortcomings of the
manual delineations and the automated algorithms. Clearly longitudi-
nal consistency is an area in which all the automated algorithms could
0.607 ( ± 0.582) [−0.693, 1.000]
improve. Our human raters were blinded to the temporal ordering of
the data, unlike the algorithms, and it is not clear at this juncture how
the human raters performance might have changed given this informa-
tion. Of course, this only covers what we would reasonably expect WML
0.432 ( ± 0.524) [−0.567, 1.000]
segmentation algorithms to do today. We should expect them to be able
to classify the three types of WML (enhancing, black hole, and T2-w)
and localize as periventricular or cortical lesions, eventually providing
0.424 ( ± 0.634) [−0.763, 0.998]
more specific location classifications such as juxtacortical, leukocorti-
cal, intracortical, and subpial. These properties may be important in
distinguishing the status of patients. The next generation of MS lesion
0.421 ( ± 0.615) [−0.919, 0.999]
detection software needs to address these issues.
An issue which we had not intended to explore was the failure of
global measures. Lesion load—as determined by lesion segmentation—
0.402 ( ± 0.634) [−0.974, 0.990] is an important clinical measure; the reduction (or stabilization) of
which through automated or semi-automatic image analysis methods is
one of the primary outcome measures to determine the efficacy of MS
0.376 ( ± 0.654) [−0.636, 1.000] therapies. Lesion load and several other global measures fail to predict
the disease course; instead we need to use location specific measures—
as mentioned above—to serve as outcome predictors or staging criteria
0.340 ( ± 0.623) [−0.955, 0.998] for monitoring therapies (Filippi et al., 2014). Beyond this there is a
desire for measures that help in identifying the pathophysiologic stages
of MS lesions (pre-active, active, chronic active, or chronic inactive)
0.327 ( ± 0.679) [−0.900, 0.970] (Jonkman et al., 2015).
Acknowledgements
0.220 ( ± 0.728) [−0.981, 0.984]
This work was supported in part by the NIH/NINDS grant R01-
NS070906, by the Intramural Research Program of NINDS, and by the
0.181 ( ± 0.540) [−0.785, 0.976] National MS Society grant RG-1507-05243. Prizes for the challenge
were furnished by the National MS Society.
Contributors to the Challenge had the following support: F. Prados
is funded by the National Institute for Health Research University
0.171 ( ± 0.634) [−0.899, 0.969]
College London Hospitals Biomedical Research Centre (NIHR BRC
UCLH/UCL High Impact Initiative-BW.mn.BRC10269); C.H. Sudre is
funded by the Wolfson Foundation and the UCL Faculty of
0.153 ( ± 0.746) [−0.930, 1.000]
Engineering; S. Ourselin receives funding from the EPSRC (EP/
H046410/1, EP/J020990/1, EP/K005278), the MRC (MR/J01107X/
1), the NIHR Biomedical Research Unit (Dementia) at UCL and the
0.042 ( ± 0.675) [−0.999, 0.991]
NIHR BRC UCLH/UCL (BW.mn.BRC10269); Teams CMIC and TIG
were supported by the UK Multiple Sclerosis Society (grant 892/08)
and the Brain Research Trust.
0.031 ( ± 0.683) [−0.980, 0.999]
The following protocol was used in the creation of the MS lesion masks, which were created in our 1 mm isotropic MNI space.
88
A. Carass et al. NeuroImage 148 (2017) 77–102
1. Review the possibilities for presentation of MS lesions in brain scans, an excellent resource is Sahraian and Radue (Sahraian and Radue, 2007).
It is also a good idea to familiarize yourself with the paint and mask functions in MIPAV (McAuliffe et al., 2001; Bazin et al., 2005) before you
begin, although this protocol description can serve as a basic guide.
2. Open MIPAV. If you have not done so in the past, add the Paint toolbar to the interface (Toolbars > Paint toolbar). The Image toolbar should be
present by default; if not, go to Toolbars > Image toolbar .
3. Open two copies of the FLAIR scan and one copy each of the T1-w, T2-w, and PD-w scans (File > Open image (A) from disk > select files > Open).
These should be appropriately co-registered in the axial view with identical slice thickness and field of view values before beginning this process.
4. Click on the T1-w scan to select it, then click on the WL button on the MIPAV toolbar to bring up the Level and Window adjuster. This should
automatically result in a reasonable tissue contrast for viewing potential lesions on the T1-w. If the contrast is inadequate, change the window
and level settings. Close the tool when you are satisfied with the contrast.
5. Enlarge each image three times using the magnifying glass + button on the Image toolbar for a total magnification of 4×. Arrange the images on
your display with the two FLAIR copies next to each other. Ensure that the scans are properly aligned with one another horizontally. This will
enable you to quickly check the other images to identify and verify tissue abnormalities as lesions (or not) while working on the FLAIR mask.
6. Link the scans together by first clicking on the Sync slice number button on the Image toolbar (two arrows one pointing left and the other
pointing right). Then click on each scan and select the Link images button (broken links next to Sync slice number button; the broken links
change to an intact link when activated). This will ensure that all of the scans stay on the same slice as the FLAIR while you work. Click on one of
the scans, then scroll up and down while looking from side to side over the images to verify proper registration and check for image processing
errors (e.g., missing pieces of brain).
7. Select one of the FLAIR scans, then click on the Paint Grow button (looks like a paint bucket) on the Paint toolbar to open the intensity and
connectivity-based paint mask generator.
8. Open the Paint Power Tools plugin. The icon (lightbulb) should be at the right end of the Paint toolbar. Look at the Threshold section. Find the
maximum intensity value present in the scan by observing the number in the right-hand box (upper threshold).
9. Look at the Paint Grow tool. Find the section marked Set maximum slider values. Change the maximum slider values in the paint mask
generator to reflect the maximum intensity in the scan, and click Set.
10. Choose a lesion with well-defined borders and strong hyperintensity on the FLAIR scan. Click on the most hyperintense area in the lesion.
11. Move the second slider (Delta below selected voxel intensity) to the right until it encompasses most of the lesioned tissue.
12. Scroll up and down through the image to ensure that the selection is limited to the lesioned area and does not include hyperintensities due to
noise or artifact. If non-lesioned tissue is included, move the slider back to the left until this tissue is deselected.
13. Move the first slider (Delta above selected voxel intensity) to the right to ensure that all voxels of higher intensity in the lesion are selected.
14. Repeat this process until all well-defined lesions in the FLAIR scan have been selected, remembering to scroll up and down frequently to prevent
masking of non-lesioned tissue. In general, this process will result in a rough draft of a lesion mask.
Do not use this process for any area that is affected by scan artifact or for any hyperintensity that is not clearly a lesion. Investigate
questionable areas during the later stages of the delineation process.
If a decision is to be made between fully encompassing a lesion and additional non-lesioned tissue or partially covering a lesion without
extraneous tissue, choose the latter option. It tends to be easier, within MIPAV, to add to a mask than subtract from it.
15. When you are satisfied with your rough mask, save it as an unsigned byte mask (VOI > Paint conversion > Paint to Unsigned byte mask). This will
give you the binary mask data you have generated thus far. When the mask image appears, go to the File menu and choose Save image as. Enter
the file name and desired extension, then click Save.
16. Close the binary mask and the Paint Grow tool.
17. To begin, move your pointer over the area around the edge of a lesion, hold the mouse button down, and notice the intensity difference between
the interior and exterior of the lesion. Record the intensity value for the area at the edge of the lesion.
Because many MS lesions are found in close proximity to the ventricles, it is useful to start in the middle slice in the axial view. Delineate
from the middle axial slice to the superior aspect of the brain, scroll down to check your work, and then delineate from the middle to the inferior
view.
18. On the Paint Power Tools interface, click the box next to Threshold. Enter the intensity value for the outside edge of the lesion in the first box;
this will restrict your paint to voxels between that intensity (lower threshold) value and the value listed in the box to the right (upper threshold).
There is no need to change the value in the right box unless you are delineating lacunes. In that case, you should set the left box to the lowest
possible value, and change the upper threshold to the highest value found on the edge of the lacune.
19. Click on the paintbrush icon on the paint toolbar. Paint around the edge of the lesion to test your threshold. You may need to paint and erase
(paint=left mouse button, erase=right mouse button) the first time you do it, and then the threshold should be activated. You may also need to
adjust the lower threshold value (left box). If too many voxels are being excluded from the lesion mask, lower the threshold value for a more
inclusive range. If too many voxels are being included, increase the threshold value.
20. If you wish, you can change the paintbrush size by clicking on the drop-down menu in the center of the Paint toolbar and selecting one of the
options.
You may also customize your paintbrush options by clicking on the Paint brush editor button (looks like a group of paintbrushes) to the right
of this menu. This will open a grid size selector that allows you to specify the width and height of the grid for your paintbrush in pixels (default is
12×12). Click OK, and the grid will open. Draw the shape you want for your paintbrush, then go to the “Grid options” menu to save
(Grid options > Save paint brush > input file name > Save). Your custom paintbrush will appear in the menu the next time MIPAV is opened, so
restart the program if you want to use it immediately.
21. As you move to different slices, you may need to readjust the lower threshold. Not all of the lesion edges have the same intensity value, and
intensities often differ between lesions at the anterior vs posterior areas of the same slice.
22. During this process, it is extremely important to scroll up and down frequently in order to get a sense of each lesion's shape and ensure mask
continuity. For every hyperintensity identified, scrolling up and down can also help to rule out false positives. Be sure to look at the other scans,
particularly the T2-w, in order to verify that what you are selecting is a lesion.
In some cases, a lesion may be much more readily visible on the T2-w scan. If this occurs, it is possible to delineate that portion directly on
89
A. Carass et al. NeuroImage 148 (2017) 77–102
the T2-w scan and add this small mask area to your FLAIR mask. This is particularly relevant when the FLAIR image contains a great deal of
artifacts. If you cannot adequately capture the lesion on the FLAIR, use the T2-w.
23. When debating what to include in the mask, keep these things in mind.
(a) Lesions usually have rounded or smoothed edges.
(b) Lesions appear distinctly hyper- or hypo- intense when compared with surrounding tissue (usually hyperintense on FLAIR, PD-w, and T2-w
scans, and hypointense on T1-w),
(c) Lesions will usually be found near the ventricles, in the corpus callosum, or in the deep white matter, though juxtacortical lesions are not
uncommon.
(d) Lesions may appear in the cerebellum, brainstem, temporal lobes, or basal ganglia at a lower intensity relative to the majority of the lesions.
It is especially important to use information from the other scans when attempting to detect and delineate lesions in these areas.
(e) Include white matter encompassed by closed, well-defined clusters of lesions. Do not include internal white matter if the cluster is open.
(f) Include all CSF inside lacunes.
(g) If a lesion is adjacent to clearly hyperintense areas near the ventricles, and you can confirm that these areas appear damaged in the T1-w
scan, include them in the mask. Lesioned tissue bordering the ventricles looks ragged and dark on T1-w scans.
(h) Do not include diffusely abnormal white matter (DAWM) in the masks for ISBI scans. The intensity of DAWM is between normal white
matter and lesioned tissue on FLAIR. DAWM looks mottled on T1-w, may radiate outward like a halo from a focal lesion, and is usually
found around the ventricles.
24. Save your work frequently or use the automatic save function in the Paint Power Tools interface. Check the box next to Auto save under Misc.,
then set the number in the box to reflect how often you want the mask to be automatically saved (default is 10 minutes).
25. For some lesions, you may need to turn the paint threshold off and use the standard paint option, which will not restrict your paint to any
specific intensity values. To do this, simply uncheck the box next to Threshold.
26. When you have finished delineating the lower portions of the brain, go back through the entire scan and check your work against the other
images, focusing specifically on any areas that may have been difficult to verify as lesions. Edit as necessary.
27. Save your final mask.
28. To load a mask that you have worked on previously, select the FLAIR scan, then click on the second button from the left on the paint toolbar
(appears to be a folder opening with a four-square gradient in front of it). Choose your mask file, click Open, and your mask will be loaded over
the FLAIR.
29. If you would like to edit your mask after opening it from a saved file, open the Power Paint Tools, click on Mask to Paint under the Import/
Export section at the bottom of the interface, and continue working.
Appendix B. Methods
For completeness, we provide descriptions of the Challenge Participants in B.1 and in B.2 we describe other methods that were not part of the
Challenge which we included in our evaluation. Where we present descriptions or results of the methods, we use a colored square to help identify
methods and within that square we denote methods that are unsupervised with the letter U and those that require some training data (supervised
methods) with the letter S.
Table B1 provides a synopsis of these methods and the MR sequences used by each individual team during the Challenge.
.
Multi-Contrast PatchMatch Algorithm for Multiple Sclerosis Lesion Detection.
(F. Prados, M. J. Cardoso, N. Cawley, O. Ciccarelli, C. A. M. Wheeler-Kingshott, & S. Ourselin).
Team CMIC used the PatchMatch (Barnes et al., 2010) algorithm for MS lesion detection. The main contribution of this work is the
generalization of the optimized PatchMatch algorithm to the context of MS lesion detection and its extension to multimodal data.
The original PatchMatch algorithm was designed to look for similarities between two 2D patches within the same image (Barnes et al., 2010).
Later, the Optimized PAtchMatch Label (OPAL) fusion approach extended patch correspondences between a target 3D image and a reference library
of 3D training templates (Ta et al., 2014). Here, the PatchMatch algorithm is used to locate pathological regions through the use of a template
library comprising a series of multimodal images with manually segmented MS lesions. By matching patches between the target multimodal image
and the multimodal images in the template library, PatchMatch can provide a rough estimate of the location of the lesions in the target image.
OPAL uses the sum of the squared differences (SSD) between two patches over one single modality to measure patch similarity. This is replaced
with an l2-norm over the multimodal patches, which are assumed to be in the same space. To improve computational speed, as in the original OPAL
method, the computation of the patch similarity is stopped if the current sum is superior to the previous minimal multimodality SSD. As this
PatchMatch algorithm has a non-binary output, an adaptive threshold value is used to binarize the probabilistic mask. A robust range (with 2%
outliers on both tails) of all voxels with non-zero probabilities is calculated, and then the mean of the values inside the robust range is computed.
This mean is then used as the threshold to binarize the probabilistic segmentation. Finally, if the highest probability within the robust range is below
0.1 the method assumes that no lesions have been detected, meaning that the patient is lesion-free.
.
Automatic Graph Cut Segmentation of Multiple Sclerosis Lesions.
(L. Catanese, O. Commowick, and C. Barillot).
Team VISAGES GCEM uses a robust Expectation-Maximization (EM) algorithm to initialize a graph, followed by a min-cut of the graph to detect
lesions, and an estimate of the WM to help remove false positives. GCEM stands for G raph-c ut with E xpectation-M aximisation.
A region of interest is defined based on the thresholded T2-w image. Each voxel within the region of interest is represented in a graph and
connected to two terminal nodes, known as the source and sink, which respectively represent the object class for MS lesions and normal appearing
brain tissues (NABT). Spatially neighboring nodes are connected by n-links weighted by boundary values that reflect the similarity of the two
90
A. Carass et al. NeuroImage 148 (2017) 77–102
Table B1
An overview of the methods and data used by the Challenge participants. We denote methods that are unsupervised with the letter U and those that require some training data
(supervised methods) with the letter S.
considered voxels. The contour information contained in the n-links weights is computed using a spectral gradient (García-Lorenzo et al., 2009).
The regional term represents how the voxel fits into the given models of object and background. The edges between a node of the image and the
terminal source and sink nodes are called t-links. Normally these models are estimated using seeds given as manual input. Instead, the Team uses
an automated version of the graph cut where the object and background seeds for the initialization are computed from the images. To do so a 3-class
multivariate GMM is employed, representing CSF, GM, and WM with lesions being treated as outliers to these three classes.
The seeds are estimated using a robust EM algorithm (García-Lorenzo et al., 2011), which optimizes a trimmed likelihood in order to be robust
to outliers. The algorithm then alternates between the computation of the GMM parameters and the % of outlier voxels. From the GMM NABT
parameters, the Mahalanobis distance is computed of each voxel to each of the classes in the GMM NABT model. This distance is then used to
compute a p-value for determining the probability of each voxel belonging to each of the three classes. For each voxel i its smallest p-value pi is
retained. As the sinks represent voxels that are close to NABT, the t-link weights Wbi are defined as Wbi = 1 − pi . To help distinguish MS lesions from
other outliers (vessels, etc.), the fact that MS lesions are hyperintense compared to WM in T2-w sequences is used. A fuzzy logic approach is used to
model this based on the previously computed model of GMM NABT, which determines fuzzy weights from which the corresponding t-link weights
are computed, see García-Lorenzo et al. (2009) for complete details.
The MS lesions are assumed to appear surrounded by WM and not adjacent to the cortical mask border. Any candidate lesions that violate either
of these criteria are removed. Finally, all candidate lesions smaller than 3 mm3 are discarded.
.
Sparse Representations and Dictionary Learning Based Longitudinal Segmentation of Multiple Sclerosis Lesions.
(H. Deshpande, P. Maurel, and C. Barillot).
Team VISAGES DL used sparse representation and a dictionary learning paradigm to automatically segment MS lesions within the longitudinal
MR data. Dictionaries are learned for the lesion and healthy brain tissue classes, and a reconstruction error-based classification approach for
prediction.
Modeling signals using sparse representation and a dictionary learning framework has achieved promising results in image classification
(Deshpande et al., 2015; Mairal et al., 2009; Roy et al., 2014a, 2015b; Weiss et al., 2013). Sparse coding finds a sparse coefficient vector a ∈ k for
representing a given signal x ∈ n using a few atoms of an over-complete dictionary D ∈ k × n . The sparse representation problem is represented as
mina ∥ a ∥0 such that ∥ x − Da ∥22 ≤ ϵ where ϵ is the error in the representation. This l0 problem can be more efficiently solved as the l1 minimization
problem
91
A. Carass et al. NeuroImage 148 (2017) 77–102
min ∥ x − Da ∥22 + λ ∥ a ∥1 ,
a
where λ balances the trade-off between error and sparsity. For a set of signals {x}1m , a dictionary D is found from the underlying data such that each
signal is sparsely represented by a linear combination of atoms,
m
minm ∑ ∥ xi − Dai ∥22 + λ ∥ ai ∥1 .
D,{ai }1
1
The optimization is an iterative two-step process involving sparse coding with a fixed D followed by a dictionary update for fixed atoms {ai}1m .
The following preprocessing steps are used in the approach. Artifacts in the Challenge data are removed through denoising the images using a
non-local means approach (Coupé et al., 2008). The images are then linearly rescaled to the range [0, 255] followed by a longitudinal intensity
normalization (Karpate et al., 2014). A leave-one-out cross-validation experiment was used to determine an optimal patch size of 5×5×5. Patches of
this size were then extracted and rasterized centered on every second voxel in the input images, this was done to reduce the computational
complexity inherent in using every voxel. Patches in the training data are determined to belong to either the healthy tissue class or the lesion class,
based on the manual delineations. Patches are finally normalized to limit their individual norms below or equal to unity. From the training data
class-specific dictionaries are learned for the two classes.
Given a test patch, the patch classification is performed in two steps: First the sparse coefficients for each class are learned. The test patch is then
assigned to the class with which it has minimum representation error. As the healthy class data represents complex anatomical structures such as
CSF, GM, and WM, it has more variability in comparison to the lesion class. To account for this, the healthy class is allowed to have a larger
dictionary size than the lesion class. As the patches are centered on every other voxel in the image, a majority vote multi-patch scheme is used to
determine the classification of each voxel. All patches that overlap a particular voxel contribute their classification to determine the winner of the
majority voting. The following parameters are used in solving the l1 minimization problem: a sparsity parameter of λ = 0.95 with a dictionary size of
5000 for the healthy tissue class and a size of 700 to 2500 for the lesion class, depending on the total lesion load.
TEAM VISAGES DL performed longitudinal intensity normalization as a pre-processing step to negate the intensity differences across the
different time points for a single MS patient. However, there are large intensity differences across several patients in the provided data set. The
Team believes that improved classification results could be obtained after performing intensity normalization across all patients.
.
Model of Population and Subject (MOPS) Segmentation.
(X. Tomas-Fernandez and S. K. Warfield).
Inspired by the ability of experts to detect lesions based on their local signal intensity characteristics, Team CRL proposes an algorithm that
achieves lesion and brain tissue segmentation through simultaneous estimation of a spatially global within-the-subject intensity distribution and a
spatially local intensity distribution derived from a healthy reference population.
To address the limitations of intensity-based MS lesion classification, the imaging data used to identify lesions is augmented to include both an
intensity model of the patient under consideration and a collection of intensity and segmentation templates that provide a model on normal tissue.
The approach is called a Model of Population and Subject (MOPS) intensities (Tomas-Fernandez and Warfield, 2015). Unlike classical approaches
in which lesions are characterized by their intensity distribution compared to all brain tissues, MOPS aims to distinguish locations in the brain with
an abnormal intensity level when compared with the expected value at the same location in a healthy reference population.
A reference population of fifteen healthy volunteers was acquired including T1-w, T2-w FSE (Fast spin echo), FLAIR-FSE, and diffusion weighted
images on a 3T clinical MR scanner from GE Medical Systems (Waukesha, WI, USA, see Tomas-Fernandez and Warfield, 2015 for details about
acquisition and spatial alignment). The MOPS algorithm combines a local intensity GMM derived from the reference population with a global
intensity GMM estimated from the imaging data. Intuitively, the local intensity model down weights the likelihood of those voxels having an
abnormal intensity given the reference population. Since MRI structural abnormalities will show an abnormal intensity level compared to similarly
located brain tissues in healthy subjects, MS lesions are identified by searching for areas with low likelihood.
.
Longitudinal Multiple Sclerosis Lesion Segmentation using 3D Convolutional Neural Networks.
(S. Vaidya, A. Chunduru, R. Muthuganapathy, and G. Krishnamurthi).
Team IIT Madras modeled a voxel-wise classifier using multi-channel 3D patches of MRI volumes as input. Two convolutional neural networks
(CNNs) were trained, each of which represented one of the trained raters. The final segmentation is obtained by combining the probability outputs
of these two CNNs. Efficient training is achieved by using sub-sampling methods and sparse convolutions.
The provided data is preprocessed such that all subjects and time-points are histogram-matched to the first provided patient and time-point,
then normalized using the mean CSF value, followed by a robust (1%) data truncation (Avants et al., 2011). A voxel-wise classifier is employed to
perform the segmentation task with 3D patches from each of the four channels (T1-w, T2-w, PD-w, and FLAIR) being fed to the classifier. As MS
Lesions only constitute a very small percentage of the MRI volume, the data is sampled to reduce the class imbalance between WML and NABT.
Each image volume is divided into subvolumes of equal size, with patches only selected from those subvolumes that contain lesion voxels greater
than a set threshold. This sampling technique speeds up the training of the CNNs for segmentation by using the sparse convolution method (Li et al.,
2014).
All convolutional layers in the CNN use the softplus activation function, with training done using a logarithmic likelihood as cost function, and
optimization carried out using mini-batch gradient descent with momentum. The CNN consists of four layers with the input being image patches of
size 19×19×19 voxels from each of the four modalities concatenated. The second and third layers consist of 60 filters of size 4×4×4 and 3×3×3
respectively, with the third layer being a multi-layer perceptron (1×1×1×200) and the final output being a binary classification. Two CNNs were
trained for each of the two trained raters, and the posterior probability maps of the lesion class from the CNNs generate the initial prediction of the
MS lesions. As the Challenge is focused on WML, a WM mask is applied to the predictions by registering the test images with a brain template and
removing any lesion predictions that are outside the template WM mask.
.
Hierarchical MRF and Random Forest Segmentation of MS Lesions and Healthy Tissues in Brain MRI.
(A. Jesson an T. Arbel).
92
A. Carass et al. NeuroImage 148 (2017) 77–102
Team PVG One built a hierarchical framework for the segmentation of a variety of healthy tissues and lesions. At the voxel level, lesion and tissue
labels are estimated through a MRF segmentation framework that leverages spatial prior probabilities for nine healthy tissues through multi-atlas
label fusion (MALF). A random forest (RF) classifier then provides region level lesion refinement.
Training consists of three stages: Stage one involves building a set of lesion and healthy tissue atlases, referred to as pathological atlases as they
are based on MS patient data. These are to be used as spatial priors for new test data. Stage two involves performing an initial segmentation of 9
healthy tissue structures in each of the patient training cases in order to build healthy and lesion intensity distributions. Stage three involves
training the RF.
Stage One included the 21 subjects from the Challenge training data with intensity values averaged over several time-points and 20 subjects
from the training data provided by Styner et al. (2008): these are combined to create atlases of pathological tissues. Healthy tissue labels for each
pathological atlas are generated through MALF from multiple labels from 35 subjects from the MICCAI 2012 Grand Challenge on Multi-Atlas
Labeling. The 134 provided labels are concatenated into nine tissue classes: CSF, lateral ventricles, other ventricles, deep GM, cortical GM,
cerebellar GM, WM, cerebellar WM, and brainstem. The NABT tissues labels are augmented by the provided manual delineations to complete the
pathological atlases.
Stage Two involves performing the same procedure as Stage One on the 21 training time-points provided. This leads to a set of healthy and
lesion labels and associated weights, which are used to guide voxel sampling for building intensity distributions of healthy tissues and lesions. Here
intensity distributions of each class are modeled as GMMs.
Stage Three involves determining the labels at each voxel for each time-point of each training subject using the models determined in Stages One
and Two. The resulting segmentations are used to group together lesion voxels into lesion candidates. A regional random forest model (RRF) is then
trained using the distance minimum, mean, and variance of each candidate lesion to each healthy tissue; the size, volume, and solidity of each
candidate lesion; and the principal moments and inertia matrix of the ellipse estimating the shape of each candidate lesion as features.
The MALF estimation of spatial tissue priors uses the rigid and affine components of ANTs (Avants et al., 2008) and the non-linear framework of
MIND (Heinrich et al., 2012, 2013). Label fusion is performed through a regional similarity method, and lesion priors are augmented through
outlier detection. In addition to the preprocessing provided by the challenge, intensity normalization was performed using a sigmoidal function,
where the parameters are determined by the mean and variance of intensities over several regions of interest. To reduce within image artifacts the
data was de-noised based on a non-local means method (Coupé et al., 2008).
.
MS-Lesion Segmentation in MRI with Random Forests.
(O. Maier and H. Handels).
Team IMI trained a RF with supervised learning to infer the classification function underlying the training data. The classification of brain
lesions in MRI is a complex task with high levels of noise, hence a total of 200 trees are trained without any growth-restriction. Contrary to reported
observations, no overfitting occurred.
All data was preprocessed to harmonize each sequences intensity profile by a learning-based intensity standardization method. From each of the
four MRI sequences, the following features are extracted: (1) voxel intensity; (2) voxel intensity after Gaussian smoothing (σ=3, 5, and 7 mm); (3)
three different local histogram configurations; and (4) each voxels' distance to the image center. Features 1–3 provide information about gray-level
values at different scales and mean intensity distributions in small areas around each voxel, see Maier et al. (2015) for a more complete explanation
of these features. The Challenge is concerned with WML, thus a probability based tissue segmentation is obtained (Zhang et al., 2001) on the T1-w
MPRAGE sequence providing probabilities for CSF, GM, and WM. The feature vector is computed with voxel gray value and voxel gray value after
Gaussian smoothing (σ=3, 7, 15, and 31 mm).
Stratified random sampling is employed to extract a representative sub-set from the training data, reducing the amount of training samples and
thus the training time. The original background-to-lesion ratio of each subject is kept intact, leading to an unequal class representation, which has
been found to be advantageous (Maier et al., 2015). To obtain a binary segmentation mask, the RFs probability output is thresholded at a value of
0.4, introducing a slight bias in favor of the lesion class that compensates for the unbalanced class ratio in the training set. Finally, single
unconnected lesion voxels are removed as outliers, holes in binary lesion objects are closed and a single-iteration closing operation with a 3D
square-connected component is applied.
.
Automatic Longitudinal Multiple Sclerosis Lesion Segmentation.
(S. Jain, D. M. Sima, and D. Smeets).
MSmetrix (Jain et al., 2015) is presented, which performs lesion segmentation while segmenting brain tissue into CSF, GM, and WM, with
lesions identified based on a spatial prior and hyperintense appearance in FLAIR.
The lesion segmentation has four stages: brain segmentation, outlier estimation, pruning, and lesion filling. The brain segmentation uses an EM
algorithm to formulate a probabilistic model of CSF, GM, and WM, from the T1-w image. In the outlier estimation step, an outlier class is estimated
from the FLAIR image of the same patient using the three tissue class segmentations from the previous step as prior information. This is also done
with an EM algorithm with the inclusion of an outlier map. The pruning stage segments the lesions in the outlier map, as not every outlier is a lesion.
To differentiate lesions from NABT, some additional a priori information about the location and the appearance of the lesions is incorporated.
Lesions need to be in the WM region and the underlying intensities of the outliers should be hyperintense compared to the GM intensities from the
FLAIR. Finally, the lesion segmentation is used to fill in the lesions in the bias corrected T1-w image with WM intensities. These four stages are
repeated until convergence and the lesion segmentation is produced as an output.
Each time-point was initially processed independently with a subsequent temporal consistency correction, similar to Xue et al. (2006). The
temporal consistency, Ci t , for a voxel i at time-point t is defined based on its temporal neighborhood 5 iTemp t ∈ {t − 1, t , t + 1} as
δ N Temp
it
Ci t = 1 − ,
|5 iTemp
t | −1
where δ N Temp is the number of times the segmentation label changes in 5 iTemp
t . The label at voxel i for time-point t, Li t , is defined based on the
it
temporal consistency of its 3×3×3 spatial neighborhood, 5 Spa
i t , as follows,
93
A. Carass et al. NeuroImage 148 (2017) 77–102
⎧ 1 T
⎪ Li t if ∑t =1 Ci t ≥ 0.5,
⎪ T
⎪ ⎛⎧
Li t = ⎨ ⎫T ⎞
⎪ mode⎜⎪ ⎨
⎪ ⎟
⎪ ⎜ argmax Cj t⎬ ⎟ otherwise.
⎪ ⎜ ⎪ Spa ⎪ ⎟
⎩ ⎝ ⎩ j ∈ 5 jt ⎭ t =1⎠
Thus, if the consistency is high enough, the labels remain unchanged by the temporal consistency; otherwise, it is replaced with the modal value of
the segmentation labels of its most consistent neighbors.
Convolution Neural Networks for MS Lesion Segmentation.
(M. Ghafoorian and B. Platel).
Team DIAG utilizes a deep CNN with five layers in a sliding window fashion to create a voxel-based classifier.
The image intensity is normalized using a 95th percentile with values at and above that set to 1; all values below that are linearly rescaled in the
range [0, 1]. The CNN learns to label n × n patches indicating if the central voxel is a lesion or NABT. A leave-one-out cross validation is employed to
provide training data for the CNN. While sampling from a patient, all available time-points and all possible lesion patches are used. An equal
number of NABT patches are randomly chosen to ensure balance between the two classes in the training data. The approximate final sizes of the five
created training data sets are 430k, 320k, 540k, 570k, and 560k respectively. No data augmentation methods have been applied to artificially
increment the size of the data. Since human experts are usually better at specificity than sensitivity, the logical OR operation is used to create a
better reference standard from the two provided human expert annotations.
To classify the image patches, a five layer CNN is trained that takes 32×32 patches from the available four channels (T1-w, T2-w, PD-w, &
FLAIR) as its input samples. There are four convolutional layers with rectified linear non-linearities that have respectively 15 filters of size 13×13,
25 filters of size 9×9, 60 filters of size 7×7, and finally 130 filters of size 3×3. Pooling is not used since it results in a sort of translation invariance
that is not desirable for a classifier that assigns the label of the whole patch to its central voxel. A final logistic regression model classifies the
resulting responses to the filters in the last convolutional layer. Stochastic gradient descent is used for the optimization with a batch size of 64 and a
learning rate of 0.0001. We run the optimization for 50 epochs and pick the best classifier based on the validation set misclassification rate.
.
Model Selection Propagation for Application on Longitudinal MS Lesion Segmentation.
(C. H. Sudre, M. J. Cardoso, and S. Ourselin).
Based on the assumption that the structural anatomy of the brain should be temporally consistent for a given patient, Team TIG proposes a
lesion segmentation method that first derives a GMM separating healthy tissues from pathological and unexpected ones on a multi-time-point intra-
subject group-wise image. This average patient-specific GMM is then used as an initialization for a final time-point specific GMM from which final
lesion segmentations are obtained.
The proposed model can be divided into four major steps. First, the provided T1-w and T2-w data are rigidly registered to the FLAIR image of
each time-point. ICBM atlases are also aligned to the transformed T1-w image and used as an initialization for a three modalities EM segmentation
in a framework that not only corrects for any possible remaining bias field but also for an initial separation between inliers and outliers. This is done
on log-transformed and bounded intensities. The second step creates an intra-subject multi-time-point group-wise average. This is performed
through an iterative set of affine registrations refined afterwards by non-rigid deformations (Modat et al., 2014). To standardize the intensity
information, histogram-matching is progressively performed between the individual time-points and the group-wise image using only the model
inliers and applying a polynomial fit of degree 2. The intensity matching allows for a direct transfer of the selected group-wise model to each specific
time-point. The third step involves running a GMM on the matched group-wise images (T1-w, T2-w, and FLAIR). The number of classes to correctly
model the inlier and outlier components of the four main anatomical regions (CSF, GM, WM, and non-brain) is determined automatically, by
finding a balance between model fit and complexity. Once the final model converges, one can obtain a group-wise tissue segmentation and an inlier/
outlier classification. To finalize the result, the group-wise tissue segmentation is transformed back to each time-point and subsequently smoothed
out using a Gaussian filter. For each time-point, this smoothed segmentation is used as a prior for a new GMM model fit improving on the inlier/
outlier separation. The lesion extraction process relies simply on the choice of the relevant component from the outlier part of the model based on
the location and intensity heuristics.
Team TIG submitted new results after the completion of the Challenge to address a bug in their code, the second submitted results are denoted
TIG BF. Both sets of results are reported in Table C3. However, only the originally submitted results are included in Tables C.1 and C.2.
Table B2 provides an overview of these methods and the data they use.
.
Automatic White Matter Hyperintensity Segmentation using FLAIR MRI.
(L. O. Iheme and D. Unay).
BAUMIP is a method based on intensity thresholding and 3D voxel connectivity analysis. A simple model is trained that is optimized by
searching for the maximum obtainable Dice overlap.
Firstly, a mapping is constructed of the intensities of every training image to those of a reference image, which in this case is the first time-point
for the first subject. The histogram of the whole brain foreground voxels is computed from the FLAIR image, with the assumption that the peak is
that of a normal distribution so that its 7 dB drop is more than twice its Full Width at Half Maximum (FWHM). The intensity I7 dB of this point is
guaranteed to be amongst the highest intensity values of the image. With this value as a minimum threshold for the WM hyperintensity, the
threshold is defined as
T = IPeak (1 − w ) + I7 dB,
where w is a to be determined weight. Voxels that exceed this threshold are segmented as WM lesions. For a more detailed description and
evaluation of the method, see Iheme et al. (2013).
94
A. Carass et al. NeuroImage 148 (2017) 77–102
Table B2
An overview of the methods and data used by the eleventh Challenge submission, and three state-of-the-art methods. We denote methods that are unsupervised with the letter U and
those that require some training data (supervised methods) with the letter S.
The 3D connectivity analysis involves examining every detected voxel for the degree of connectivity with each of its neighboring voxels. This is
equivalent to analyzing the volumetric significance of every detected lesion. The training data was used to determine a minimum volume for lesions;
connected components that are below this volume threshold are deemed insignificant and assumed to be false positives. To further reduce the
incidence of false positives at the corpus callosum, the interhemispheric fissure is estimated using a RANSAC-based approach (Ekin, 2006). Lesions
that fall within a prescribed distance of the interhemispheric fissure are also removed as false positives.
.
Multi-Output Random Forests for Lesion Segmentation in Multiple Sclerosis.
(A. Jog, A. Carass, D. L. Pham, and J. L. Prince).
MORF is an automated algorithm to segment WML in MR images using multi-output random forests. The work is similar to Geremia et al.
(2011) in that it uses binary decision trees that are learned from intensity and context features. However, instead of predicting a single voxel, an
entire neighborhood or patch is predicted for a given input feature vector. The multi-output decision trees implementation has similarities to output
kernel trees (Geurts et al., 2007). Predicting entire neighborhoods gives further context information such as the presence of lesions predominantly
inside WM. This approach was originally presented in Jog et al. (2015).
From the co-registered T1-w, T2-w, FLAIR, and expert manual delineations, 3×3×3 sized patches for each voxel i are extracted. Small patches
provide local context for a particular voxel with the patch for the manual segmentation being the desired output of the multi-output decision trees.
The multi-modality intensity features are augmented with a global context for each voxel i consisting of the mean intensity of a large window (of size
11×11×3) calculated at a fixed radial distance from i and multiple angles within the axial plane. The final feature vector is created by concatenating
the local intensity patches from the three modalities (T1-w, T2-w, & FLAIR) and global context features, and xi is used to denote the feature vector
of i (see Jog et al., 2015 for complete details).
Learning a multi-output random forests is similar to the random forest algorithm (Breiman, 2001). With independent vectors, xi , and dependent
vectors, yi , which are the 3×3×3 patch of the manual delineation. Given a node q in a decision tree, with training samples Θq = {[x1 ; y1 ], …, [xm ; ym ]}
and the mean of the dependent vectors denoted by yq , then the squared distance from the mean is computed as
m 27
⎛ ⎞2
∑ ∑ ⎜⎝ykj − yqj ⎟ .
⎠
k =1 j =1
For a particular feature, f, and threshold πf, the data in q (Θq) are separated into two disjoint sets ΘqL = {[xi ; yi ]| ∀ i , xif ≤ πf } and
ΘqR = {[xi ; yi ]| ∀ i , xif > πf }. f and πf are chosen such that the combined squared distance of the two daughter nodes qL and qR of q is minimized.
To predict a lesion segmentation on a new image, the local and global context features from the T1-w, T2-w, and FLAIR images are constructed
as mentioned above. The trained multi-output tree ensemble is applied to each extracted feature vector. The input vector travels through the tree as
its features are evaluated against the ones in the tree nodes, until it lands in a leaf node. Leaf nodes consist of at least 50 training samples, each a 27-
dimensional label vector. These label vectors provide a percentage of lesion voxels. The output from the multi-output decision ensemble is smoothed
using a Gaussian filter with σ=1. This smoothed membership image is thresholded to create a binary lesion mask. A 3-class fuzzy k-means
segmentation (Bezdek, 1980) of the T1-w image provides an initial WM mask. Lesions inside WM are labeled as GM in this 3-class fuzzy k-means
segmentation, thus forming holes in the initial WM mask. Therefore, MORF fills the initial WM mask and regards any lesions found outside the
filled WM mask as false positives; these lesions are removed from the final MORF output.
.
A topology-preserving approach to the segmentation of brain images with multiple sclerosis lesions.
(N. Shiee, P.-L. Bazin, A. Ozturk, D. S. Reich, P. A. Calabresi, and D. L. Pham).
Lesion-TOADS is an atlas-based segmentation technique employing topological and statistical atlases. The method builds upon previous work
(Bazin and Pham, 2008) by handling lesions as topological outliers that can be addressed in a topology-preserving framework when grouped
together with the underlying tissues.
A complete description of Lesion-TOADS is available in Shiee et al. (2010), as it represents a continuation of the development of TOADS (Bazin
and Pham, 2008); a brief review of that work is provided here. TOADS segments the brain into several major structures (sulcal CSF, ventricular CSF,
cortical GM, cerebral WM, cerebellar GM, cerebellar WM, putamen, thalamus, caudate, and brainstem) and Lesion-TOADS introduces the
delineation of WML. TOADS incorporates statistical and topological atlases with a fuzzy clustering framework giving topologically consistent
95
A. Carass et al. NeuroImage 148 (2017) 77–102
segmentation of healthy brain anatomy. A topologically consistent hard segmentation of the brain is initialized from a topological atlas and used to
modulate the influence of similar intensity clusters that are non-contiguous. The statistical and topological atlases are rigidly registered to the MR
image initiating an iterative process alternating between intensity based tissue segmentation and topology preserving fast marching. Lesion-TOADS
augments TOADS by handling the union of WML and WM as a topological consistent object, with both WML and WM having the same spatial prior.
Other improvements of Lesion-TOADS over TOADS include: 1) redefining the cluster distance function to account for the intensity profile of WML
to help distinguish it from the partial volume mix of GM & WM, or CSF & WM, which can cause false positives; and 2) multichannel weights to
take advantage of the discriminative power of FLAIR images in distinguishing WML from NABT.
.
Multi-View Convolutional Neural Networks.
(A. Birenbaum and H. Greenspan).
MV-CNN is a method based on a Longitudinal Multi-View CNN (Roth et al., 2014). The classifier is modeled as a CNN, whose input for every
evaluated voxel are patches from axial, coronal, and sagittal views of the T1-w, T2-w, PD-w, and FLAIR images of current and previous time-points.
That is multiple contrasts, multiple views, and multiple time-points. MV-CNN consists of three phases: Preprocessing the Challenge data, Candidate
Extraction, and CNN Prediction. The Challenge data is preprocessed by clamping the top and bottom 1% and the intensity values are scaled to the
range [0, 1].
The Candidate Extraction phase disqualifies the majority of the image voxels from being lesions, thus dramatically improving the performance of
CNN prediction. MV-CNN bases its candidate extraction on two clinical rules (Mechrez et al., 2016):
1. Lesions appear as hyperintense in FLAIR images and can be roughly approximated by thresholding the FLAIR image;
2. Lesions tend to be found in WM or the boundary between WM and GM. Thus a probabilistic WM template (Mazziotta et al., 2001) is registered to
the FLAIR image using a mutual information cost function. Due to misregistration errors the WM template is gray-scale dilated by a radius R.
⎧1 (I
Mask(x ) = ⎨ FLAIR (x ) ≥ TFLAIR )⋂((PWM ⊕ )R)(x ) ≥ TWM ),
⎩ 0 otherwise.
Where IFLAIR(x ) is the FLAIR intensity at x, PWM(x ) is the WM probability which is dilated by )R , a ball of radius R, and the thresholds are TFLAIR and
TWM . The parameters TFLAIR , TWM , and R are determined by cross-validation.
The CNN Prediction phase assigns a lesion probability to each voxel in the image Mask(x ). The input to the CNN are 24 patches of 32×32 pixels
from all four images, three orthogonal views, and two consecutive time-points. All input patches of a single view and time-point are processed by
Table C1
Final rankings from the Challenge participants (B.1). The metrics (Dice, PPV, TPR, & LTPR) have been normalized relative to the inter-rater metrics and denoted with the prefix “N-”.
The Final Score is weighted in the following manner: 20% 1 – LFPR, 20% N-LTPR, 20% LongCorr, 20% TotalCorr, and 20% for the average of N-Dice, N-PPV, & N-TPR.
Name N-Dice N-PPV N-TPR 1–LFPR N-LTPR LongCorr TotalCorr Final Score Ranking
96
A. Carass et al. NeuroImage 148 (2017) 77–102
three convolution layers with the following parameters 24 at 4 × 5 × 5, 32 at 24 × 3 × 3, and 48 at 32 × 3 × 3. The first two convolution layers are
followed by a 2×2 max pooling layer. Thus for each time-point a 48 × 4 × 4 tensor representation is obtained. The tensors of two consecutive time-
points from a single view are concatenated and processed by a 48 at 96 × 1 × 1 convolution layer and a fully connected layer whose output is a vector
of 16 neurons, which is the full representation of a single view. Vectors from axial, coronal, and sagittal views are concatenated and processed by two
fully connected layers of 16 and 2 output neurons respectively. Softmax is applied to the output of the last fully connected layer to obtain a non-
lesion and lesion probability, while the rest of the convolution and fully connected layers are followed by Leaky ReLU activation (α=0.3) and
Dropout layers (p=0.25). Voxels are assigned the lesion label if their lesion probability is higher than a threshold TCNN .
The CNN's weights were optimized for 500 epochs by AdaDelta (Zeiler, 2012) to minimize the categorical cross-entropy. Each training batch
consisted of 64 negative samples and 64 positive samples which were extracted with random rotations in the cardinal planes, drawn from a
Gaussian distribution (μ = 0°, σ = 5°). Values for the thresholds and dilation radius were determine via cross-validation to maximize the mean Dice
score, with TFLAIR = 0.91, TWM = 0.5, TCNN = 0.99, and R=2.
In this section, we present a comparison of the ten Challenge participants, outlined in Section 3.1. Table C1 shows the score achieved by each
participant for a normalized version of each of Dice, PPV, TPR, and LTPR; the normalization is done relative to the inter-rater metrics by dividing by
the inter-rater score, so that the relative value of the metric is boosted. For example, N-Dice is computed as,
minr ∈ 9 (Dice(4 r , 4 A))
N − Dice(4 A) = ,
Dice(4 r1, 4 r2)
where 9 is the set of all raters, and the denominator is the inter-rater score. Also shown in Table C1 are the 1 – LFPR, the Longitudinal Correlation
(LongCorr), and the Total Correlation (TotalCorr). The results in Table C1 and the Challenge are ranked based on a weighted score (20% 1 – LFPR,
20% N-LTPR, 20% LongCorr, 20% TotalCorr, and 20% for the average of N-Dice, N-PPV, & N-TPR). Figs. 2 and 3 show the result generated by
each team on the same subject, as well as showing the preprocessed data, manual delineations generated by the two raters, and our Consensus
Delineation.
Table C2
The participants were timed on how long it took them to return their results for Test Set B, which are listed in the column denoted “Total Time”. The rank sum combination of Total Time
and their final place in the Challenge (see Table C1) was used to rank the efficiency of the teams.
3:18:09 2 3 1
5:59:17 6 1 2
4:03:13 3 4 3
3:02:30 1 7 4
5:05:18 4 5 5
29:08:04 8 2 6
5:53:51 5 6 7
24:47:55 7 8 8
31:56:04 9 9 9
52:38:47 10 10 10
97
A. Carass et al. NeuroImage 148 (2017) 77–102
Table C3
Rankings from the Challenge Website for the Challenge participants (B.1) and the other state-of-the-art methods (B.2).
2 90.698
3 90.283
– 90.070
6 89.807
1 89.159
5 88.744
– 88.465
4 88.009
– 87.917
– 87.376
8 87.017
7 86.916
9 86.436
10 86.068
– 84.140
a
Team TIG submitted new results after the completion of the Challenge to address a bug in their code, the second submitted results are denoted TIG BF.
The participants were told prior to downloading Test Set B, that they would be timed on how long it took them to return the results for that data
set. The results for the time taken for each of the ten Challenge participants are listed in Table C2. The various run times provide a frame of
reference for each of the methods and may serve as a guide for which method is most appropriate for a given situation. For example, consider the
convolutional neural network based approach proposed by Team DIAG which takes an order of magnitude less time and has similar Dice scores to
Team PVG One, and thus Team DIAG might be preferred over Team PVG One. Alternatively, researchers may have a minimum acceptable score in
98
A. Carass et al. NeuroImage 148 (2017) 77–102
another metric, the reported times allow them to identify the quickest method with the required performance level. The ranking for the efficiency
performance was based on a combination of the return time of Test Set B and the final ranking of the teams in the Challenge (see Table C1). The
ranking of both were summed and the team with the lowest combined sum was deemed the most efficient. This allowed us to have a balance
between speed and the accuracy of the method relative to both human raters.
To facilitate the dissemination of the data and promote the sharing of results we have created a website (see Footnote 2). Visitors to the site can
see a list of the Top 25 submitted results. Currently only fifteen results are listed: ten from the Challenge, plus a bug fixed version of a Challenge
participant, and an additional four results—which are outlined in Section 3 and described in detail in Appendix B. Groups interested in running
their methods on the data need only register for an account, download the data, and upload their results. The uploader of the results will receive an
e-mail within ten minutes detailing the results on a per subject and per time-point basis. The report includes the following computed metrics: Dice,
Jaccard, PPV, TPR, LFPR, LTPR, AVD, SSD, algorithm and manual lesion volume. For algorithm A, the Website score is computed as follows,
where : is the set of all subjects, 9 is the set of all raters, and Corr is the Pearson's correlation coefficient of the volumes. This is then linearly
normalized by the inter-rater scores between each other such that the lower inter-rater score has an overall rating of 90. This was designed to mimic
the scoring of the 2008 MICCAI MS Lesion challenge (Styner et al., 2008). Table C3 shows the ranking as displayed on the Challenge Website.
From the main Challenge results (see Table C1) it is clear that there is very little separating the performance of the top three teams. An
interesting characteristic of these three algorithms is that they are machine learning based. However Team DIAG, which finished far off from the
winning Team IIT Madras, used at its core a convolutional neural network engine. This suggests that a more refined approach to using machine
learning technologies is needed to maximize their effectiveness. This point can also be inferred from the performance of MORF on the reduced data
set made available through the Challenge Website (see Table C3). We had expected MORF to perform better than Lesion-TOADS as it should have
represented an improvement on the work of Geremia et al. (2011) which ranked first at the 2008 MICCAI Grand Challenge on MS Lesion
Segmentation (Styner et al., 2008), whereas Lesion-TOADS was ranked third at the same challenge. The disappointing performance of MORF could
be due in part to the differences in the training data and choices about how much and which portion of the available data was used to train the
method. However, it may simply reflect a basic instability in machine learning based approaches.
References convolutional encoder networks with shortcuts for multiscale feature integration
applied to multiple sclerosis lesion segmentation. IEEE Trans. Med. Imag. 35,
1229–1239.
Aït-Ali, L., Prima, S., Heiler, P., Carsin, B., Edan, G., Barillot, C., 2005. STREM: a robust Brosch, T., Yoo, Y., Tang, L.Y.W., Li, D.K.B., Traboulsee, A., Tam, R., 2015. Deep
multidimensional parametric method to segment MS lesions in MRI. In: Proceedings convolutional encoder networks for multiple sclerosis lesion segmentation. In:
of the 8th International Conference on Medical Image Computing and Computer Proceedings of the 18th International Conference on Medical Image Computing and
Assisted Intervention (MICCAI 2005). Springer Berlin Heidelberg. pp. 409–416. Computer Assisted Intervention (MICCAI 2015). Springer Berlin Heidelberg. pp. 3–
Anbeek, P., Vincken, K.L., van Osch, M.J.P., Bisschops, R.H.C., van der Grond, J., 2004. 11.
Probabilistic segmentation of white matter lesions in MR imaging. NeuroImage 21, Buonanno, F.S., Kistler, J.P., Lehrich, J.R., Noseworthy, J.H., New, P.F., Brady, T.J.,
1037–1044. 1983. 1H Nuclear magnetic resonance imaging in multiple sclerosis. Neurol. Clin. 1,
Avants, B.B., Epstein, C.L., Grossman, M., Gee, J.C., 2008. Symmetric diffeomorphic 757–764.
image registration with cross-correlation: evaluating automated labeling of elderly Carass, A., Cuzzocreo, J., Wheeler, M.B., Bazin, P.L., Resnick, S.M., Prince, J.L., 2010.
and neurodegenerative brain. Med. Image Anal. 12, 26–41. Simple paradigm for extra-cerebral tissue removal: algorithm and analysis.
Avants, B.B., Tustison, N.J., Wu, J., Cook, P.A., Gee, J.C., 2011. An open source NeuroImage 56, 1982–1992.
multivariate framework for n-tissue segmentation with evaluation on public data. Carass, A., Wheeler, M.B., Cuzzocreo, J., Bazin, P.L., Bassett, S.S., Prince, J.L., 2007. A
Neuroinformatics 9, 381–400. joint registration and segmentation approach to skull stripping. In: Proceedings of
Bakshi, R., 2005. Magnetic resonance imaging advances in multiple sclerosis. J. the 4th International Symposium on Biomedical Imaging (ISBI 2007), IEEE. pp.
Neuroimaging 15, 5–9. 656–659.
Barnes, C., Shechtman, E., Golman, D.B., Finkelstein, A., 2010. The generalized Cocosco, C.A., Kollokian, V., Kwan, R.K.S., Evans, A.C., 1997. BrainWeb: Online
patchmatch correspondence algorithm. In: 2010 European Conference on Computer interface to a 3D MRI simulated brain database. In: Proceedings of the 3rd
Vision (ECCV 2010). Springer Berlin Heidelberg. pp. 29–43. International Conference on Functional Mapping of the Human Brain, p. S425.
Battaglini, M., Rossi, F., Grove, R.A., Stromillo, M.L., Whitcher, B., Matthews, P.M., De Collins, D.L., Montagnat, J., Zijdenbos, A.P., Evans, A.C., Arnold, D.L., 2001. Automated
Stefano, N., 2014. Automated identification of brain new lesions in multiple sclerosis estimation of brain volume in multiple sclerosis with BICCR. In: 17th Inf.
using subtraction images. Mag. Reson. Im. 39, 1543–1549. Proceedings in Med. Imaging (IPMI 2001), Springer Berlin Heidelberg. pp. 141–
Bazin, P.L., Pham, D.L., 2008. Homeomorphic brain image segmentation with 147.
topological and statistical atlases. Med. Image Anal. 12, 616–625. Collins, D.L., Zijdenbos, A.P., Kollokian, V., Sled, J.G., Kabani, N.J., Holmes, C.J., Evans,
Bazin, P.L., Pham, D.L., Gandler, W., McAuliffe, M., 2005. Free software tools for atlas- A.C., 1998. Design and construction of a realistic digital brain phantom. IEEE Trans.
based volumetric neuroimage analysis. In: Proceedings of the SPIE Medical Imaging Med. Imag. 17, 463–468.
(SPIE-MI 2005), San Diego, CA, February 19–21, 2005, pp. 1824–1833. Compston, A., Coles, A., 2008. Multiple sclerosis. Lancet 372, 1502–1517.
Bezdek, J.C., 1980. A convergence theorem for the fuzzy ISO-DATA clustering Confavreux, C., Vukusic, S., 2008. The clinical epidemiology of multiple sclerosis.
algorithms. IEEE Trans. Pattern. Anal. Mach. Intell. 20, 1–8. Neuroimaging Clin. N. Am. 18, 589–622.
Bosc, M., Heitz, F., Armspach, J.P., Namer, I., Gounot, D., Rumbach, L., 2003. Automatic Coupé, P., Yger, P., Prima, S., Hellier, P., Kervrann, C., Barillot, C., 2008. An optimized
change detection in multimodal serial MRI: application to multiple sclerosis lesion blockwise nonlocal means denoising filter for 3-D magnetic resonance images. IEEE
evolution. NeuroImage 20, 643–656. Trans. Med. Imag. 27, 425–441.
Breiman, L., 2001. Random forests. Mach. Learn. 45, 5–32. Deshpande, H., Maurel, P., Barillot, C., 2015. Adaptive dictionary learning for
Bricq, S., Collet, C., Armspach, J.P., 2008. Lesion detection in 3D brain MRI using competitive classification of multiple sclerosis lesions. In: Proceedings of the 12th
trimmed likelihood estimator and probabilistic atlas. In: Proceedings of the 5th International Symposium on Biomedical Imaging (ISBI 2015), pp. 136–139.
International Symposium on Biomedical Imaging (ISBI 2008), pp. 93–96. Dice, L.R., 1945. Measures of the amount of ecologic association between species.
Brosch, T., Tang, L.Y.W., Yoo, Y., Li, D.K.B., Traboulsee, A., Tam, R., 2016. Deep 3D Ecology 26, 297–302.
99
A. Carass et al. NeuroImage 148 (2017) 77–102
Dugas-Phocion, G., Gonzalez, M.A., Lebrun, C., Chanalet, S., Bensa, C., Malandain, G., Berlin Heidelberg. pp. 469–477.
Ayache, N., 2004. Hierarchical segmentation of multiple sclerosis lesions in multi- He, J., Grossman, R.I., Ge, Y., Mannon, L.J., 2001. Enhancing patterns in multiple
sequence MRI. In: Proceedings of the 2nd International Symposium on Biomedical sclerosis: evolution and persistence. Am. J. Neuroradiol. 22, 664–669.
Imaging (ISBI 2004), pp. 157–160. Heimann, T., van Ginneken, B., Styner, M.A., Arzhaeva, Y., Aurich, V., Bauer, C., Beck,
Ekin, A., 2006. Feature-based brain mid-sagittal plane detection by RANSAC. In: A., Becker, C., Beichel, R., Bekes, G., Bello, F., Binnig, G., Bischof, H., Bornik, A.,
Proceedings of the 14th European Signal Processing Conference, pp. 1–4. Cashman, P., Ying, C., Cordova, A., Dawant, B.M., Fidrich, M., Furst, J.D., Furukawa,
Elliott, C., Arnold, D.L., Collins, D.L., Arbel, T., 2013. Temporally consistent probabilistic D., Grenacher, L., Hornegger, J., Kainmuller, D., Kitney, R.I., Kobatake, H.,
detection of new multiple sclerosis lesions in brain MRI. IEEE Trans. Med. Imag. 32, Lamecker, H., Lange, T., Lee, J., Lennon, B., Li, R., Li, S., Meinzer, H.P., Nemeth, G.,
1490–1503. Raicu, D.S., Rau, A.M., van Rikxoort, E.M., Rousson, M., Rusko, L., Saddi, K.A.,
Elliott, C., Arnold, D.L., Collins, D.L., Arbel, T., 2014. A generative model for automatic Schmidt, G., Seghers, D., Shimizu, A., Slagmolen, P., Sorantin, E., Soza, G.,
detection of resolving multiple sclerosis lesions. In: Proceedings of the 17th Susomboon, R., Waite, J.M., Wimmer, A., Wolf, I., 2009. Comparison and evaluation
International Conference on Medical Image Computing and Computer Assisted of methods for liver segmentation from CT datasets. IEEE Trans. Med. Imag. 28,
Intervention (MICCAI 2014), Springer Berlin Heidelberg. pp. 118–129. 1251–1265.
Elliott, C., Francois, S., Arnold, D.L., Collins, D.L., Arbel, T., 2010. Bayesian classification Heinrich, M.P., Jenkinson, M., Bhushan, M., Matin, T., Gleeson, F.V., Brady, M.,
of multiple sclerosis lesions in longitudinal MRI using subtraction images. In: Schnabel, J.A., 2012. MIND: Modality independent neighbourhood descriptor for
Proceedings of the 13th International Conference on Medical Image Computing and multi-modal deformable registration. Med. Image Anal. 16, 1432–1435.
Computer Assisted Intervention (MICCAI 2010), Springer Berlin Heidelberg. pp. Heinrich, M.P., Jenkinson, M., Papież, B.W., Brady, M., Schnabel, J.A., 2013. Towards
290–297. realtime multimodal fusion for image-guided interventions using self-similarities. In:
Evans, A.C., Frank, J.A., Antel, J., Miller, D.H., 1997. The role of MRI in clinical trials of Proceedings of the 16th International Conference on Medical Image Computing and
multiple sclerosis: comparison of image processing techniques. Ann. Neurol. 41, Computer Assisted Intervention (MICCAI 2013), Springer Berlin Heidelberg. pp.
125–132. 187–194.
Ferrari, R.J., Wei, X., Zhang, Y., Scott, J.N., Mitchell, J.R., 2003. Segmentation of Hsu, Y., Hagel, N., Rekkers, G., 1984. New likelihood test methods for change detection
multiple sclerosis lesions using support vector machines. In: Proceedings of SPIE in image sequences. Comput. Vision., Graph., Image Process. 26, 73–106.
Medical Imaging (SPIE-MI 2003), pp. 16–26. Iheme, L.O., Unay, D., Baskaya, O., Sennaz, A., Kandemir, M., Yalciner, Z.B., Tepe, M.S.,
Filippi, M., Grossman, R.I., 2002. MRI techniques to monitor MS evolution: the present Kahraman, T., Unal, G., 2013. Concordance between computer-based neuroimaging
and the future. Neurology 58, 1147–1153. findings and expert assessments in dementia grading. In: Signal Processing and
Filippi, M., Horsfield, M.A., Tofts, P.S., Barkhof, F., Thompson, A.J., Miller, D.H., 1995. Communications Applications Conference (SIU), pp. 1–4.
Quantitative assessment of MRI lesion load in monitoring the evolution of multiple Jain, S., Sima, D.M., Ribbens, A., Cambron, M., Maertens, A., Van Hecke, W., De Mey, J.,
sclerosis. Brain 118, 1601–1612. Barkhof, F., Steenwijk, M.D., Daams, M., Maes, F., Van Huffel, S., Vrenken, H.,
Filippi, M., Preziosa, P., Rocca, M.A., 2014. Magnetic resonance outcome measures in Smeets, D., 2015. Automatic segmentation and volumetry of multiple sclerosis brain
multiple sclerosis trials: time to rethink? Curr. Opin. Neurol. 27, 290–299. lesions from MR images. NeuroImage: Clin. 8, 367–375.
Filippi, M., Rocca, M.A., Barkhof, F., Brück, W., Chen, J.T., Comi, G., DeLuca, G., De Jog, A., Carass, A., Pham, D.L., Prince, J.L., 2015. Multi-output decision trees for lesion
Stefano, N., Erickson, B.J., Evangelou, N., Fazekas, F., Geurts, J.J.G., Lucchinetti, C., segmentation in multiple sclerosis. In: Proceedings of SPIE Medical Imaging (SPIE-
Miller, D.H., Pelletier, D., Popescu, B.F.G., Lassmann, H., 2012. Association between MI 2015), Orlando, FL, February 21–26, 2015, pp. 94131C–94131C–6.
pathological and MRI findings in multiple sclerosis for the Attendees of the Jog, A., Carass, A., Roy, S., Pham, D.L., Prince, J.L., 2017. Random forest regression for
Correlation between Pathological MRI findings in MS workshop. Lancet Neurol. 11, magnetic resonance image synthesis. Med. Image Anal. 35, 475–488.
349–360. Johnston, B., Atkins, M.S., Mackiewich, B., Anderson, M., 1996. Segmentation of
Freifeld, O., Greenspan, H., Goldberger, J., 2009. Multiple sclerosis lesion detection multiple sclerosis lesions in intensity corrected multispectral MRI. IEEE Trans. Med.
using constrained GMM and curve evolution. J. Biomed. Imaging 2009, 14:1–14:13. Imag. 15, 154–169.
Gaitán, M., Shea, C.D., Evangelou, I.E., Stone, R.D., Fenton, K.M., Bielekova, B., Jonkman, L.E., Lopez Soriano, A., Amor, S., Barkhof, F., van der Valk, P., Vrenken, H.,
Massacesi, L., Reich, D.S., 2011. Evolution of the blood-brain barrier in newly Geurts, J.J.G., 2015. Can MS lesion stages be distinguished with MRI? A
forming multiple sclerosis lesions. Ann. Neurol. 70, 22–29. postmortem MRI and histopathology study. J. Neurol. 262, 11074–11080.
Ganiler, O., Oliver, A., Diez, Y., Freixenet, J., Vilanova, J.C., Beltran, B., Ramió-Torrentà, Kamber, M., Shinghal, R., Collins, D.L., Francis, G.S., Evans, A.C., 1996. Model-based 3-
L., Rovira, A., Lladó, X., 2014. A subtraction pipeline for automatic detection of new D segmentation of multiple sclerosis lesions in magnetic resonance brain images.
appearing multiple sclerosis lesions in longitudinal studies. Neuroradiology 56, IEEE Trans. Med. Imag. 14, 442–453.
363–374. Karpate, Y., Commowick, O., Barillot, C., Edan, G., 2014. Longitudinal intensity
García-Lorenzo, D., Francis, S., Narayanan, S., Arnold, D.L., Collins, D.L., 2013. Review normalization in multiple sclerosis patients. Transl. Res. Med. Imaging 8680,
of automatic segmentation methods of multiple sclerosis white matter lesions on 118–125.
conventional magnetic resonance imaging. Med. Image Anal. 17, 1–18. Khayati, R., Vafadust, M., Towhidkhah, F., Nabavi, M., 2008. Fully automatic
García-Lorenzo, D., Lecoeur, J., Arnold, D.L., Collins, D.L., Barillot, C., 2009. Multiple segmentation of multiple sclerosis lesions in brain MR FLAIR images using adaptive
sclerosis lesion segmentation using an automated multimodal graph cuts. In: mixtures method and markov random field model. Comput. Biol. Med. 38, 379–390.
Proceedings of the 12th International Conference on Medical Image Computing and Kikinis, R., Guttmann, C.R.G., Metcalf, D., Wells, W.M., III, Ettinger, G.J., Weiner, H.L.,
Computer Assisted Intervention (MICCAI 2009), Springer Berlin Heidelberg. pp. Jolesz, F.A., 1999. Quantitative follow-up of patients with multiple sclerosis using
584–591. MRI: technical aspects. Jrnl. Magn. Reson. Imaging 9, 519–530.
García-Lorenzo, D., Prima, S., Arnold, D.L., Collins, D.L., Barillot, C., 2011. Trimmed- Kwan, R.K.S., Evans, A.C., Pike, G.B., 1999. MRI simulation-based evaluation of image-
likelihood estimation for focal lesions and tissue segmentation in multisequence MRI processing and classification methods. IEEE Trans. Med. Imag. 18, 1085–1097.
for multiple sclerosis. IEEE Trans. Med. Imag. 30, 1455–1467. Li, H., Zhao, R., Wang, X., 2014. Highly Efficient Forward and Backward Propagation of
García-Lorenzo, D., Prima, S., Collins, D.L., Arnold, D.L., Morrissey, S.P., Barillot, C., Convolutional Neural Networks for Pixelwise Classification. CoRR arXiv:1412.4526.
2008. Combining robust expectation maximization and mean shift algorithms for Lladó, X., Oliver, A., Cabezas, M., Freixenet, J., Vilanova, J.C., Quiles, A., Valls, L.,
multiple sclerosis brain segmentation. In: Proceedings of the 11th International Ramió-Torrentà, L., Rovira, À., 2012. Segmentation of multiple sclerosis lesions in
Conference on Medical Image Computing and Computer Assisted Intervention brain MRI: a review of automated approaches. Inf. Sci. 186, 164–185.
(MICCAI 2008) workshop on Medical Image Analysis on Multiple Sclerosis (MIAMS Lucas, B.C., Bogovic, J.A., Carass, A., Bazin, P.L., Prince, J.L., Pham, D.L., Landman,
2008), pp. 82–91. B.A., 2010. The Java Image Science Toolkit (JIST) for rapid prototyping and
Geremia, E., Clatz, O., Menze, B.H., Konukoglu, E., Criminisi, A., Ayache, N., 2011. publishing of neuroimaging software. Neuroinformatics 8, 5–17.
Spatial decision forests for MS lesion segmentation in multi-channel magnetic Maier, O., Menze, B.H., von der Gablentz, J., Häni, L., Heinrich, M.P., Liebrand, M.,
resonance images. NeuroImage 57, 378–390. Winzeck, S., Basit, A., Bentley, P., Chen, L., Christiaens, D., Dutil, F., Egger, K., Feng,
Geremia, E., Menze, B.H., Clatz, O., Konukoglu, E., Criminisi, A., Ayache, N., 2010. C., Glocker, B., Götz, M., Haeck, T., Halme, H.L., Havaei, M., Iftekharuddin, K.M.,
Spatial decision forests for MS lesion segmentation in multi-channel MR images. In: Jodoin, P.M., Kamnitsas, K., Kellner, E., Korvenoja, A., Larochelle, H., Ledig, C., Lee,
Proceedings of the 13th International Conference on Medical Image Computing and J.H., Maes, F., Mahmood, Q., Maier-Hein, K.H., McKinley, R., Muschelli, J., Pal, C.,
Computer Assisted Intervention (MICCAI 2010), Springer Berlin Heidelberg. pp. Pei, L., Rangarajan, J.R., Reza, S.M.S., Robben, D., Rueckert, D., Salli, E., Suetens,
111–118. P., Wang, C.W., Wilms, M., Kirschke, J.S., Krämer, U.M., Münte, T.F., Schramm, P.,
Geurts, P., Ernst, D., Wehenkel, L., 2006. Extremely randomized trees. Mach. Learn. 36, Wiest, R., Handels, H., Reyes, M., 2017. ISLES 2015 - A public evaluation
3–42. benchmark for ischemic stroke lesion segmentation from multispectral MRI. Med.
Geurts, P., Touleimat, N., Dutreix, M., d'Alche Buc, F., 2007. Inferring biological Image Anal. 35, 250–269.
networks with output kernel trees. BMC Bioinforma. 8, S4. Maier, O., Wilms, M., von der Gablentz, J., Krämer, U.M., Münte, T.F., Handels, H.,
Global Burden of Disease Study 2013 Mortality and Causes of Death Collaborators, 2015. 2015. Extra Tree forests for sub-acute ischemic stroke lesion segmentation in MR
Global, regional, and national age-sex specific all-cause and cause-specific mortality sequences. J. Neurosci. Methods 240, 89–100.
for 240 causes of death, 1990–2013: a systematic analysis for the Global Burden of Mairal, J., Bach, F., Ponce, J., Sapiro, G., Zisserman, A., 2009. Supervised dictionary
Disease Study 2013. Lancet 385, 117–171. learning. In: Advances in Neural Information Processing Systems (NIPS) 21, Curran
Harmouche, R., Collins, D.L., Arnold, D.L., Francis, S., Arbel, T., 2006. Bayesian MS Associates, Inc., pp. 1033–1040.
lesion classification modeling regional and local spatial information. In: Proceedings Mazziotta, J., Toga, A., Evans, A., Fox, P., Lancaster, J., Zilles, K., Woods, R., Paus, T.,
of the 18th International Conference on Pattern Recognition (ICPR), 2006, pp. 984– Simpson, G., Pike, B., Holmes, C., Collins, L., Thompson, P., MacDonald, D.,
987. Iacoboni, M., Schormann, T., Amunts, K., Palomero-Gallagher, N., Geyer, S.,
Havaei, M., Guizard, N., Chapados, N., Bengio, Y., 2016. HeMIS: Hetero-modal image Parsons, L., Narr, K., Kabani, N., Goualher, G.L., Boomsma, D., Cannon, T.,
segmentation. In: Proceedings of the 19th International Conference on Medical Kawashima, R., Mazoyer, B., 2001. A probabilistic atlas and reference system for the
Image Computing and Computer Assisted Intervention (MICCAI 2016), Springer human brain: International Consortium for Brain Mapping (ICBM). Philos. Trans. R.
100
A. Carass et al. NeuroImage 148 (2017) 77–102
101
A. Carass et al. NeuroImage 148 (2017) 77–102
Estimation (STAPLE): an algorithm for the validation of image segmentation. IEEE Barkhof, F., Guttmanna, C.R.G., 2006. Automated segmentation of multiple sclerosis
Trans. Med. Imag. 23, 903–921. lesion subtypes with multichannel MRI. NeuroImage 32, 1205–1215.
Weiss, N., Rueckert, D., Rao, A., 2013. Multiple sclerosis lesion segmentation using Xie, Y., Tao, X., 2011. White matter lesion segmentation using machine learning and
dictionary learning and sparse coding. In: Proceedings of the 16th International weakly labeled MR images. In: Proceedings of SPIE Medical Imaging (SPIE-MI
Conference on Medical Image Computing and Computer Assisted Intervention 2011), Orlando, FL, February 12–17, 2011, pp. 79622G–79622G–9.
(MICCAI 2013), Springer Berlin Heidelberg. pp. 735–742. Xue, Z., Shen, D., Davatzikos, C., 2006. CLASSIC: consistent longitudinal alignment and
Welti, D., Gerig, G., Radü, E.W., Kappos, L., Székely, G., 2001. Spatio-temporal segmentation for serial image computing. NeuroImage 30, 388–399.
segmentation of active multiple scleroris lesions in serial MRI data. In: 17th Inf. Zeiler, M.D., 2012. ADADELTA: an adaptive learning rate method arXiv:1212.5701.
Proceedings in Med. Imaging (IPMI 2001), Springer Berlin Heidelberg. pp. 438– Zhang, Y., Brady, M., Smith, S., 2001. Segmentation of brain MR images through a
445. hidden Markov random field model and the expectation-maximization algorithm.
Wilcoxon, F., 1945. Individual comparisons by ranking methods. Biom. Bull. 1, 80–83. IEEE Trans. Med. Imag. 20, 45–57.
World Health Organization, 2008. Atlas: Multiple Sclerosis Resources in the World 2008. Zijdenbos, A.P., Dawant, B.M., Margolin, R.A., Palmer, A.C., 1994. Morphometric
Springer, Geneva, Switzerland. analysis of white matter lesions in MR images: method and validation. IEEE Trans.
Wu, Y., Warfield, S.K., Tan, I.L., Wells, W.M., III, Meier, D.S., van Schijndel, R.A., Med. Imag. 13, 716–724.
102