Parameter Setting. The method depends on sev-
eral parameters: the radius R and R
max
, the ration of
non-landmark versus landmark voxels P, the number
of voxels N
p
to extract for computing a prediction,
the number of trees T , the size of the window W and
the number of resolutions D. These parameters are
either set to their maximum value given the avail-
able computing resources (T, N
p
) or tuned through
cross-validation. Trees were fully grown (n
min
= 2)
and the K parameter was set to its default value
p
3D((2W + 1)
2
) (Geurts et al., 2006).
2.2 Multimodal Landmark-based Rigid
Registration
Once anatomical landmark coordinates have been
predicted in both images, the registration of the re-
sulting matching pairs of landmark positions is for-
mulated as the least-square optimization problem pre-
sented in (1).
min
X,T
N
∑
i=1
||p
0
i
− (X p
i
+ T )||
2
(1)
N is the number of landmarks, p
i
and p
0
i
are the co-
ordinates of the ith landmark in the two images, X
is a 3 × 3 rotation matrix, and T a 3 × 1 translation
vector. To solve this problem, we use the noniterative
SVD-based algorithm proposed in (Arun et al., 1987).
It is important to notice that, as opposed to volume
registration based on local feature detectors and in-
variant descriptors (e.g. (Lukashevich et al., 2011)),
our method does not require matching of landmark
descriptors accross modalities.
3 EXPERIMENTS AND RESULTS
In this section, we first describe our dataset, divided
into a training and a test set. Then, we study system-
atically the influence of the main parameters of our
landmark detection method by leave-one-patient-out
validation on the training set. Finally, we present reg-
istration results on the test set and compare them to
a semi-automated volume registration algorithm (Fe-
dorov et al., 2012).
3.1 Datasets
Our dataset contains images related to 45 patients
(male and female) and was acquired at the Radiother-
apy and Oncology Department, University of Li
`
ege,
Belgium. For each of these patients, we have one
pelvic CT scan as the reference (45 CTs in total), and
Table 1: Sets of values tested during cross-validation for
each parameter. In bold, the default value of each parameter
used in the first stage of cross-validation.
Parameter Tested values
R 2, 4, 5,6, 7, 8, 10,12, 14, 16
R
max
10, 25, 40,50, 75, 100, 200,500, 1000, 2000
P 0.1, 0.25, 0.5,1, 1.5, 2, 3,4, 6, 8
N
p
1, 10, 100,1000, 5000, 10000, 50000,
100000, 200000, 500000
T 1,5, 10, 25, 50,75, 100, 150, 200, 300
W 2, 3, 4,5, 6, 7, 8,9, 10, 12
D 1,2, 3, 4, 5,6, 7, 8, 9, 10
at least one corresponding CBCT scan of the pelvis
(68 CBCTs in total). We divided this dataset into a
training set of 30 patients, each with one CT and at
least one CBCT (i.e 53 CBCTs in total), and a test
set of 15 patients, each with exactly one CT and one
CBCT.
Because our algorithm works better with volumes
of identical resolutions and the original resolution
information is always available, each CT and each
CBCT were resized to 1 × 1 × 1mm voxel resolution.
Originally, CT scan resolutions were comprised be-
tween 0.5 and 3mm. The CBCT scans were acquired
with an Elekta XVI scanner, that were reconstructed
to 1 × 1 × 1mm resolution. More information about
the quality of the CBCT image acquisition procedure
can be found in (Kamath et al., 2011).
On each CT and each CBCT, 8 landmarks dis-
tributed in the pelvis were manually annotated two
times by the same skilled operator. The mean distance
between the two annotation runs is shown in Table 2
(Manual Err.). The position of each landmark is pre-
sented in Figure 4 for CT scans. We used as ground-
truth for each landmark the mean coordinates of the
two manual annotations provided by the operator.
3.2 Landmark Detection Results
3.2.1 Protocol
For our experiments, we fixed extremely random-
ized tree parameters to recommended values (K =
p
3D((2W + 1)
2
), n
min
= 2) (Geurts et al., 2006).
Other parameter values were evaluated in the ranges
presented in Table 1.
These values were tested for both the regression
and the classification approaches using leave-one-
patient-out in the training set. Since it is not possible
to explore all parameter combinations, we use a two-
stage approach. In the first stage, for each parameter
in turn, all its values were tested with the other pa-
rameters set to some default value (in bold in Table1).
In the second stage, the exact same procedure was ap-
VISAPP 2017 - International Conference on Computer Vision Theory and Applications
336