
Table 1: Results for recall, precision and F1, for different N values. Recall values for random guess (RND), as well as recall
and precision for default guess are also shown
Recall Prec.
1 0,147 0,287 0,194 0,089 0,277 0,134 0,091 0,425 0,150 0,054 0,360 0,093 0,003 0,011 0,019
2 0,194 0,288 0,201 0,106 0,213 0,142 0,100 0,352 0,155 0,059 0,299 0,098 0,007 0,017 0,015
3 0,217 0,168 0,189 0,115 0,189 0,143 0,103 0,318 0,155 0,061 0,269 0,099 0,010 0,029 0,013
5 0,240 0,129 0,168 0,121 0,163 0,139 0,105 0,282 0,153 0,063 0,241 0,100 0,017 0,034 0,012
10 0,261 0,095 0,140 0,125 0,138 0,131 0,109 0,252 0,152 0,065 0,218 0,101 0,034 0,057 0,010
20 0,272 0,076 0,119 0,127 0,128 0,128 0,111 0,236 0,151 0,066 0,210 0,101 0,069 0,108 0,009
N
Prec. Prec.Recall RecallPrec.Recall Prec. F1
(Recall)
RND
Default
F1
minsup=0,003
minconf=0,1
minsup=0,005
minconf=0,5
F1
minsup=0,005
minconf=0,1
minsup=0,003
minconf=0,5
F1Recall
been retrieved by the system, i.e., the proportion of
resources in the hidden set that are adequately
recommended. The value of recall tends to increase
with N, the number of recommendations made for a
single team.
||
||
Hidden
RecHidden
Recall
∩
=
Precision gives us the average quality of an
individual recommendation. As N increases, the
quality of each recommendation decreases.
||
||
Rec
RecHidden
Precision
∩
=
F1 has been suggested as a measure that
combines recall and precision with equal weights. It
ranges from 0 to 1 and higher values indicate a more
balanced combination between recall and precision.
It is useful as a summary of the other two measures.
PrecisionRecall
PrecisionRecall
F1
+
××
=
2
The data used in these experiments refer to the
period between September 2001 and November
2002. For this period we have 290 resources and
26234 baskets. The average number of resources per
basket is 2,68. With the train and test split we got
20987 baskets for train set and 5247 baskets for test
set.
To build the set of association rules we tried
different combinations of minimum support and
minimum confidence. Table 1 shows the results for
recall, precision and F1, for different N values. The
best results for recall were achieved with minimum
support = 0,003 and with minimum confidence = 0,1.
For these parameters, the number of rules in the
model was 8957.
Recall is around 15% when only one
recommendation is made (N = 1) – this means that
we are able to retrieve 15% of the relevant
recommendations. In this case, precision is higher
(0,287) because a recommendation is not made
when no rule applies. The recommender model
recall value is 49 times higher than the resource
random guess (Rnd column). These random values
were obtained by dividing N by the total number of
resources (290).
We have also compared the predictive accuracy
of our model with the default recommendations (the
most likely resources a priori). When N = 1, the
default recommendation for every basket in the
observable set is the resource with the highest
support in the training set; when N = 2, the default
recommendations for every basket in the observable
set are the two resources with the highest support in
the train set, and so on. In Figure 5 we can see the
comparison of recall values between our model and
default recommendations, for different N values.
In the case of precision, it drops smoothly as the
number of recommendations N increases, as it was
expected. When N = 1 each one the collaborative
filtering recommendations made has a 28,7% chance
of being relevant. In Figure 6 we can see the
comparison of precision values between our model
and default recommendations, for different N values.
The F1 measure indicates that the best
combination of recall and precision is achieved
when N = 2. This can be used if we want to give the
team manager a list of recommendations with a good
balance between recall and precision.
Minsup=0,003 Minconf=0,1
0
0,05
0,1
0,15
0,2
0,25
0,3
0,35
12351020
N
Recall Precision F1 RND
Figure 4: Results for recall, precision and F1, for
different N values – minimum support=0,003 and
minimum confidence=0,1.
MODEL-BASED COLLABORATIVE FILTERING FOR TEAM BUILDING SUPPORT
245