more reasonable results with respect to random guess-
ing. Moreover, none of the 15 ensembles was ranked
in the first position across different datasets. The E0F
ensembles were more reasonable than the other en-
sembles in four datasets; the ENF ensembles were
the best in two datasets (COCOMO81 and Miyazaki).
However, the E1F ensembles were the less reasonable
in all datasets.
However, the accuracy results in terms of 8 perfor-
mance measures suggest that the ENF, in particular
with the combiners IR or ME, outperformed the E1F
and E0F ensembles in 5 out of 6 datasets. This im-
plies that using different feature subsets by ensemble
members can lead to more accurate estimations than
when members use the same feature subset or all the
available features. In fact, the success of the ENF en-
sembles is mainly due to the fact that their members
were diverse and generate different estimations at the
same point (i.e. diversity) than the members of E1F or
E0F. Moreover, E0F ensembles generate slightly bet-
ter estimates than E1F. Therefore, we conclude that
ensembles without feature selection were better and
easier to construct than ensembles with one filter.
Ongoing work will focus on investigating the im-
pact of other feature selection techniques, including
filters or wrappers, on the accuracy of homogenous
and heterogeneous ensembles.
REFERENCES
Azzeh, M., Nassif, A. B., and Minku, L. L. (2015). An
empirical evaluation of ensemble adjustment methods
for analogy-based effort estimation. The Journal of
Systems and Software, 103:36–52.
Cohen, J. (1992). A power primer. Psychological Bulletin,
112(1):155–159.
Hira, Z. M. and Gillies, D. F. (2015). A review of feature se-
lection and feature extraction methods applied on mi-
croarray data. Advances in Bioinformatics, 2015(1).
Hosni, M. and Idri, A. (2017). Software effort estimation
using classical analogy ensembles based on random
subspace. In Proceedings of the ACM Symposium on
Applied Computing, volume Part F1280.
Hosni, M., Idri, A., and Abran, A. (2017a). Investigating
heterogeneous ensembles with filter feature selection
for software effort estimation. In ACM International
Conference Proceeding Series, volume Part F1319.
Hosni, M., Idri, A., Abran, A., and Nassif, A. B. (2017b).
On the value of parameter tuning in heterogeneous en-
sembles effort estimation.
Hosni, M., Idri, A., Nassif, A., and Abran, A. (2017c). Het-
erogeneous Ensembles for Software Development Ef-
fort Estimation. In Proceedings - 2016 3rd Interna-
tional Conference on Soft Computing and Machine In-
telligence, ISCMI 2016.
Idri, A., Abnane, I., and Abran, A. (2017). Evaluating
Pred( p ) and standardized accuracy criteria in soft-
ware development effort estimation. Journal of Soft-
ware: Evolution and Process, (September):e1925.
Idri, A. and Cherradi, S. (2016). Improving Effort Esti-
mation of Fuzzy Analogy using Feature Subset Se-
lection. In Computational Intelligence (SSCI), 2016
IEEE Symposium Series on.
Idri, A., Hosni, M., and Abran, A. (2016a). Improved Esti-
mation of Software Development Effort Using Classi-
cal and Fuzzy Analogy Ensembles. Applied Soft Com-
puting.
Idri, A., Hosni, M., and Abran, A. (2016b). Systematic
Mapping Study of Ensemble Effort Estimation. In
Proceedings of the 11th International Conference on
Evaluation of Novel Software Approaches to Software
Engineering, number Enase, pages 132–139.
Idri, A., Hosni, M., and Alain, A. (2016c). Systematic Liter-
ature Review of Ensemble Effort Estimation. Journal
of Systems and Software, 118:151–175.
Jovic, A., Brkic, K., and Bogunovic, N. (2015). A re-
view of feature selection methods with applications.
In 2015 38th International Convention on Information
and Communication Technology, Electronics and Mi-
croelectronics (MIPRO), number May, pages 1200–
1205.
Khoshgoftaar, T., Golawala, M., and Hulse, J. V. (2007). An
Empirical Study of Learning from Imbalanced Data
Using Random Forest. In 19th IEEE International
Conference on Tools with Artificial Intelligence(ICTAI
2007), volume 2, pages 310–317.
Kocaguneli, E., Kultur, Y., and Bener, A. (2009). Combin-
ing Multiple Learners Induced on Multiple Datasets
for Software Effort Prediction. In Proceedings of In-
ternational Symposium on Software Reliability Engi-
neering.
Minku, L. L. and Yao, X. (2013). Ensembles and locality:
Insight on improving software effort estimation. Infor-
mation and Software Technology, 55(8):1512–1528.
Scott, A. J. and Knott, M. (1974). A Cluster Analysis
Method for Grouping Means in the Analysis of Vari-
ance. Biometrics, 30(3):507–512.
Shepperd, M. and MacDonell, S. (2012). Evaluating pre-
diction systems in software project estimation. Infor-
mation and Software Technology, 54(8):820–827.
Song, L., Minku, L. L., and Yao, X. (2013). The impact of
parameter tuning on software effort estimation using
learning machines. In Proceedings of the 9th Interna-
tional Conference on Predictive Models in Software
Engineering.
Zhou, Z.-H. (2012). Ensemble Methods. CRC Press.
Zhu, D. (2010). A hybrid approach for efficient ensembles.
Decision Support Systems, 48(3):480–487.
ICSOFT 2018 - 13th International Conference on Software Technologies
412