Nonetheless, the study provides first-time insight
of the use of a stacked ensemble architecture in the
domain of learning analytics and early detection of
academically at-risk students. We recognize that
much more could be written about each of these
topics. However, we will provide more complete
coverage of these topics at the conference.
6 SUMMARY AND CONCLUDING
COMMENTS
Stacked ensembles are powerful and flexible
machine learning frameworks with the potential of
delivering better and more stable predictions. This
paper demonstrates how to create a stacked
ensemble and perform predictions of academically
at-risk students. The impetus of this research stems
from the need of introducing novel approaches that
can be used in practical settings to predict academic
performance and carry out early detection of
students at risk. The methodology presented in this
paper is the subject of intensive research and
exploration at our institution, inclusive of the
analysis of different configurations of classifiers,
model tuning criteria, arrangements of predictors,
and its impact on the stack’s predictive performance.
A pilot on a group of course sections has been run at
the College during Fall 2017 and will continue
through Spring 2018 using the stacked ensemble
methodology described in this paper. The model
output will be visualized through the LMS using a
graphical user interface –a dashboard- augmented
with statistics generated from the prediction.
Hopefully, this paper will provide the motivation for
other researchers and practitioners to begin
exploring the use of stacked ensembles for
predictive modeling in the learning analytics
domain.
REFERENCES
Breiman, L., 2001. Random Forests. Mach Learn 45, 5–
32.
Chen, T., Guestrin, C., 2016. XGBoost: A Scalable Tree
Boosting System. CoRR abs/1603.02754.
Džeroski, S., Ženko, B., 2004. Is Combining Classifiers
with Stacking Better Than Selecting the Best One?
Mach Learn 54, 255–273.
Fan, R.-E., Chang, K.-W., Hsieh, C.-J., Wang, X.-R., Lin,
C.-J., 2008. LIBLINEAR: A Library for Large Linear
Classification. J Mach Learn Res 9, 1871–1874.
Jayaprakash, S. M., Moody, E. W., Lauría, E. J. M.,
Regan, J. R., Baron, J. D., 2014. Early Alert of
Academically At-Risk Students: An Open Source
Analytics Initiative. J. Learn. Anal. 1, 42.
John, G.H., Langley, P., 1995. Estimating Continuous
Distributions in Bayesian Classifiers, in: Proceedings
of the Eleventh Conference on Uncertainty in
Artificial Intelligence, UAI’95. Morgan Kaufmann
Publishers Inc., San Francisco, CA, USA, pp. 338–
345.
Kaggle Team, 2017. No Free Hunch: Allstate Claims
Severity Competition, 2nd Place Winner’s Interview:
Alexey Noskov. Off. Blog Kagglecom.
Kuhn, M., 2008. Building Predictive Models in R Using
the caret Package. J. Stat. Softw. Artic. 28, 1–26.
Lam, L., 2000. Classifier Combinations: Implementations
and Theoretical Issues, in: Multiple Classifier
Systems: First International Workshop, MCS 2000
Cagliari, Italy, June 21–23, 2000 Proceedings.
Springer Berlin Heidelberg, Berlin, Heidelberg, pp.
77–86.
Landwehr, N., Hall, M., Frank, E., 2005. Logistic Model
Trees. Mach. Learn. 59, 161–205.
Lauría, E.J.M., Presutti, E., Sokoloff, M., Guarino, M.,
2016. Crossing the Chasm to Big Data: Early
Detection of at-Risk Students in a Cluster Computing
Environment. Presented at the Proceedings of the 7th
International Learning Analytics & Knowledge
Conference (LAK’17)- Practitioner Track. Vancouver,
Canada.
Pistilli, M.D., Arnold, K.E., 2010. Purdue Signals: Mining
Real-time Academic Data to Enhance Student
Success. Campus 22–24.
Polikar, R., 2006. Ensemble based systems in decision
making. IEEE Circuits Syst. Mag. 6, 21–45.
Romero, C., Ventura, S., Garcia, E., 2008. Data mining in
course management systems: Moodle case study and
tutorial. Comput Educ 51, 368–384.
Sill, J., Takács, G., Mackey, L.W., Lin, D., 2009. Feature-
Weighted Linear Stacking. CoRR abs/0911.0460.
Smith, V.C., Lange, A., Huston, D.R., 2012. Predictive
Modeling to Forecast Student Outcomes and Drive
Effective Interventions in Online Community College
Courses. J. Asynchronous Learn. Netw. 16, 51–61.
van der Laan, M., Polley, E., Hubbard, A., 2007. Super
Learner (Working Paper No. 222), Berkeley Division
of Biostatistics Working Paper Series. University of
California, Berkeley.
Wolpert, D.H., 1992. Stacked generalization. Neural
Netw. 5, 241–259.