chine learning approaches and existing expert algo-
rithms. In general, methods based on machine learn-
ing yielded promising results (with the best result ac-
cording to our main evaluation metric obtained by an
LSTM). Still, some caveats need to be expressed:
• machine-learning methods, including LSTMs,
might give good results as measured with an eval-
uation metric, but still behaving in an impracti-
cal manner and breaking natural assumptions (e.g.
probability of retention not decreasing even for
very long intervals or even higher probabilities of
retention for longer intervals),
• this can be alleviated with modifications trans-
planted from expert methods (e.g. forgetting
curve as proposed by Wo´zniak),
• LSTM is susceptible to large variance and, in
practical terms, is more complicated to use than
expert methods,
• the ranking of methods depends heavily on the
evaluation metric chosen, we claim the evaluation
method we called MacroAvgMAE is the most rea-
sonable, but still it is far from obvious how this re-
lates to quality of learning, when a given method
is embedded within a real learning application.
One area of improvement for the method based on
LSTMs is to equip it with a mechanism to adapt for a
specific user/course, just the way expert methods such
as SM-17 do.
REFERENCES
Chen, T. and Guestrin, C. (2016). XGBoost: A scalable
tree boosting system. In Proceedings of the 22nd
ACM SIGKDD International Conference on Knowl-
edge Discovery and Data Mining, KDD ’16, pages
785–794, New York, NY, USA. ACM.
Grali
´
nski, F., Jaworski, R., Borchmann, Ł., and Wierz-
cho
´
n, P. (2016). Gonito.net – open platform for re-
search competition, cooperation and reproducibility.
In Branco, A., Calzolari, N., and Choukri, K., editors,
Proceedings of the 4REAL Workshop: Workshop on
Research Results Reproducibility and Resources Ci-
tation in Science and Technology of Language, pages
13–20.
Grali
´
nski, F., Wróblewska, A., Stanisławek, T., Grabowski,
K., and Górecki, T. (2019). GEval: Tool for debug-
ging NLP datasets and models. In Proceedings of the
2019 ACL Workshop BlackboxNLP: Analyzing and In-
terpreting Neural Networks for NLP, pages 254–262,
Florence, Italy. Association for Computational Lin-
guistics.
Hochreiter, S. and Schmidhuber, J. (1997). Long short-term
memory. Neural computation, 9(8):1735–1780.
Leitner, S. (1999). So lernt man lernen angewandte
Lernpsychologie - ein Weg zum Erfolg. Weltbild Ver-
lag.
Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J.,
Chanan, G., Killeen, T., Lin, Z., Gimelshein, N.,
Antiga, L., Desmaison, A., Kopf, A., Yang, E., De-
Vito, Z., Raison, M., Tejani, A., Chilamkurthy, S.,
Steiner, B., Fang, L., Bai, J., and Chintala, S. (2019).
Pytorch: An imperative style, high-performance deep
learning library. In Wallach, H., Larochelle, H.,
Beygelzimer, A., d'Alché-Buc, F., Fox, E., and Gar-
nett, R., editors, Advances in Neural Information Pro-
cessing Systems 32, pages 8024–8035. Curran Asso-
ciates, Inc.
Settles, B. and Meeder, B. (2016). A trainable spaced repe-
tition model for language learning. In Proceedings of
the 54th annual meeting of the Association for Com-
putational Linguistics (volume 1: Long papers), pages
1848–1858.
Sinha, S. (2019). Using deep reinforcement learning for
personalizing review sessions on e-learning platforms
with spaced repetition.
Upadhyay, U., De, A., and Gomez Rodriguez, M. (2018).
Deep reinforcement learning of marked temporal
point processes. Advances in Neural Information Pro-
cessing Systems, 31.
Wo´zniak, P. A., Gorzela
´
nczyk, E. J., and Murakowski, J. A.
(1995). Two components of long-term memory. Acta
neurobiologiae experimentalis, 55(4):301–305.
Wo´zniak, P. (1990). Optimization of learning. Master’s
thesis, University of Technology in Poznan. See also
https://www.supermemo.com/en/archives1990-2015/
english/ol/sm2.
Wo´zniak, P. (1999). Effective learning: Twenty rules of
formulating knowledge. https://www.supermemo.co
m/en/archives1990-2015/articles/20rules. Accessed:
2022-05-09.
Wo´zniak, P. (2018a). The true history of spaced repetition.
https://www.supermemo.com/en/articles/history.
Accessed: 2022-05-09.
Wo´zniak, P. (2018b). Universal metric for cross-comparison
of spaced repetition algorithms. https://supermemo.
guru/wiki/Universal_metric_for_cross-comparison_
of_spaced_repetition_algorithms. Accessed: 2022-
05-09.
Wo´zniak, P. A., Gorzela
´
nczyk, E. J., and A., M. J. (2005).
Building memory stability through rehearsal. https:
//www.supermemo.com/en/archives1990-2015/articl
es/stability. Accessed: 2022-05-09.
Yang, Z., Shen, J., Liu, Y., Yang, Y., Zhang, W., and Yu, Y.
(2020). Tads: learning time-aware scheduling policy
with dyna-style planning for spaced repetition. In Pro-
ceedings of the 43rd International ACM SIGIR Con-
ference on Research and Development in Information
Retrieval, pages 1917–1920.
Modeling Spaced Repetition with LSTMs
95