
Overcoming catastrophic forgetting in neural net-
works. Proceedings of the National Academy of Sci-
ences, 114(13):3521–3526.
Krizhevsky, A. (2009). Learning multiple layers of features
from tiny images. Technical Report.
LeCun, Y., Bottou, L., Bengio, Y., and Haffner, P. (1998).
Gradient-based learning applied to document recogni-
tion. Proceedings of the IEEE, 86(11):2278–2324.
Li, J., Zhou, P., Xiong, C., and Hoi, S. (2021). Prototyp-
ical contrastive learning of unsupervised representa-
tions. International Conference on Learning Repre-
sentations.
Li, Z. and Hoiem, D. (2016). Learning without forgetting.
European Conference on Computer Vision.
Lin, H., Zhang, B., Feng, S., Li, X., and Ye, Y. (2023).
Pcr: Proxy-based contrastive replay for online class-
incremental continual learning. Conference on Com-
puter Vision and Pattern Recognition.
Lopez-Paz, D. and Ranzato, M. (2017). Gradient episodic
memory for continual learning. Neural Information
Processing Systems conference.
Madaan, D., Yoon, J., Li, Y., Liu, Y., and Hwang, S. J.
(2022). Representational continuity for unsuper-
vised continual learning. International Conference on
Learning Representations.
Mallya, A. and Lazebnik, S. (2018). Packnet: Adding mul-
tiple tasks to a single network by iterative pruning.
Conference on Computer Vision and Pattern Recog-
nition.
Purushwalkam, S., Morgado, P., and Gupta, A. (2022). The
challenges of continuous self-supervised learning. Eu-
ropean Conference on Computer Vision.
Rusu, A. A., Rabinowitz, N. C., Desjardins, G., Soyer,
H., Kirkpatrick, J., Kavukcuoglu, K., Pascanu, R.,
and Hadsell, R. (2016). Progressive neural networks.
arXiv preprint arXiv:1606.04671.
Shin, H., Lee, J. K., Kim, J., and Kim, J. (2017). Continual
learning with deep generative replay. Neural Informa-
tion Processing Systems conference.
Thrun, S. (1995). A lifelong learning perspective for mobile
robot control. Elsevier.
Tiwari, R., Killamsetty, K., Iyer, R., and Shenoy, P. (2022).
Gcr: Gradient coreset based replay buffer selection for
continual learning. Conference on Computer Vision
and Pattern Recognition.
Wang, L., Zhang, X., Yang, K., Yu, L., Li, C., Hong, L.,
Zhang, S., Li, Z., Zhong, Y., and Zhu, J. (2022). Mem-
ory replay with data compression for continual learn-
ing. International Conference on Learning Represen-
tations.
Ye, M., Zhang, X., Yuen, P. C., and Chang, S.-F. (2019).
Unsupervised embedding learning via invariant and
spreading instance feature. Conference on Computer
Vision and Pattern Recognition.
Yoon, J., Yang, E., Lee, J., and Hwang, S. J. (2018).
Lifelong learning with dynamically expandable net-
works. International Conference on Learning Repre-
sentations.
Zbontar, J., Jing, L., Misra, I., LeCun, Y., and Deny, S.
(2021). Barlow twins: Self-supervised learning via
redundancy reduction. International Conference on
Machine Learning.
Zenke, F., Poole, B., and Ganguli, S. (2017). Continual
learning through synaptic intelligence. International
Conference on Machine Learning.
Zhang, C., Zhang, K., Pham, T. X., Niu, A., Qiao, Z., Yoo,
C. D., and Kweon, I. S. (2022). Dual temperature
helps contrastive learning without many negative sam-
ples: Towards understanding and simplifying moco.
Conference on Computer Vision and Pattern Recogni-
tion.
Zhang, H., Cisse, M., Dauphin, Y. N., and Lopez-Paz, D.
(2017). mixup: Beyond empirical risk minimization.
arXiv preprint arXiv:1710.09412.
Diverse Data Selection Considering Data Distribution for Unsupervised Continual Learning
535