Coleman, C., Yeh, C., Mussmann, S., Mirzasoleiman, B.,
Bailis, P., Liang, P., Leskovec, J., and Zaharia, M.
(2020). Selection via proxy: Efficient data selection
for deep learning. In International Conference on
Learning Representations (ICLR).
Davidson, T., Warmsley, D., Macy, M., and Weber, I.
(2017). Automated hate speech detection and the
problem of offensive language. In Proceedings of
the International AAAI Conference on Web and Social
Media, volume 11.
Devlin, J., Chang, M.-W., Lee, K., and Toutanova, K.
(2019). Bert: Pre-training of deep bidirectional trans-
formers for language understanding. In Proceed-
ings of the 2019 Conference of the North American
Chapter of the Association for Computational Lin-
guistics: Human Language Technologies, volume 1,
pages 4171–4186.
Doherty, R. A. and Sorenson, P. (2015). Keeping users in
the flow: mapping system responsiveness with user
experience. Procedia Manufacturing, 3:4384–4391.
Dor, L. E., Halfon, A., Gera, A., Shnarch, E., Dankin, L.,
Choshen, L., Danilevsky, M., Aharonov, R., Katz, Y.,
and Slonim, N. (2020). Active learning for bert: An
empirical study. In Proceedings of the 2020 Confer-
ence on Empirical Methods in Natural Language Pro-
cessing (EMNLP), pages 7949–7962.
Fails, J. A. and Olsen Jr, D. R. (2003). Interactive machine
learning. In Proceedings of the 8th international con-
ference on Intelligent user interfaces, pages 39–45.
Hu, R., Mac Namee, B., and Delany, S. J. (2016). Active
learning for text classification with reusability. Expert
systems with applications, 45:438–449.
Joulin, A., Grave, E., Bojanowski, P., Douze, M., J
´
egou,
H., and Mikolov, T. (2016). Fasttext.zip: Com-
pressing text classification models. arXiv preprint
arXiv:1612.03651.
Joulin, A., Grave,
´
E., Bojanowski, P., and Mikolov, T.
(2017). Bag of tricks for efficient text classification.
In Proceedings of the 15th Conference of the Euro-
pean Chapter of the Association for Computational
Linguistics, volume 2, pages 427–431.
Lewis, D. D. and Catlett, J. (1994). Heterogeneous uncer-
tainty sampling for supervised learning. In Machine
learning proceedings 1994, pages 148–156. Elsevier.
Lewis, D. D. and Gale, W. A. (1994). A sequential algo-
rithm for training text classifiers. In SIGIR’94, pages
3–12. Springer.
Lewis, D. D., Yang, Y., Russell-Rose, T., and Li, F. (2004).
Rcv1: A new benchmark collection for text catego-
rization research. Journal of machine learning re-
search, 5(Apr):361–397.
Lowell, D., Lipton, Z. C., and Wallace, B. C. (2019). Prac-
tical obstacles to deploying active learning. In Pro-
ceedings of the 2019 Conference on Empirical Meth-
ods in Natural Language Processing and the 9th Inter-
national Joint Conference on Natural Language Pro-
cessing (EMNLP-IJCNLP), pages 21–30.
Maalej, W., Kurtanovi
´
c, Z., Nabil, H., and Stanik, C.
(2016). On the automatic classification of app reviews.
Requirements Engineering, 21(3):311–331.
Martin, G. L. and Corl, K. G. (1986). System response time
effects on user productivity. Behaviour & Information
Technology, 5(1):3–13.
Peters, M. E., Neumann, M., Iyyer, M., Gardner, M., Clark,
C., Lee, K., and Zettlemoyer, L. (2018). Deep con-
textualized word representations. In Proceedings of
NAACL-HLT, pages 2227–2237.
Prabhu, A., Dognin, C., and Singh, M. (2019). Sampling
bias in deep active classification: An empirical study.
In Proceedings of the 2019 Conference on Empirical
Methods in Natural Language Processing and the 9th
International Joint Conference on Natural Language
Processing (EMNLP-IJCNLP), pages 4058–4068.
Pranckevi
ˇ
cius, T. and Marcinkevi
ˇ
cius, V. (2017). Compari-
son of naive bayes, random forest, decision tree, sup-
port vector machines, and logistic regression classi-
fiers for text reviews classification. Baltic Journal of
Modern Computing, 5(2):221.
Qiu, X., Sun, T., Xu, Y., Shao, Y., Dai, N., and Huang,
X. (2020). Pre-trained models for natural language
processing: A survey. Science China Technological
Sciences, pages 1–26.
Reimers, N. and Gurevych, I. (2019). Sentence-bert: Sen-
tence embeddings using siamese bert-networks. In
Proceedings of the 2019 Conference on Empirical
Methods in Natural Language Processing and the 9th
International Joint Conference on Natural Language
Processing (EMNLP-IJCNLP), pages 3982–3992.
Scheffer, T., Decomain, C., and Wrobel, S. (2001). Active
hidden markov models for information extraction. In
International Symposium on Intelligent Data Analy-
sis, pages 309–318. Springer.
Settles, B. (2009). Active learning literature survey. In
ComputerSciences Technical Report 1648. University
of Wisconsin–Madison.
Settles, B. (2011). From theories to queries: Active learn-
ing in practice. In Active Learning and Experimental
Design workshop In conjunction with AISTATS 2010,
pages 1–18. JMLR.
Sugiyama, M. and Nakajima, S. (2009). Pool-based active
learning in approximate linear regression. Machine
Learning, 75(3):249–274.
Tolia, N., Andersen, D. G., and Satyanarayanan, M.
(2006). Quantifying interactive user experience on
thin clients. Computer, 39(3):46–52.
Tomanek, K. and Morik, K. (2011). Inspecting sample
reusability for active learning. In Active Learning and
Experimental Design workshop In conjunction with
AISTATS 2010, pages 169–181. JMLR.
Yang, Y. (1999). An evaluation of statistical approaches to
text categorization. Information retrieval, 1(1):69–90.
Yarlagadda, S., Scroggins, D. J., Cao, F., Devabhaktuni, Y.,
Buitron, F., and Brown, E. T. (2021). Doctable: Table-
oriented interactive machine learning for text corpora.
In 2021 IEEE Workshop on Machine Learning from
User Interactions (MLUI), pages 1–11. IEEE.
Zhu, J. and Ma, M. (2012). Uncertainty-based active learn-
ing with instability estimation for text classification.
ACM Transactions on Speech and Language Process-
ing (TSLP), 8(4):1–21.
ICAART 2023 - 15th International Conference on Agents and Artificial Intelligence
32