cause it mirrors the inherent data distribution. Con-
versely, strategies like tag count or fluctuation history
appear to select beneficial subsets of data, thereby
mitigating errors in low-frequency labels. This is also
illustrated in Figure 5, where these strategies outper-
form random selection even in the region where the
data sets begin to converge (∼ 10k documents), fur-
ther demonstrating their efficacy in reducing bias.
7 CONCLUSION
This paper conducted a comparative performance
analysis of Active Learning (AL) strategies in the con-
text of entity recognition (ER). Based on a systematic
selection of corpora and strategies, guided by a com-
prehensive scoping review, we conducted 115 exper-
iments within a standardized evaluation setting. Our
assessment referred to both performance and runtime.
We identified conditions where AL achieved signifi-
cant improvements, as well as situations where its re-
sults are more limited. Two strategies came out as
clear winners: tag count and fluctuation history.
Future work may expand the evaluation to a
broader range of AL strategies and corpora, includ-
ing those that do not adhere to the rigorous construc-
tion standards of benchmark datasets, to explore their
specific challenges.
REFERENCES
Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J. D.,
Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G.,
Askell, A., Agarwal, S., Herbert-Voss, A., Krueger,
G., Henighan, T., Child, R., Ramesh, A., Ziegler, D.,
Wu, J., Winter, C., Hesse, C., Chen, M., Sigler, E.,
Litwin, M., Gray, S., Chess, B., Clark, J., Berner,
C., McCandlish, S., Radford, A., Sutskever, I., and
Amodei, D. (2020). Language Models are Few-Shot
Learners. In Advances in Neural Information Pro-
cessing Systems, volume 33, pages 1877–1901. Cur-
ran Associates, Inc.
Chen, Y., Lasko, T. A., Mei, Q., Denny, J. C., and Xu, H.
(2015). A study of active learning methods for named
entity recognition in clinical text. Journal of Biomed-
ical Informatics, 58:11–18.
Covington, M. A. and McFall, J. D. (2010). Cutting the
Gordian Knot: The Moving-Average Type–Token Ra-
tio (MATTR). Journal of Quantitative Linguistics,
17(2):94–100.
Esuli, A., Marcheggiani, D., and Sebastiani, F. (2010).
Sentence-based active learning strategies for informa-
tion extraction. In CEUR Workshop Proceedings, vol-
ume 560, pages 41–45.
Feng, S. Y., Gangal, V., Wei, J., Chandar, S., Vosoughi,
S., Mitamura, T., and Hovy, E. (2021). A Survey of
Data Augmentation Approaches for NLP. In Zong,
C., Xia, F., Li, W., and Navigli, R., editors, Find-
ings of the Association for Computational Linguistics:
ACL-IJCNLP 2021, pages 968–988, Online. Associa-
tion for Computational Linguistics.
Finlayson, M. A. and Erjavec, T. (2017). Overview of An-
notation Creation: Processes and Tools. In Ide, N. and
Pustejovsky, J., editors, Handbook of Linguistic An-
notation, pages 167–191. Springer Netherlands, Dor-
drecht.
Hassan, S. and Alikhani, M. (2023). D-CALM: A Dynamic
Clustering-based Active Learning Approach for Mit-
igating Bias. In Rogers, A., Boyd-Graber, J., and
Okazaki, N., editors, Findings of the Association for
Computational Linguistics: ACL 2023, pages 5540–
5553, Toronto, Canada. Association for Computa-
tional Linguistics.
Huang, K.-H. (2021). DeepAL: Deep Active Learning in
Python.
Jayakumar, T., Farooqui, F., and Farooqui, L. (2023).
Large Language Models are legal but they are not:
Making the case for a powerful LegalLLM. In
Preo\textcommabelowtiuc-Pietro, D., Goanta, C.,
Chalkidis, I., Barrett, L., Spanakis, G., and Aletras,
N., editors, Proceedings of the Natural Legal Lan-
guage Processing Workshop 2023, pages 223–229,
Singapore. Association for Computational Linguis-
tics.
Jehangir, B., Radhakrishnan, S., and Agarwal, R. (2023).
A survey on Named Entity Recognition — datasets,
tools, and methodologies. Natural Language Process-
ing Journal, 3:100017.
Kaddour, J., Key, O., Nawrot, P., Minervini, P., and Kus-
ner, M. J. (2023). No Train No Gain: Revisiting
Efficient Training Algorithms For Transformer-based
Language Models. Advances in Neural Information
Processing Systems, 36:25793–25818.
Kettunen, K. (2014). Can Type-Token Ratio be Used
to Show Morphological Complexity of Languages?
Journal of Quantitative Linguistics, 21(3):223–245.
Kholghi, M., De Vine, L., Sitbon, L., Zuccon, G., and
Nguyen, A. (2017). Clinical information extraction
using small data: An active learning approach based
on sequence representations and word embeddings.
68(11):2543–2556.
Kholghi, M., Sitbon, L., Zuccon, G., and Nguyen, A.
(2015). External knowledge and query strategies in
active learning: A study in clinical information extrac-
tion. In Proceedings of the 24th ACM International
on Conference on Information and Knowledge Man-
agement, CIKM ’15, pages 143–152, New York, NY,
USA. Association for Computing Machinery.
Klie, J.-C., Bugert, M., Boullosa, B., Eckart de Castilho, R.,
and Gurevych, I. (2018). The INCEpTION Platform:
Machine-Assisted and Knowledge-Oriented Interac-
tive Annotation. In Proceedings of the 27th Interna-
tional Conference on Computational Linguistics: Sys-
tem Demonstrations, pages 5–9, Santa Fe, New Mex-
ico. Association for Computational Linguistics.
Kohl, P., Freyer, N., Kr
¨
amer, Y., Werth, H., Wolf, S., Kraft,
B., Meinecke, M., and Z
¨
undorf, A. (2023). ALE: A
Comparative Performance Analysis of Active Learning Strategies for the Entity Recognition Task
487