proach is proposed for questionnaire classification by
using concatenated question-answers pairs to perform
text classification rather than separately analysing the
pairs. A pre-trained BERT is used in the training
to get contextual and bidirectional word embeddings
to capture the correlation between the pairs in the
whole text. A BiLSTM layer on top of BERT is
used to represent the sequential dependencies of word
embeddings for further improvement. We utilized a
data-centric approach, using clustering to group in-
consistent data to mitigate the effects of noise caused
by open-ended questions that provide deeper insights
into a questionnaire and affect the annotation process.
The model architecture we proposed for questionnaire
classification performed better than a simple text clas-
sification architecture. Also, we have observed mean-
ingful improvement in the classification performance
with models trained on the data where the clustering
approach is applied.
The proposed novel approach for classification
can be used in a dataset in a similar setting that has
multiple question-answer pairs with the task of clas-
sifying these pairs as a single unit and not as sepa-
rate parts. The method we used doesn’t involve any
domain-centric or language-centric technique, thus
one can assume the methods are applicable to simi-
lar data in other contexts or languages. Our work fo-
cuses on Turkish data in the banking domain due to
not having any public data available. However, the
results prove the classification is successful in a noisy
dataset that is labelled without supervision.
For future research, we intend to experiment with
semi-supervised methods like self-learning to lessen
the impact of incorrect labels. This will help us to
cover the examples in our dataset that our approach
could not affect. We also believe data-centric ap-
proaches will improve NLP applications, especially
for low-resource languages like Turkish. And using a
data-centric approach to handle inconsistent data will
further help in situations where manual labour is not
feasible. For further work, we aim to develop our
method using Explainable AI approaches to under-
stand which question-answer pair mostly contributed
to the outcome.
REFERENCES
Bhowmick, A. and Hazarika, S. M. (2016). Machine learn-
ing for e-mail spam filtering: Review,techniques and
trends.
Bojanowski, P., Grave, E., Joulin, A., and Mikolov, T.
(2017). Enriching word vectors with subword infor-
mation. Transactions of the Association for Computa-
tional Linguistics, 5:135–146.
Clark, K., Luong, M.-T., Le, Q. V., and Manning, C. D.
(2020). Electra: Pre-training text encoders as dis-
criminators rather than generators. arXiv preprint
arXiv:2003.10555.
Devlin, J., Chang, M.-W., Lee, K., and Toutanova, K.
(2019). BERT: Pre-training of deep bidirectional
transformers for language understanding. In Pro-
ceedings of the 2019 Conference of the North Amer-
ican Chapter of the Association for Computational
Linguistics: Human Language Technologies, Volume
1 (Long and Short Papers), pages 4171–4186, Min-
neapolis, Minnesota. Association for Computational
Linguistics.
Graves, A. and Schmidhuber, J. (2005). Framewise
phoneme classification with bidirectional lstm and
other neural network architectures. Neural networks,
18(5-6):602–610.
Jiang, Z., Yang, M., Tsirlin, M., Tang, R., Dai, Y., and
Lin, J. (2023). “low-resource” text classification: A
parameter-free classification method with compres-
sors. In Findings of the Association for Computational
Linguistics: ACL 2023, pages 6810–6828.
Jiang, Z.-H., Yu, W., Zhou, D., Chen, Y., Feng, J., and Yan,
S. (2020). Convbert: Improving bert with span-based
dynamic convolution. Advances in Neural Informa-
tion Processing Systems, 33:12837–12848.
Jolliffe, I. T. (1986). Principal Component Analysis.
Springer-Verlag, Berlin; New York.
Kaliyar, R. K., Goswami, A., and Narang, P. (2021). Fake-
bert: Fake news detection in social media with a bert-
based deep learning approach. Multimedia tools and
applications, 80(8):11765–11788.
Larson, S. and Leach, K. (2022). A survey of intent classi-
fication and slot-filling datasets for task-oriented dia-
log.
Mikolov, T., Chen, K., Corrado, G., and Dean, J. (2013).
Efficient estimation of word representations in vector
space.
Munikar, M., Shakya, S., and Shrestha, A. (2019). Fine-
grained sentiment classification using bert. 2019 Arti-
ficial Intelligence for Transforming Business and So-
ciety (AITB), 1:1–5.
Northcutt, C., Jiang, L., and Chuang, I. (2021). Confident
learning: Estimating uncertainty in dataset labels. J.
Artif. Int. Res., 70:1373–1411.
Pennington, J., Socher, R., and Manning, C. (2014). GloVe:
Global vectors for word representation. In Proceed-
ings of the 2014 Conference on Empirical Methods in
Natural Language Processing (EMNLP), pages 1532–
1543, Doha, Qatar. Association for Computational
Linguistics.
Reimers, N. and Gurevych, I. (2019). Sentence-bert: Sen-
tence embeddings using siamese bert-networks. In
Proceedings of the 2019 Conference on Empirical
Methods in Natural Language Processing. Associa-
tion for Computational Linguistics.
Sanh, V., Debut, L., Chaumond, J., and Wolf, T. (2019).
Distilbert, a distilled version of bert: smaller, faster,
cheaper and lighter. arXiv preprint arXiv:1910.01108.
Schweter, S. (2020). Berturk - bert models for turkish.
Classification of Questionnaires with Open-Ended Questions
419