-get pronoun in the reading passage. Questions were
automatically generated from existing human made
paragraphs so we believe that the present study con-
tributes to reducing teacher burden on creating those
questions and accommodating learners with abundant
question exercises as well.
In our proposed method, nonrestrictive relative
clauses (NRC) were utilised to generate the reading
passage using the sentence splitting technique. For
generating correct answers, the parse tree search and
the dependency parser worked together to enhance the
reliability of the generated answer. In creating distrac-
tors, the coreference resolver and the part-of-speech
tagger were employed.
According to the subjective evaluation of the gener-
ated questions, 53% of the questions were acceptable.
That means the half of the generated questions could
be used for the real test, but the performance still re-
mains far from fully automatic question generation.
Our system generated reading passage with 97% in
accuracy, correct answer with 70% in accuracy, and
distractors with 89% in accuracy; those results show
promising potential to generate the reference ques-
tions automatically. With the current performance
as is, it will be practical to incorporate the proposed
components into a kind of authoring system for creat-
ing reference questions, so as to reduce the burden of
human experts in creating questions.
At the same time, we need to further refine each com-
ponent generation module to obtain a better perfor-
mance of the total system. For instance, in order
to improve the performance of correct answer gen-
eration, checking the agreement between the relative
clause and the antecedent could help antecedent iden-
tification perform better. In order to remedy the ani-
macy error in distractor generation illustrated in Fig-
ure 7, we could incorporate a named entity recogniser
in our system to distinguish a person from an inani-
mate entity. We are also planning to conduct exper-
iments on real English learners to see to what extent
the automatically generated questions by our method
could discriminate test takers’ ability in reading com-
prehension.
REFERENCES
Brennan, S. E., Friedman, M. W., and Pollard, C. J. (1987).
A centering approach to pronouns. In Proceedings of
the 25th annual meeting on Association for Compu-
tational Linguistics, pages 155–162. Association for
Computational Linguistics.
Chen, D. and Manning, C. (2014). A fast and accurate de-
pendency parser using neural networks. In Proceed-
ings of the 2014 Conference on Empirical Methods in
Natural Language Processing (EMNLP), pages 740–
750, Doha, Qatar. Association for Computational Lin-
guistics.
Davis, B. G. (2009). Tools for teaching. John Wiley & Sons.
Gordon, P. C. and Scearce, K. A. (1995). Pronominalization
and discourse coherence, discourse structure and pro-
noun interpretation. Memory & Cognition, 23(3):313–
323.
Grosz, B. J., Joshi, A. K., and Weinstein, S. (1995). Cen-
tering: A framework for modeling the local coherence
of discourse. Computational Linguistics, 21(2):203–
225.
Heilman, M. (2011). Automatic factual question generation
from text. PhD thesis, Carnegie Mellon University.
Karamanis, N., Ha, L. A., and Mitkov, R. (2006). Gen-
erating multiple-choice test items from medical text:
A pilot study. In Proceedings of the Fourth Inter-
national Natural Language Generation Conference,
pages 111–113. Association for Computational Lin-
guistics.
Klein, D. and Manning, C. D. (2003). Accurate unlexical-
ized parsing. In Proceedings of the 41st Annual Meet-
ing on Association for Computational Linguistics-
Volume 1, pages 423–430. Association for Computa-
tional Linguistics.
Lee, H., Chang, A., Peirsman, Y., Chambers, N., Surdeanu,
M., and Jurafsky, D. (2013). Deterministic coref-
erence resolution based on entity-centric, precision-
ranked rules. Computational Linguistics, 39(4):885–
916.
Levy, R. and Andrew, G. (2006). Tregex and Tsurgeon:
tools for querying and manipulating tree data struc-
tures. In Proceedings of the fifth international confer-
ence on Language Resources and Evaluation, pages
2231–2234. ELRA.
Service, E. T. (2012). Official Guide to the TOEFL Test, 4th
Edition. McGraw-Hill Education.
Siddharthan, A. (2002). Resolving attachment and clause
boundary ambiguities for simplifying relative clause
constructs. In Proceedings of the Student Workshop,
40th Meeting of the Association for Computational
Linguistics (ACL02), pages 60–65.
Skalban, Y., Ha, L. A., Specia, L., and Mitkov, R. (2012).
Automatic question generation in multimedia-based
learning. In Proceedings of COLING 2012: Posters,
pages 1151–1160, Mumbai, India. The COLING 2012
Organizing Committee.
Susanti, Y., Iida, R., and Tokunaga, T. (2015). Automatic
generation of english vocabulary tests. In Proceed-
ings of the 7th International Conference on Computer
Supported Education (CSEDU 2015), pages 77–78.
Toutanova, K., Klein, D., Manning, C. D., and Singer, Y.
(2003). Feature-rich part-of-speech tagging with a
cyclic dependency network. In Proceedings of the
2003 Conference of the North American Chapter of
the Association for Computational Linguistics on Hu-
man Language Technology-Volume 1, pages 173–180.
Association for Computational Linguistics.
Walker, M. A., Joshi, A. K., and Prince, E. F. (1998). Cen-
tering theory in discourse. Oxford University Press.
CSEDU 2017 - 9th International Conference on Computer Supported Education
386