Table 2: Results and effective features in Exp. 1, 2 and 3.
Exp. effective features accuracy F-measure
Exp. 1 S1, S2, S3, S4, S5, S6 86.04% 0.8443
Exp. 2 S1, S4, S5, S9, S12, S15, S16 91.65% 0.4773
Exp. 3 S1, S4, S5, S9, S12, S16 86.04% 0.6503
S1 word unigrams of the target sentence
S2 word bigrams of the target sentence
S3 word trigrams of the target sentence
S4 number of sentence of the question/answer and
sentence number of the target sentence
S5 number of words of the question/answer
S6 word unigrams of the non-target sentences and
relative position to the target sentence (be-
fore/after)
S7 word bigrams of the non-target sentences and rel-
ative position to the target sentence (before/after)
S8 word trigrams of the non-target sentences and rel-
ative position to the target sentence (before/after)
S9 word unigrams of the question
S10 word bigrams of the question
S11 word trigrams of the question
S12 word unigrams of the important sentence in the
question
S13 word bigrams of the important sentence in the
question
S14 word trigrams of the important sentence in the
question
S15 nouns which are found both in the question and
its answer
S16 number of nouns which are found both in the
question and its answer
Figure 1: The features used in machine learning (SVM) on
Yahoo! chiebukuro.
Exp. 1 extract important sentences from questions
posted on a Q&A site
Exp. 2 extract sentences including clues as to which
information should be described in a question
from answers posted on a Q&A site
Exp. 3 extract sentence including information which
a questioner does not know but is easy to confirm
from answers posted on a Q&A site
We conducted Exp. 1, 2, and 3 using TinySVM (Ku-
doh 00) with polynomial kernel (d = 2, c = 1). In this
experiments, we used 2219 questions and their 2251
answer in Table 1 as the experimental data.
All experimental results were obtained with 10-
fold cross-validation. To calculate the accuracyand F-
measure, the experimental data was manually tagged
in the preparation of the experiments.
Table 2 shows the results and effective features in
Exp. 1, 2, and 3.
Finally, we discuss the features which were not
designated as effective features in Exp. 2 and 3. Both
in Exp. 2 and 3, S6, S7, and S8 were not designated
as effective features. These features were based on
word n-grams in the non-target sentences of SVM ex-
traction process. It shows that sentences including in-
formation for supporting a learner to recognize what
he/she did not understand can be extracted, not by us-
ing non-target sentences of SVM extraction process.
Furthermore, it may show that although the user only
read sentences which include information for support-
ing a learner to recognize and never read other sen-
tences, he/she can understand and use it.
REFERENCES
Dumais, Banko, Brill, Lin, and Ng: Web question answer-
ing: Is more always better?, ACM SIGIR 2002, 2002.
Kiyota, Kurohashi, and Kido: “Dialog Navigator” A Ques-
tion Answering System based on Large Text Knowl-
edge Base, COLING02, 2002.
Matsuike, Zettsu, Oyama, and Tanaka, Supporting the
Query Modification by Making Keyword Formula of
an Outline of Retrieval Result, IEICE DEWS2005,
1C-i9, 2005 (in Japanese).
Xu and Croft: Query expansion using local and global doc-
ument analysis, ACM SIGIR 1996, 1996.
Hayashi and Kikui: Design and Implementation of Rewrit-
ing Support Functions in a Japanese Text Revi-
sion System, Trans. of IPSJ, Vol.32, No.8, 1991 (in
Japanese).
Yamazaki, Yamamura, and Ohnishi: A Computer Aided
System for Writing in the way of Changable
Styles, IEICE technical report, NLC98-54, 1998 (in
Japanese).
Kudoh: TinySVM: Support Vector Machines,
(http://chasen.org/
˜
taku/software/TinySVM/index.html,
2002).
Lin, Fernandes, Katz, Marton, and Tellex: Extracting an-
swers from the Web using knowledge annotation and
knowledge mining techniques, 11th TREC, 2002.
Watanabe, Nishimura, and Okada: A Question Answer Sys-
tem Based on Confirmed Knowledge Acquired from a
Mailing List, Internet Research, Vol.18, No.2, 2008.
Kurohashi and Kawahara: JUMAN Manual version 5.1,
Kyoto University, 2005 (in Japanese).
CSEDU 2009 - International Conference on Computer Supported Education
374