3.2 Student Answer Normalization
Although model answers are given for each
question, student responses significantly have so
many variants of the correct or incorrect answers.
So, identifying the characteristics of each question is
also important.
Table 1: Example of student responses and result of the
normalization.
Answer
t
es
Original text Normalized text
Model
answe
마음을치유
Heal the min
마음을치유
Heal the min
response 1
A: 마음을치유
A: heal the min
마음을치유
Heal the min
response 2
마음을치유
Healthemin
마음을치유
Heal the min
response 3
(상처받은)마음을치유
Heal the
hurted
min
상처받은마음을치유
Heal the hurted min
For instance, human graders should consider
spelling errors about several questions of the Korean
language courses. The system should normalize
every student responses including model answers
with certain considerations. Three of four options
are used for normalization stage, except the cueword
option. Table 1 is an example of student responses
and result of the normalization with all options
enabled; spelling error correction, word-spacing
error correction and eliminating unnecessary
characters or symbols.
3.3 Automated Scoring
Once an answer template file and normalized student
responses are prepared, the system is now ready to
proceed to the actual scoring phase. Automated
scoring process consists of four sequential steps;
model answers matching, high frequency answers
exact matching, concept-based assessment and
cueword-based handling for incorrect responses. A
cueword-based task can be skipped by the option.
Each scoring step processes every student responses,
and then passes the unprocessed responses to the
next step.
Model answers and high frequency answers are
already evaluated by human graders and normalized
by the system. Since the information which are
stored in the template file and student responses are
already normalized, scoring process can be done
with exact matching method.
After exact matching, the system proceeds to the
concept-based scoring step. A concept consists of
one or more tokens. There are two token types of
lexical and grammatical morpheme. A Korean word
is split into two tokens through morphological
analysis. The system produces concepts from
unprocessed student responses then tries to find a
concept from the answer template in a
conceptualized student response. Concept-based
method can handle responses even if part of the
answer matches with a concept. Figure 3 shows how
the system handles student responses with concepts.
The fourth element in the concept has been modified
to an asterisk ‘*’ so that any token is matched to this
token.
Figure 3: Example of concept matching.
Last step handles incorrect responses with
cueword-based methods. Cue word is an essential
keyword of the question for writing a correct
response. In contrast to the previous step, each
response will be treated as incorrect if the response
does not include any of the cue words. As mentioned
earlier, this step can be skipped by the option. If
cueword option is disabled, unprocessed responses
after concept-based step will be treated as non-
assessed. Non-assessed responses are passed to a
post processing module.
3.4 Post Processing
After automated scoring, unprocessed student
responses may exist. The purpose of post processing
is merging student responses. As a result, a list of
merged concepts will be produced. Human graders
can examine the produced list and select concepts to
create.
Post processing will proceed in the following
order; conceptualizing and sorting student responses
then merging into concepts. First of all, every
student responses are conceptualized. Then post
processor sorts student responses by size because
merging concept requires same sized responses. If
the size of target concepts are same, the system sorts
them alphabetically.
Next step is merging the concepts. The system
investigates every responses of the same size and
picks targets for merging. With the picked
responses, the processor performs a comparison,
CSEDU2014-6thInternationalConferenceonComputerSupportedEducation
228