4.2 Dialog Context Model
Our representation of a dialog context consists of
diverse pieces of discourse and subtask information
as shown in Table 2. The task of predicting the
probable user intention in a given dialog context can
be viewed as searching for dialog contexts that are
similar to the current one in dialog context space and
then inferring to the expected user intention from the
user intentions of the dialog contexts found.
Therefore, we can formulate the task as the k-nearest
neighbors (KNN) problem (Dasarathy, 1990). We
had a number of reasons for choosing instance-based
learning methodology. First, instance-based learning
provides high controllability for tuning the model
incrementally during operation, which is practically
very desirable property. Second, an elaborate
similarity function can be applied. Many of other
approaches, e.g. maximum entropy model used in
the utterance model, express the similarity between
states in a simplest manner through the features that
the states share, losing elaborate regularities between
features. For the dialog context model, we can easily
predict which features become important features to
measure similarity conditioning on certain values of
other features using general discourse knowledge.
For example, if the current system dialog act is
“inform”, the number of database query results
becomes an important feature. If the number of
results is greater than one, the most likely expected
user intention would be “select”. If the number of
results equals one, “ack” would be the most probable
intention. Otherwise, the users might want to modify
their requirements. Another example, if all
exchanges of information are confirmed and the
current system intention is “wh-question”, the
current system intention itself becomes the most
important feature to determine the next user
intention.
However, the conventional KNN model has two
drawbacks. First, it considers no longer the degree of
similarity after selecting k nearest contexts, hence
intentions that occur rarely cannot have a chance to
be chosen regardless of how close they are to the
given dialog context. The second drawback is that if
dialog contexts with, say, intention A, are locally
condensed rather than widely distributed, then A is
specifically fitted intention to the local region of the
dialog context. So the intention A should be given
greater preference than other intentions. To cope
with these drawbacks, we introduce a new concept,
locality, and take both similarity and locality into
account in estimating the probability distribution of
the dialog context model (Eq. 9, 10).
The similarity function is defined as the
following equation:
,
,
(7)
where is the number of features,
denotes the
feature functions,
the weighted parameters for
features. Our feature functions first include the
simplest tests, whether a feature is shared or not, for
each feature of a dialog context (Table 2). For
composite features, individual tests are also included
for each constituent to alleviate data sparseness
problems. For example, we include feature functions
not only for system intention but also for its
constituents, system dialog act and type of subtask.
In addition, we include a number of feature
functions which test the elaborate rules as illustrated
in the examples above. The weighted parameters are
given initial values based on general discourse and
task knowledge and optimized on the development
data set with minimum error rate criteria.
The locality function is the ratio between the
number of elements of the set
,
and the number of
elements of the set
:
,
,
,
(8)
where
|
,
and
,
|
,
and
is the set of k nearest neighbors of the given
dialog context .
The score function calculates the score of the
intention I based on the set of k nearest dialog
contexts using both similarity and locality:
,
,
,
,
(9)
To let the dialog context model be a probability
distribution, the score function is divided by the
normalization factor:
|
,
,
(10)
5 EXPERIMENTS
To verify the proposed model, we conducted a case
study for dialogs in a system devoted to immigration
into an English-speaking country. We used the
example based dialog management method (Lee et.
CSEDU 2010 - 2nd International Conference on Computer Supported Education
14