ommendations: Users should not feel stressed or dic-
tated to when interacting with the system. While
visual feedback is displayed by the system, users
have the choice or autonomy over what feedback they
choose to react to. The other point in relation to feed-
back is that the feedback is non-directive. The pur-
pose of the feedback is to make the user aware of
their speaking behaviour, not to tell them what to do.
During a review of the multimodal system for pub-
lic speaking, Presentation Trainer, experts found that
the system should ’shift focus and become a tool to
develop awareness of nonverbal communication, in-
stead of correcting it’ (Schneider et al., 2017). Users
should see themselves represented on screen as a full
3D avatar because this allows them to assess their full
3D body pose but does not distract them with details
of their personal appearance. The research question
posed in this paper is, how do users respond to these
two approaches?
Figure 1: The Spheres of Positive Computing (Calvo and
Peters, 2014).
3 SYSTEM DESCRIPTION
We will present a brief description of our multi-
modal Positive Computing system for public speak-
ing. It has been described in greater depth in our
previous work (Dermody and Sutherland, 2016),(Der-
mody and Sutherland, 2018a),(Dermody and Suther-
land, 2018b).The term ‘multimodal’ refers to the fact
that the system detects multiple speaking modes in
the speaker such as their gestures, voice and eye con-
tact. Gestures, body posture, gestures and gaze di-
rection are all important aspects of public speaking
(Toastmasters International, 2011), (Toastmasters In-
ternational, 2008). The user can select if they want
to receive feedback on all speaking modes or a subset
of them. The system consists of a Microsoft Kinect 1
connected to a laptop. The system uses the Microsoft
Kinect to sense the user’s body movements, facial ex-
pressions and voice. The user stands in front of the
system and speaks. The speaker can see themselves
represented on screen as an avatar. Visual feedback is
given on a laptop screen in front of the user. The feed-
back is displayed in proximity to the area it relates to.
The objective of the system is to enable the user to
speak freely without being interrupted, distracted or
confused by the visual feedback on screen.
3.1 System Feedback
Feedback on the speaker’s voice is given by a track,
which consists of a moving line where the horizontal
axis represents time and the vertical axis represents
pitch. The width of the line represents volume i.e. the
loudness with which the user speaks. Syllables are
represented by different colours. The density of the
syllables represents the speed with which the user is
speaking.
The system also displays a visual feedback icon
near the avatar’s hands to indicate that the user is
touching their hands. Feedback is also given on gaze
direction using arrows near the avatar’s head. The
aforementioned feedback can be seen in Figure 2. For
the purposes of this study, we chose to only look at
these feedback items. However, the system can pro-
vide feedback on other speaking behaviours as de-
tailed in our previous work. These particular speaking
behaviours were chosen because they have been rated
as important by experts in public speaking (Toastmas-
ters International, 2011), (Toastmasters International,
2008). A speaker’s open gestures and varying eye
contact have been found to impact on audience en-
gagement.
4 STUDY DESIGN
The study had 10 participants (4F, 6M). Participants
were selected from the staff and student body at our
university. The study was designed to be a one-
time recruitment with a duration of 25 minutes per
participant. The participants completed a prelimi-
nary questionnaire on demographic information and a
post-questionnaire. 9 of the participants were novice
speakers who had done some public speaking but
wished to improve their skills in this area. One partic-
ipant described himself as an accomplished speaker
who was keen to participate in the study. None of the
participants had used a multimodal system for public
speaking previously.
The post-questionnaire consisted of eleven items.
User experience was evaluated using three questions
on naturalness, motivation to use the application again
HUCAPP 2019 - 3rd International Conference on Human Computer Interaction Theory and Applications
196