
3.4 Autonomous conversation
Besides networking communication, autonomous
communication is strongly required for single-
resident elderies user, as the pet robot could be a
partner and then that may avoid the user from loneli-
ness. Autonomous communication consists of speech
recognition and speech generation. Speech media are
very useful for elderies, who are not accustomed to
use computers straightforwardly.
The robot at the moment can tell his name, current
date/time to the user(s). The robot can also use over
200 Japanese words, including
• “Good morning!”,
• “Wake up!”,
• “Bye Bye”,
• etc.
If a user talk to the robot “Wake up!”, the the robot
talks back to the user the greeting and gives one ar-
bitrary health advice at random, which give user(s)
some feelings of “live creature”. Furthermore, for the
purpose of getting friendly, the robot can sing several
short songs.
Note that those speech is not synthesized but just
composed from parts of speech pieces, which were
previously recorded and edited from human speech.
This composed method found to be more natural to
listen to, and that leads quite important characteris-
tics for elderies to be easy to communicate rather than
synthesized ones.
4 SYSTEM ARCHITECTURE
To reinforce interaction, including communication
with user(s), more naturally and friendlily, the robot
has some sensors and motors.
Figure 3 presents functions of the robot skeleton.
Those components enable the robot to behave
much more like a real (living) pet, which afford to be
more friendly and easy to contact (Gibson and Walk,
1960).
Sensors work for catching some signals of friend-
liness, which makes the robot cheer up/down, as well
as for interrupting its action, speech, etc. Considering
necessity and sufficiency, the robot has four motors;
one for both ears, one for both eyes, one for the nose,
and one for the neck. Motions generates by those mo-
tors symbolize emotions of the robot, which is essen-
tial for our object. For instance, the head, followed by
neck, can move vertically, which imply the emotion
of bowing, and horizontally, which imply negation.
The robot consists of two units, stuffed toy unit and
control unit, and those two are connected by a serial
line. the toy (doll) part has motors and sensors, and
Figure 3: Functions of the skeleton.
motion commands and sensed signals are sent to the
control unit. To afford user(s) to use easily, or pleas-
antly, a camera is embedded into a toy camera, mi-
crophones are set at both ears, and a speaker is set
just under the mouth.
To/from the information center, communication, or
information interchange is done through the Internet.
Telephone calls are not to use VoIP protocol, but done
using facility of cellular phones through public tele-
phone network.
5 DISCUSSION
We have examined a field test in Ikeda-city in Japan.
Targets are 7 single-resident elderies, 3 males (ave.
83 years old) and 4 females (ave. 78 years old). Each
person use the pet-type robot ave. 62 days, and in
that period we have done 4 interviews and 2 question-
naires.
Table 1 presents questionnaire results, where “Q”
stands for Questions, “V” for “Very good”, “G” for
“Good”, “N” for “Not so good”, and “B” for “Bad”.
It actually shows our system can be used quite well,
though it is not complete. Especially, the item of “un-
derstanding” is not of good point which may cause
from insufficiency of speech recognition, and thus the
function of speech recognition and autonomous con-
versation should be improved.
Table 2 shows statistics of conversation data, de-
rived from log-messages in the center (server). The
system is used Approximately each two days, which
seems to be moderate. Users turn on the switch
twice a day, which may causes from mail checking
on morning and evening . Conversation time is about
6 minutes, and it may be enough for get and give in-
formation. Conversation success rate is counted by
hand from logs of the server. 46.8 % is not so good
ICINCO 2004 - ROBOTICS AND AUTOMATION
324