mechanical pulses at a single frequency within the
natural range of the human voice. The transducer
uses an a coil-magnet arrangement that vibrates
against a diaphragm when the output of an electrical
oscillator is applied to its winding. The device is
pressed against the mandible and this vibrates the
pharynx which in turn resonates the air in the vocal
and/or nasal tract. The vibrations are formed into
speech by the articulators of the upper vocal tract.
Research to date has focused on the
improvement of the quality of speech produced by
the electro-larynx. Some significant contributions
have been made by (Houston et al. 1999) who
developed an electro-larynx which used digital
signal processing to create a superior quality of
sound. (Shoureshi et al. 2003) used neural-based
signal processing and smart materials to improve the
sound created. (Liu et al. 2006) and (Cole et al.
1997) focused on removing the buzzing sound
created by the transducer. (Uemi et al. 1994)
developed a system that utilized measurements from
air pressure that was obtained from a resistive
component that was placed over the stoma to
maintain the electrolarynx’s fundamental frequency.
(Ma et al. 1999) used cepstral analysis of speech to
replace the electrolarynx excitation signal with a
normal speech excitation signal.
Despite all the acoustic improvements these
studies have shown, they have however, been
performed in isolation and have been deemed to be
difficult to implement into the existing technology.
Therefore, the basic concept of design which was
first introduced by Barney et al. in the 1950’s
remains the same to this day. It has been shown that
up to 50%-66% of all laryngectomees use some
form of electrolarynx speech (Gray et al. 1976) and
(Hillman et al. 1998): either as a method of
communication for speech rehabilitation post-
surgery or as a reliable back-up in situations where
esophageal or trcheo-esophageal speech is proving
difficult.
1.2 Speech Intelligibility
When determining the intelligibility of a speech
signal, it is important to choose a suitable linguistic
level at which to make measurements. Is it necessary
to measure the accuracy at with which each phonetic
element is communicated in order to assess whether
each word is identifiable. It is also necessary to
investigate whether the communication of a sentence
is clear.
This type of linguistic dismissal can introduce an
additional difficulty in that individual human
listeners will ultimately differ in their capability to
make use of these linguistic constraints. Even
though it may be deemed necessary to assess the
utility of a particular channel in order to convey the
meanings of real spoken utterances, listeners will
inevitably vary in their capacity to comprehend the
speech, depending on their own linguistic ability.
Many speech intelligibility tests consist of either
phonetic unit, which are composed into: nonsense
syllables, words which are used in isolation or in
short sentences spoken in one breath for comfort
(Crystal et al 1982) and (Mitchell et al. 1996).
An issue that arises through the use of nonsense
syllables is that many listeners could require training
in order to be able to identify the component
phonetic units, and they may be confused by
phonemes which don't compare well with the
spelling e.g. there, their, they’re. Therefore by
limiting listener reply’s to real words thus allowing
them to respond in ordinary spelling. This can
however introduce other difficulties: firstly, that
varying listeners may possess differing degrees of
familiarity with the words that are being used;
secondly, that some words are memorable and
having heard a word once, some listeners may be
biased in their usage of a particular word another
time.
A possible solution to these problems includes
the formulation of multiple word lists of reasonable
difficulty, allowing a listener to be used within a test
more than once. Another option is to create tests
consisting of closed response sets, making every
listener needing to make the matching choices about
the word which is under test.
(Egan et al. 1948) pioneered one of the first lists
of words for an intelligibility test in 1948. He
created the list by using the concept of “phonetic
balance” which meant that the relative frequency of
the phonemes in the word lists corresponded to the
relative frequency of phonemes in conversational
speech. He constructed 20 lists containing 50
monosyllabic words and his intention was to balance
average difficulty and range of difficulty throughout
the lists whilst ensuring that the phonetic units that
were present were represented equally.
2 METHODS
2.1 Novel Actuator Design
This design consists of a simple pager motor (which
is typically found in a mobile phone) attached to a
thin piece of high-density polyethylene (HDPE) by
BIOSIGNALS 2011 - International Conference on Bio-inspired Systems and Signal Processing
266