Figure 2: An exemplary speech signal (fragment) for a pa-
tient with polyp.
analysis, we propose using neural networks with the
capability of extracting the phoneme articulation pat-
tern for a given patient (articulation is an individual
patient feature) and the capability of assessment of its
replication in the whole examined signal. Preliminary
observations show that significant replication distur-
bances in time, appear for patients with the clinical
diagnosis of disease.
The capabilities mentioned are possessed by re-
current neural networks. One class of them are the
Elman neural networks (Elman, 1990). In real time-
decision making, an important thing is to speed up
a learning process for neural networks. Moreover,
accuracy of learning of signal patterns plays an im-
portant role. Therefore, in this paper, we propose
some improvement of learning ability of the Elman
networks by combining them with another kind of
recurrent neural networks, namely, the Jordan net-
works (Jordan, 1986) and by providing some addi-
tional modification.
Our paper is organized as follows. After introduc-
tion, we shortly describe a structure and features of
the modified Elman-Jordan neural network used for
supporting diagnosis of laryngopathies (Section 2). In
Section 3, we present results obtained by experiments
done on real-life data. Some conclusions and final re-
marks are given in Section 4.
2 RECURRENT NEURAL
NETWORKS
In most cases, neural network topologies can be di-
vided into two broad categories: feedforward (with no
loops and connections within the same layer) and re-
current (with possible feedback loops). The Hopfield
network, the Elman network and the Jordan network
are the best known recurrent networks. In the paper
we are interested in the two last ones.
In the Elman network (Figure 3) (Elman, 1990),
the input layer has a recurrent connection with the
hidden layer. Therefore, at each time step the output
values of the hidden units are copied to the input units,
which store them and use them for the next time step.
This process allows the network to memorize some
information from the past, in such a way to detect
periodicity of the patterns in a better manner. Such
capability can be exploited in our problem to recog-
nize temporal patterns in the examined speech signals.
The Jordan networks (Jordan, 1986) are similar to the
Elman networks. The context layer is, however, fed
from the output layer instead of the hidden layer. To
accelerate a learning (training) process of the Elman
neural network we propose a modified structure of the
network. We combine the Elman network with the
Jordan network and add another feedback for an out-
put neuron as it is shown in Figure 4.
The pure Elman network consists of four layers:
• an input layer (in our model: the neuron I
1
),
• a hidden layer (in our model: the neurons H
1
, H
2
,
..., H
40
),
• a context layer (in our model: the neurons C
1
, C
2
,
..., C
40
),
• an output layer (in our model: the neuron O
1
).
z
−1
is a unit delay here.
To improve some learning ability of the pure El-
man networks, we propose to add additional feed-
backs in network structures. Experiments described
in Section 3 validate this endeavor. We create (see
Figure 4):
• feedback between an output layer and a hidden
layer through the context neuron (in our model:
the neuron C
41
), such feedback is used in the Jor-
dan networks,
• feedback for an output layer.
A new network structure will be called the modified
Elman-Jordan network.
The Elman network, according to its structure, can
store an internal state of a network. There can be val-
ues of signals of a hidden layer in time unit t −1. Data
are stored in the memory context. Because of storing
values of a hidden layer for t −1 we can make predic-
tion for the next time unit for a given input value. In
the case of learning neural networks with different ar-
chitectures, we can distinguish three ways for making
prediction for x(t + s), where s > 1:
1. Training a network on values x(t), x(t − 1), x(t −
1), ....
IMPROVING LEARNING ABILITY OF RECURRENT NEURAL NETWORKS - Experiments on Speech Signals of
Patients with Laryngopathies
361