
the training parameters of the ANN.
The ANN consists from a fully connected and
biased with 2 layers network structure of 2:11:15:1.
The training for the ANN was carried out on a
Pentium-400 based PC. Learning rate was initially
equal to 0.9 and later it was changed to 0.7, 0.5, 0.3,
0.2 and 0.1. Momentum factor for the training was
0.9 and was later changed to 0.8 and finally 0.7. The
training was many times forced to restart again. This
was because of different reasons such as need for
more nodes in the network, over training or lack of
progress in the training. The training process took
around 10 days. In the final reported training, the
total number of elapsed epochs was 1382755
epochs. Total training time was 285899.0 seconds
(3 days, 7 hours, 24 minutes and 59 seconds). The
final MSE using normalised data on the output was
0.00027 and the final MSE using de-normalised (in
order to make it comparable with the PKA values)
data on the output was 6.04.
Table 1 compares two methods. In this table by
the term “the whole generalised points” the author
means the sampled points plus the generalised
interim points. Figure 4-a to Figure 4-c respectively
depict the output from the reference equation, PKA
and ANN. In Figure 3-a to Figure 3-c the
aforementioned dataset is generalised using
respectively the reference equation, PKA and ANN.
In analysing the data it is important to
understand that in any estimation the size of the
sampling step plays an important role in the final
resulting estimation error. If the speed of change in
the dataset is very high then a small sampling step is
required. This is some how the case in this example.
The ratio of the rise in the height of the graph is
sometimes too fast for the applied sampling step
size. Figure 5 depicts this concept.
As it can be seen in Figure 5-a, in the PKA
method the error values are higher at the centre of
the function domain than the rest of the points. In the
PKA method, the error is not evenly distributed in
the function domain. In this example in the PKA
method, although the maximum error seems to be
higher than the ANN method (Figure 5-b), most of
the points are well below the MSE value.
Figure 5-b depicts the distribution of the error
values in the ANN method. This figure shows that
the ANN method has a more evenly distributed error
pattern than the PKA method. However most of the
error values in this method are close to its MSE.
5.2 Analysis of the Result
Analysing the result, first the differences can be
divided into two categories training stage and the
execution stage. Later both of these categories will
be divided into smaller subsets such as difference in
execution speed, accuracy and some differences due
to the properties of the PKA methodology.
In the training stage as it is for the ANN, the
difference is huge. No training for the PKA in
comparison to the 3 days training for the ANN is a
clear advantage for the PKA methodology. Contrary
to the ANN that the required time for the training
can be different from one dataset to another, the time
required for the PKA to learn knowledge (the
training stage) is always the same. This property
plus ability to update the knowledge during the
execution are two of the main advantages of the
PKA. On the other hand PKA requires a large
memory to hold the samples that makes it a
memory-bound process, the fully connected and
biased ANN using the BP model is mostly a CPU-
bound process. Currently cost of the CPU time and
the Memory module are falling.
Sensitivity to the complexity in the training
pattern and sample population affects the training
stage for the ANN and in the supervised training it
might be necessary to reorganise the network and
restart the training. However, in the PKA none of
these problems will be encountered. After the
training for the ANN using extra sample points the
result has to be verified, but in the PKA due to the
consistency in the linear approximation this is not
necessary. In the reported experiment maximum
error in the PKA method was larger than the ANN
method. This difference is due to the large
: A numerical comparison between the PKA and the ANN. All the values are in millimetres
(MSE is
calculated using de-normalised data values). (*) Means that: Where output should be 91.98 it was 78.95
The ANN The PKA
Maximum
3
error for the sampled points. 1.3 0.0
MSE for the sampled points. 6.0 0.0
Maximum error (difference) for the interim points. 8.7 13.0 (*)
MSE for the interim points. 9.0 4.1
Maximum error (difference) for the whole generalised points. 8.7 13.0 (*)
MSE for the whole generalised points. 8.0 3.6
. Maximum error is a relati
ve error (between the reference
value and the generalised value.
ICEIS 2004 - ARTIFICIAL INTELLIGENCE AND DECISION SUPPORT SYSTEMS
162