have the capability to learn the underlying
relationships between the inputs and outputs of a
process, without needing the explicit knowledge of
how these variables are related.
Recently, numerous applications of AANs to
estimate air temperature data have been presented,
e.g. in areas with sparse network of meteorological
stations (Snell et al., 2000); (Chronopoulos et al.,
2008) for the prediction of hourly (Tasadduq,
Rehman and Bubshait, 2002), daily (Dombayc and
Golcu, 2009) and year-round air temperature (Smith
et al., 2009) or room temperature (Mustafaraj et al.,
2011) as well as for simulating the Heat Island
(Mihalakakou et al., 2002).
In this work first we briefly present the
theoretical background of ANN methodologies
applicable to the field of air temperature time series
and spatial modeling. Next, we focus on
implementation issues and on evaluating the
accuracy of the aforementioned methodologies using
a set of metrics in the case of a specific region with
complex terrain at Chania, Crete Island, Greece. A
number of alternative Feed-forward ANN topologies
are applied in order to assess the spatial and time
series air temperature prediction capabilities in
different time horizons.
2 ANN PREDICTION MODELING
Artificial Neurons are Process Element (PE) that
attempt to simulate in a simplistic way the structure
and function of the real physical biological neurons.
A PE in its basic form can be modelled as non-liner
element that first sums its weighted inputs x
1
, x
2
, x
3
,
...x
n
(coming either from original data, or from the
output of other neurons in a neural network) and
then passes the result through an activation function
Ψ (or transfer function) according to the formula:
n
i
jjiii
wxy
1
(1)
where y
j
is the output of the artificial neuron, θ
j
is an
external threshold (or bias value) and w
ji
are the
weight of the respective input x
i
which determines
the strength of the connection from the previous
PE’s to the corresponding input of the current PE.
Depending on the application, various non-linear or
linear activation functions Ψ have been introduced
(Fausett, 1994); (Bishop, 1995) like the: signum
function (or hard limiter), sigmoid limiter, quadratic
function, saturation limiter, absolute value function,
Gaussian and hyperbolic tangent functions. Artificial
Neural Networks (ANN) are signal or information
processing systems constituted by an assembly of a
large number of simple Processing Elements, as they
have been described above. The PE of a ANN are
interconnected by direct links called connections and
cooperate to perform a Parallel Distributed
Processing in order to solve a specific computational
task, such as pattern classification, function
approximation, clustering (or categorization),
prediction (or forecasting or estimation),
optimization and control. One the main strength of
ANNs is their capability to adapt themselves by
modifying the interaction between their PE. An-
other important feature of ANNs is their ability to
automatically learn from a given set of
representative examples.
The architectures of ANNs can be classified into
two main topologies: a) Feed-forward multilayer
networks (FFANN) in which feedback connections
are not allowed and b) Feedback recurrent networks
(FBANN) in which loops exist. FFANNs are
characterized mainly as static and memory-less
systems that usually produce a response to an input
quickly (Jain et al., 1996). Most FFANNs can be
trained using a wide variety of efficient conventional
numerical methods. FBANNs are dynamic systems.
In some of them, each time an input is presented, the
ANN must iterate for a potentially long time before
it produces a response. Usually, they are more
difficult to train FBANNs compared to FFANNs.
FFANNs have been found to be very effective
and powerful in prediction, forecasting or estimation
problems (Zhang et al., 1998). Multilayer
perceptrons (MLPs) and radial basis function (RBF)
topologies are the two most commonly-used types of
FFANNs. Essentially, their main difference is the
way in which the hidden PEs combine values
coming from preceding layers: MLPs use inner
products, while RBF constitutes a multidimentional
function which depends on the distance
cxr
between the input vector x and the center c (where
denotes a vector norm) (Powell, 1987). As a
consequence, the training approaches between MLPs
and RBF based FFANN is not the same, although
most training methods for MLPs can also be applied
to RBF ANNs. In RBF FFANNs the connections of
the hidden layer are not weighted and the hidden
nodes are PEs with a RBF, however the output layer
performs simple weighted summation of its inputs,
like in the case of MLPs. One simple approach to
approximate a nonlinear function is to represent it as
a linear combination of a number of fixed nonlinear
ICPRAM2013-InternationalConferenceonPatternRecognitionApplicationsandMethods
670