1. Preprocessing
2. Wavelet decomposition
3. Multi resolution forecasting
4. Combination
2.1 Preprocessing
In this level the input signal (time series) is being
denoised and down sampled.
2.2 Wavelet Decomposition
As long as “db.4” (Daubechies, 1992) doesn’t have
sharp edges we have used it as a desirable wavelet
because better adoption of neural networks (MLPs
and RBFs) was observed. To avoid border
distortion, symmetric padding (Matlab’s toolbox
default DWT mode) of the time series signal was
applied. A six level decomposition followed by
single level reconstruction was applied to the input
time series. In this point we have seven
reconstructed signals (named XREC) that can be
forecasted separately. Since for better performance
of the neural networks we need the training set to be
between -1 and +1 (Hagan, 1996) therefore we
assume a maximum value M for our input time
series and divide all of the reconstructed signals to
M. Also we assume a minimum value m, the need
for this minimum value is described in combination
level. It is clear that, M and m change for different
time series.
2.3 Multi-resolution Forecasting
As described above, forecasting models are different
for each resolution. The resolutions are divided to
two separated parts, first the four lower resolutions,
second the three higher resolutions.
For first group Focused time lagged feed forward
network which is a nonlinear filter is used (Figure 2)
(
Haykin, 1999). As seen in (Figure 2) the prediction
network is made up of a short term memory
followed by a static neural network. We have used a
tapped delay line of length 12 as the short term
memory. The input SREC(t) (reconstructed signals)
is fed in to the tapped delay line. These delayed
signals are then inputs to the static neural network
(MLP or RBF) which is trained to predict the next
value of the input signal SREC(t+1). So for every
level of resolution there is a separate forecasting
block, for the lowest resolution a 2-layer ordinary
RBF network was used as static neural network,
which is fast in training in comparison to MLPs, for
others more complex static neural networks(MLPs
mainly) were used. For the second group LRN
which is a dynamical network is used instead of
static neural network. This makes the forecasting
model more adaptive with the frequency conditions
of the high resolutions. This structure is like FTLFF
diagrammatically.
Figure 1: Layer recurrent network structure.
Figure 2: Focused time lagged feed forward network
(Haykin, 1999).
2.4 Combination
In this level all the forecasted signals are being
added together and multiplied by M; the result will
be compared to m and M (Matlab’s satlin function)
for acceptable output values.
3 EXPERIMENTS AND RESULTS
Two data sets were used for comparing the MRF
with MLP and LRN. 1. Annual sunspot average 2.
Normalized intensity data recorded from a Far-
Infrared-Laser in a chaotic state (Table1 & Table2).
Table1: NMSE for one step prediction of annual sun spot
average. Results obtained predicting 200 points from the
data set.
Table1 MLP LRN MRF
NMSE 0.5182 0.5068 0.2917
Table2: NMSE for one step prediction of Far-Infrared-
Laser in a chaotic state.
Table2 MLP LRN MRF
NMSE 0.3130 0.2058 0.1127
)( n
)( ny
)(nd
A MULTI RESOLUTION FORECASTING METHOD FOR SHORT LENGTH TIME SERIES DATA USING NEURAL
NETWORKS
533