stagewise inclusion of regressors in the overall learn-
ing procedure for an ART model (see Section 3) is a
main reason for irrelevant regressors to not pose much
of a problem for our approach.
7 CONCLUSIONS AND FUTURE
WORK
We have presented a spectral splitting method that
improves segmentation in regime-switching time se-
ries, and use the segmentation in the process of cre-
ating a switching regression tree for the regimes in
the model. We have exemplified our approach by
an extension to the ART models–the only type of
models that to our knowledge applies switching trees
in regime-switching models. However, the spectral
splitting method does not rely on specific paramet-
ric assumptions, which allows for our approach to
be readily generalized to a wide range of regime-
switching time-series models that go beyond the limit
of the Gaussian error assumption in the ART mod-
els. Our proposed method is very parsimonious in
the number of split predicates it proposes for candi-
date split nodes in the tree. It only proposes a single
oblique split candidate for the regressors in the model
and possibly a few very targeted time-split candidates,
thus keeping computational complexity of the over-
all algorithm under control. Both types of split can-
didates rely on a spectral clustering, where different
views–phase and trace–on the time-series data give
rise to the two different types of candidates.
Finally, we have given experimental evidence that
our approach, when applied for the exemplifying ART
models, dramatically improves predictive accuracy
over the current approach. Regarding time splits,
we hope to be able to find–and are actively looking
for–real-world time series for future experiments that
will allow us to factor out and compare the quality of
our spectral time-split proposer method to the current
ART approach.
The focus in this paper has been on learning
regime-switching time-series models that will easily
lend themselves to explanatory analysis and interpre-
tation. In future experiments we also plan to evaluate
the potential tradeoff in modularity, interpretability,
and computational efficiency with forecast precision
for our simple learning approach compared to more
complicated approaches that integrates learning of
soft regime switching and the local regimes the mod-
els, such as the learning of Markov-switching (e.g.
Hamilton, 1989, 1990) and gated expert (e.g. Wa-
terhouse and Robinson, 1995; Weigend et al., 1995)
models.
REFERENCES
Bishop, C. M. (1995). Neural Networks for Pattern Recog-
nition. Oxford University Press, Oxford.
Breiman, L., Friedman, J. H., Olshen, R. A., and Stone,
C. J. (1984). Classification and Regression Trees.
Wadsworth International Group, Belmont, California.
Brodley, C. E. and Utgoff, P. E. (1995). Multivariate deci-
sion trees. Machine Learning, 19(1):45–77.
Chaudhuri, P., Huang, M. C., Loh, W.-Y., and Yao, R.
(1994). Piecewise polynomial regression trees. Sta-
tistica Sinica, 4:143–167.
Chickering, D., Meek, C., and Rounthwaite, R. (2001).
Efficient determination of dynamic split points in a
decision tree. In Proc. of the 2001 IEEE Inter-
national Conference on Data Mining, pages 91–98.
IEEE Computer Society.
DeGroot, M. (1970). Optimal Statistical Decisions.
McGraw-Hill, New York.
Dobra, A. and Gehrke, J. (2002). Secret: A scalable lin-
ear regression tree algorithm. In Proc. of the 8th
ACM SIGKDD International Conference on Knowl-
edge Discovery and Data Mining, pages 481–487.
ACM Press.
Donath, W. E. and Hoffman, A. J. (1973). Lower bounds for
the partitioning of graphs. IBM Journal of Research
and Development, 17:420–425.
Fiedler, M. (1973). Algebraic connectivity of graphs.
Czechoslovak Mathematical Journal, 23:298–305.
Gama, J. (1997). Oblique linear tree. In Proc. of the Second
International Symposium on Intelligent Data Analy-
sis, pages 187–198.
Hamilton, J. (1994). Time Series Analysis. Princeton Uni-
versity Press, Princeton, New Jersey.
Hamilton, J. D. (1989). A new approach to the economic
analysis of nonstationary time series and the business
cycle. Econometrica, 57(2):357–84.
Hamilton, J. D. (1990). Analysis of time series subject to
changes in regime. Journal of Econometrics, 45:39–
70.
Iyengar, V. S. (1999). Hot: Heuristics for oblique trees. In
Proc. of the 11th IEEE International Conference on
Tools with Artificial Intelligence, pages 91–98, Wash-
ington, DC. IEEE Computer Society.
Jordan, M. I. and Jacobs, R. A. (1994). Hierarchical mix-
tures of experts and the EM algorithm. Neural Com-
putation, 6:181–214.
Li, K. C., Lue, H. H., and Chen, C. H. (2000). Interactive
tree-structured regression via principal Hessian direc-
tions. Journal of the American Statistical Association,
95:547–560.
Mandelbrot, B. (1966). Forecasts of future prices, unbiased
markets, and martingale models. Journal of Business,
39:242–255.
Mandelbrot, B. B. and Hudson, R. L. (2006). The
(mis)Behavior of Markets: A Fractal View of Risk,
Ruin And Reward. Perseus Books Group, New York.
Meek, C., Chickering, D. M., and Heckerman, D. (2002).
Autoregressive tree models for time-series analysis. In
Proc. of the Second International SIAM Conference
on Data Mining, pages 229–244. SIAM.
ICPRAM 2012 - International Conference on Pattern Recognition Applications and Methods
124