Deng, L. and Platt, J. (2014). Ensemble deep learning for
speech recognition.
Eigen, D., Ranzato, M., and Sutskever, I. (2013). Learning
factored representations in a deep mixture of experts.
arXiv preprint arXiv:1312.4314.
Fei-Fei, L., Fergus, R., and Perona, P. (2007). Learning gen-
erative visual models from few training examples: An
incremental bayesian approach tested on 101 object
categories. Computer vision and Image understand-
ing, 106(1):59–70.
Ge, W. and Yu, Y. (2017). Borrowing treasures from
the wealthy: Deep transfer learning through selective
joint fine-tuning. arXiv preprint arXiv:1702.08690.
Girshick, R., Donahue, J., Darrell, T., and Malik, J. (2014).
Rich feature hierarchies for accurate object detec-
tion and semantic segmentation. In Proceedings of
the IEEE conference on computer vision and pattern
recognition, pages 580–587.
Griffin, G., Holub, A., and Perona, P. (2007). Caltech-256
object category dataset.
He, K., Zhang, X., Ren, S., and Sun, J. (2015). Delv-
ing deep into rectifiers: Surpassing human-level per-
formance on imagenet classification. In Proceedings
of the IEEE international conference on computer vi-
sion, pages 1026–1034.
He, K., Zhang, X., Ren, S., and Sun, J. (2016). Deep resid-
ual learning for image recognition. In Proceedings of
the IEEE conference on computer vision and pattern
recognition, pages 770–778.
Ju, C., Bibaut, A., and van der Laan, M. J. (2017). The rela-
tive performance of ensemble methods with deep con-
volutional neural networks for image classification.
arXiv preprint arXiv:1704.01664.
Kim, Y.-D., Jang, T., Han, B., and Choi, S. (2016). Learn-
ing to select pre-trained deep representations with
bayesian evidence framework. In Proceedings of the
IEEE Conference on Computer Vision and Pattern
Recognition, pages 5318–5326.
Maaten, L. v. d. and Hinton, G. (2008). Visualizing data
using t-sne. Journal of Machine Learning Research,
9(Nov):2579–2605.
Masoudnia, S. and Ebrahimpour, R. (2014). Mixture of ex-
perts: a literature survey. Artificial Intelligence Re-
view, pages 1–19.
Nilsback, M.-E. and Zisserman, A. (2006). A visual vo-
cabulary for flower classification. In Computer Vi-
sion and Pattern Recognition, 2006 IEEE Computer
Society Conference on, volume 2, pages 1447–1454.
IEEE.
Nilsback, M.-E. and Zisserman, A. (2008). Automated
flower classification over a large number of classes.
In Proceedings of the Indian Conference on Computer
Vision, Graphics and Image Processing.
Oquab, M., Bottou, L., Laptev, I., and Sivic, J. (2014).
Learning and transferring mid-level image represen-
tations using convolutional neural networks. In Pro-
ceedings of the IEEE conference on computer vision
and pattern recognition, pages 1717–1724.
Parkhi, O. M., Vedaldi, A., Zisserman, A., and Jawahar,
C. V. (2012). Cats and dogs. In IEEE Conference
on Computer Vision and Pattern Recognition.
Pennington, J., Socher, R., and Manning, C. (2014). Glove:
Global vectors for word representation. In Proceed-
ings of the 2014 conference on empirical methods in
natural language processing (EMNLP), pages 1532–
1543.
Qiu, X., Zhang, L., Ren, Y., Suganthan, P. N., and Ama-
ratunga, G. (2014). Ensemble deep learning for re-
gression and time series forecasting. In Compu-
tational Intelligence in Ensemble Learning (CIEL),
2014 IEEE Symposium on, pages 1–6. IEEE.
Schmidhuber, J. (2015). Deep learning in neural networks:
An overview. Neural networks, 61:85–117.
Sharif Razavian, A., Azizpour, H., Sullivan, J., and Carls-
son, S. (2014). Cnn features off-the-shelf: an as-
tounding baseline for recognition. In Proceedings of
the IEEE conference on computer vision and pattern
recognition workshops, pages 806–813.
Shazeer, N., Mirhoseini, A., Maziarz, K., Davis, A., Le,
Q., Hinton, G., and Dean, J. (2017). Outrageously
large neural networks: The sparsely-gated mixture-of-
experts layer. arXiv preprint arXiv:1701.06538.
Simonyan, K. and Zisserman, A. (2014). Very deep con-
volutional networks for large-scale image recognition.
arXiv preprint arXiv:1409.1556.
Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., and Wo-
jna, Z. (2016). Rethinking the inception architecture
for computer vision. In Proceedings of the IEEE Con-
ference on Computer Vision and Pattern Recognition,
pages 2818–2826.
Theagarajan, R., Pala, F., and Bhanu, B. (2017). Eden: En-
semble of deep networks for vehicle classification. In
Computer Vision and Pattern Recognition Workshops
(CVPRW), 2017 IEEE Conference on, pages 906–913.
IEEE.
Wah, C., Branson, S., Welinder, P., Perona, P., and Be-
longie, S. (2011). The Caltech-UCSD Birds-200-2011
Dataset. Technical report.
Yao, B., Jiang, X., Khosla, A., Lin, A. L., Guibas, L.,
and Fei-Fei, L. (2011). Human action recognition by
learning bases of action attributes and parts. In Com-
puter Vision (ICCV), 2011 IEEE International Con-
ference on, pages 1331–1338. IEEE.
Yosinski, J., Clune, J., Bengio, Y., and Lipson, H. (2014).
How transferable are features in deep neural net-
works? In Advances in neural information processing
systems, pages 3320–3328.
Zhao, T., Yu, J., Kuang, Z., Zhang, W., and Fan, J. (2017).
Deep mixture of diverse experts for large-scale visual
recognition. arXiv preprint arXiv:1706.07901.
ICPRAM 2019 - 8th International Conference on Pattern Recognition Applications and Methods
144