7 CONCLUSION AND FUTURE
WORK
In this paper, we have proposed a framework that has
investigated the viability of heterogeneous machine
learning model against GAN-based evasion attacks.
These kinds of attack are rarely detected by ML-
based malware detectors that are not backed by
defensive techniques. For instance, 87.6% of
adversarial malware examples constructed by
WGAN bypass the Logistic Regression classifier.
The presented empirical results have proven that
combining supervised and unsupervised models can
thwart roughly 42.8% of vanilla GAN-based attacks
and 19.2% of WGAN-based attacks.
There are several aspects that can be investigated
in the future. First, there is the impact of increasing
the number of clusters and classifiers in C3E-SL to
improve the malware detection rate. Second, there is
changing the squared loss in equation (1) for another
Bregman divergence. The third is by applying our
model in different application examples such as spam
and intrusion detection to definitely verify our claims
about the hardness of our proposed method. Fourth is
injecting adversarial examples in the training set and
retraining the hybrid model. We believe that
adversarial retraining will improve the heterogeneous
model’s robustness and encourage it to generalize
well to unseen data.
REFERENCES
Acharya, A., Hruschka, E. R., Ghosh, J., & Acharyya, S.
(2011). C 3 e: a framework for combining ensembles of
classifiers and clusterers. International Workshop on
Multiple Classifier Systems, 269–278.
Al-Dujaili, A., Huang, A., Hemberg, E., & O’Reilly, U.-M.
(2018). Adversarial deep learning for robust detection
of binary encoded malware. 2018 IEEE Security and
Privacy Workshops (SPW), 76–82.
Ao, X., Luo, P., Ma, X., Zhuang, F., He, Q., Shi, Z., & Shen,
Z. (2014). Combining supervised and unsupervised
models via unconstrained probabilistic embedding.
Information Sciences, 257, 101–114.
Arjovsky, M., Chintala, S., & Bottou, L. (2017).
Wasserstein generative adversarial networks.
International Conference on Machine Learning, 214–
223.
Bai, T., Zhao, J., Zhu, J., Han, S., Chen, J., Li, B., & Kot,
A. (2021). Ai-gan: Attack-inspired generation of
adversarial examples. 2021 IEEE International
Conference on Image Processing (ICIP), 2543–2547.
Banerjee, A., & Ghosh, J. (2008). Clustering with balancing
constraints. Constrained Clustering: Advances in
Algorithms, Theory, and Applications, 171–200.
Biggio, B., Fumera, G., & Roli, F. (2010). Multiple
classifier systems for robust classifier design in
adversarial environments. International Journal of
Machine Learning and Cybernetics, 1(1), 27–41.
Chakraborty, T. (2017). Ec3: Combining clustering and
classification for ensemble learning. 2017 IEEE
International Conference on Data Mining (ICDM),
781–786.
Chen, L., Hou, S., & Ye, Y. (2017). Securedroid:
Enhancing security of machine learning-based
detection against adversarial android malware attacks.
Proceedings of the 33rd Annual Computer Security
Applications Conference, 362–372.
Chinavle, D., Kolari, P., Oates, T., & Finin, T. (2009).
Ensembles in adversarial classification for spam.
Proceedings of the 18th ACM Conference on
Information and Knowledge Management, 2015–2018.
Coletta, L. F. S., Hruschka, E. R., Acharya, A., & Ghosh, J.
(2015). A differential evolution algorithm to optimise
the combination of classifier and cluster ensembles.
International Journal of Bio-Inspired Computation,
7(2), 111–124.
Gao, J., Liang, F., Fan, W., Sun, Y., & Han, J. (2011). A
graph-based consensus maximization approach for
combining multiple supervised and unsupervised
models. IEEE Transactions on Knowledge and Data
Engineering, 25(1), 15–28.
Gibert, D., Mateu, C., & Planes, J. (2020). The rise of
machine learning for detection and classification of
malware: Research developments, trends and
challenges. Journal of Network and Computer
Applications, 153, 102526.
Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B.,
Warde-Farley, D., Ozair, S., Courville, A., & Bengio,
Y. (2014). Generative adversarial nets. Advances in
Neural Information Processing Systems,
27.
Grosse, K., Manoharan, P., Papernot, N., Backes, M., &
McDaniel, P. (2017). On the (statistical) detection of
adversarial examples. ArXiv Preprint
ArXiv:1702.06280.
Grosse, K., Papernot, N., Manoharan, P., Backes, M., &
McDaniel, P. (2016). Adversarial perturbations against
deep neural networks for malware classification. ArXiv
Preprint ArXiv:1606.04435.
Hu, W., & Tan, Y. (2017). Generating adversarial malware
examples for black-box attacks based on GAN. ArXiv
Preprint ArXiv:1702.05983.
Hu, W., & Tan, Y. (2018). Black-box attacks against RNN
based malware detection algorithms. Workshops at the
Thirty-Second AAAI Conference on Artificial
Intelligence.
IBM Support. (2020). Clustering binary data with K-Means
(should be avoided). IBM.
Kong, D., & Yan, G. (2013). Discriminant malware
distance learning on structural information for
automated malware classification. Proceedings of the