clean dataset into new dataset. Then the new dataset
is inputted in the LightGBM model. In Section II, we
analyse the related work. In Section III, we introduce
the detail of our design. In Section IV, we introduce
raw data, data pre-processing and evaluation metrics.
In Section V, we give the results of the experiment
and compare performance of various methods.
Finally, in Section VI, we conclude this paper.
2 RELATED WORK
The concept of intrusion detection system from a
technical report submitted to the US Air Force by
Anderson (1980), which details what is intrusion
detection. The core of intrusion detection is to use
existing computer technology to analyse and detect
network traffic, and then take corresponding
measures according to certain rules. After more than
30 years of development, the intrusion detection
technology has achieved many exciting results
(Aljawarneh, Aldwairi, & Yassein, 2018). The existing
mainstream intrusion detection methods are based
on different machine learning algorithms and typical
neural network algorithms, such as support vector
machine (SVM) (Mahmood, 2018), Naive Bayes
Multiclass Classifier, DNN, CNN (Nguyen et al.,
2018).
One of the earliest work found in literature used
SVM with various Kernel functions and
regularization parameter C values for the design of
the model (Kim & Park, 2003). In its paper, Dong
Seong Kim and Jong Sou Park used 10% of the
KDD 99 training dataset for training and all test
dataset for testing. As expected, the training data
was divided into train set and validation set. Instead
of k-fold cross validation, they optimized the model
by repeatedly sampling training set randomly. It is
worth noting that the experimental results are
improved a lot by proper feature selection. The
experimental results proved that SVM achieved a
high accuracy in IDS.
Inspired by SVM model, Sungmoon Cheong
proposed new model named SVM-BTA to improve
SVM (Cheong, Oh, & Lee, 2004). He produced a novel
structure which includes SVM and decision tree.
This work built a binary decision tree which each
node was a SVM classifier. Besides of this,
Sungmoon Cheong produced a modified SOM to
convert multi-class tree into binary tree. As
expected, this method got took advantage of both the
efficient computation of the tree architecture and
high accuracy.
Deep learning has become more and more
popular since researchers were satisfied with data
and computation. Javaid et al. (2016) proposed a 2-
level deep learning structure. In the first level, there
is a self-taught learning model, or more specifically,
a Sparse Auto-Encoder for a more expressive feature
representation. Sparse Auto-Encoder can excavate
more relationships between features and labels. The
second level is a softmax regression classifier.
Moreover, Farahnakian and Heikkonen (2018)
recently have proposed a deep Auto-Encoder based
approach for IDS. There are five Auto-Encoder
stacked together in their model. Then they used a
supervised learning algorithm to avoid overfitting
and local optima. Finally, a softmax classifier is
added to get the results.
In this paper, we proposed an deep Auto-Encoder
and LightGBM based approach for improving IDS
performance. Our main contributions are as follows:
Firstly, an Auto-Encoder model is added to
discover efficient feature representations.
Secondly, our model concatenated the
intermediate result of the deep Auto-Encoder and the
clean dataset into new dataset so that we can avoid
the loss of feature transformation and feature
reduction. We employed a LightGBM model to
classify data.
Finally, the performance of our model is
evaluated by KDD-CUP’99 dataset. A series of
experiments is conducted to explore the performance
of different parameters.
3 DEEP AUTO-ENCODER BASED
LIGHTGBM MODEL
3.1 Auto-Encoder
An Auto-Encoder is a deep learning model which
uses a backpropagation algorithm to make the output
value equal to the input value. It first compresses the
input into a latent spatial representation and then
reconstructs the output by this characterization. An
Auto-Encoder includes two parts.
Encoder: This part compresses the input into a
latent representation, which can be represented by
the encoding function
.
Decoder: This part can reconstruct the input from
the latent representation, which can be represented
by the decoding function
.
Auto-Encoder is an unsupervised learning
algorithm whose structure is consistent with BP
neural network, but its objective function is different:
A Deep Auto-Encoder based LightGBM Approach for Network Intrusion Detection System
143