attribute that most effectively divides its sample set
into a set of enriched sections in one class or another.
The criterion is the acquisition of normalized infor-
mation that results from the selection of attributes to
separate data. The attribute with the highest normal-
ized information acquisition was chosen to make a de-
cision (Korting, 2006).
In the C4.5 algorithm, the gain value is used to de-
termine which variable will be the node of a decision
tree (the variable with the highest gain).
Gain(A) = Entropi(S) − Σ
k
i=1
|S
i
|
|S|
xEntropi(S
i
) (1)
This process uses the parameter ”entropy” to mea-
sure the level of heterogeneity of the dataset, where
the greater the value of entropy, the greater the level
of heterogeneity of a data set.
Entropi(S) = Σ
k
j=1
− p jlog
2
p j (2)
Information : S = dataset (case) k = number of
partitions S pj = probability obtained from Sum (Yes)
divided by total cases
2.2 Neural Network Algorithm
Neural Network or better known as ANN (Artifi-
cial Neural Network) is a data mining method that is
widely used to do classification and prediction (Mc-
Culloch and Pitts, 1943). A Neural Network generally
consists of input, output, and hidden layer. And one of
the most popular algorithms used in learning of ANN
is Backpropagation (McClelland et al., 1986). ANN
or Artificial Neural Networks (ANN) is a parallel sys-
tem consisting of many, special non-linear processors,
known as neurons (Markopoulos et al., 2016). Like
the human brain, they can learn from its examples,
they can generalize and fault tolerance, and they can
respond intelligently to new triggers. Each neuron is a
primary processing unit, which receives one or more
external inputs and uses it to produce an output. The
whole system is considered parallel because many
neurons can implement calculations simultaneously.
The most important feature of neural networks is the
structure of the neurons that are connected because
they determine how the calculations are performed.
Starting from the source layer that receives input and
the output layer where the input layer is mapped, neu-
ral networks can have one or more hidden layers be-
tween. Neural networks, known as one or more hid-
den layers, are multilayer perceptron (MLP). These
networks, unlike simple perceptron, are capable of
linearly classifying inseparable patterns and can solve
complex problems. Examples of ANN with a single
hidden layer consisting of four units, six source units,
and two output units are shown in Figs.1
Figure 1: Single Hidden Layer Feed Forward ANN 6-4-2
(Markopoulos et al., 2016)
3 RESEARCH METHOD
3.1 Research Design
Figure 2: Research Design
Data is collected and selected from a collection of
operational data, then processed to obtain data with
good, complete, and consistent quality. The data that
has been pre-processed is determined as a dataset
which will then be used to build a classification model
with the Decision Tree C.45 and Neural Network al-
gorithm and at the same time be evaluated using the
Confusion Matrix method with several test parame-
ters. The classification model and evaluation process
are carried out using WEKA 3.8.3 data mining tools.
The results of the evaluation are then compared and
analysed so that the algorithm with the best model
is chosen based on the level of accuracy and classi-
fication modelling categories on the ROC (Receiver
Operating Characteristic) Curve, to be used in mak-
ing predictions of new data in the form of prospective
social assistance data.
3.2 Datasets
The data used in this study were recipients of so-
cial assistance data sourced from the Department of
Social Services of Gorontalo City in the database
of aid distribution totalling 123 records. Each data
record consists of 11 criteria with numeric and string
types, namely Trans Code, KKK, Name, Address,
Village, Sub-District, Education, Employment, Num-
ber of Children, Age, and Type of Assistance.
Comparison of Data Mining Classification Algorithm Performance for Data Prediction Type of Social Assistance Distribution
337