Table 6: Values with k-NN as classifier.
Dataset Approach Accuracy Feature
Reduction
Senti140
No FS 50.07% -
GA 63.00% 49.97%
Binary Bat 59.69% 50.02%
Binary Wolf 56.36 11.67%
Cornell
No FS 58.28% -
Movie
GA 55.96 % 50.23%
Reviews
Binary Bat 55.20% 0.2%
Binary Wolf 53.98% 8.4%
with max-depth being used to limit the number of fea-
tures used. However, the computation of Information
Gain that has to be done for every feature would sig-
nificantly speed up with a lesser number of features.
Genetic Algorithm again turns out to be the most effi-
cient in terms of feature reduction.
When using k-NN as the classifier, the results
seem to be mixed. While more than significant ac-
curacy gains over the baseline have been obtained on
the Senti140 Dataset, we also observe worsened per-
formance over the baseline for the Cornell Movie Re-
views Dataset. This is probably because of the ran-
dom nature of k-NN as it simply performs a majority
voting within k-nearest neighbors and does not actu-
ally pick up any patterns. This shows that for some
classifiers, any sort of feature selection will not guar-
antee an increase in the accuracy.
5 CONCLUSION
In this paper, we have compared the performances
of meta-heuristic and evolutionary feature selection
methods to the problem of sentiment analysis using
various classifiers on two different domains of tweets
and movie reviews. While we can see, that meth-
ods such as Random Forest that have in-built param-
eters to limit the features used, do not gain any suf-
ficient improvement in accuracy, other methods such
as SVM and k-NN can have gain in accuracy upto
25%. While the performance of Binary Bat and Ge-
netic Algorithm was similar in terms of accuracy gain,
the performance of Binary Grey Wolf Algorithm was
consistently lower than these two. Also, the percent-
age decrease in the number of features is another im-
portant ground to consider while making a choice.
Genetic Algorithm was observed to be the most effi-
cient in terms of feature reduction percentage. More-
over, there is a difference in the number of hyperpa-
rameters that need to be tuned to make each algorithm
work optimally, with Binary Grey Wolf Algorithm be-
ing the easiest to tune. Hence, a multitude of factors
need to be considered when selecting a method for
feature selection. The results reported in this paper
can be used as a guidance for extended work in dif-
ferent domains.
REFERENCES
Abbasi, A., Chen, H., and Salem, A. (2008). Sentiment
analysis in multiple languages: Feature selection for
opinion classification in web forums. ACM Transac-
tions on Information Systems (TOIS), 26(3):12.
Aghdam, M. H., Ghasem-Aghaee, N., and Basiri, M. E.
(2009). Text feature selection using ant colony
optimization. Expert systems with applications,
36(3):6843–6853.
Ahmad, S. R., Bakar, A. A., and Yaakub, M. R. (2015).
Metaheuristic algorithms for feature selection in senti-
ment analysis. In 2015 Science and Information Con-
ference (SAI), pages 222–226. IEEE.
Eirinaki, M., Pisal, S., and Singh, J. (2012). Feature-based
opinion mining and ranking. Journal of Computer and
System Sciences, 78(4):1175–1184.
Emary, E., Zawbaa, H. M., and Hassanien, A. E. (2016).
Binary grey wolf optimization approaches for feature
selection. Neurocomputing, 172:371–381.
Fong, S., Yang, X.-S., and Deb, S. (2013). Swarm search for
feature selection in classification. In 2013 IEEE 16th
International Conference on Computational Science
and Engineering, pages 902–909. IEEE.
Forman, G. (2003). An extensive empirical study of fea-
ture selection metrics for text classification. Journal
of machine learning research, 3(Mar):1289–1305.
Gandomi, A. H., Yang, X.-S., and Alavi, A. H. (2013).
Cuckoo search algorithm: a metaheuristic approach
to solve structural optimization problems. Engineer-
ing with computers, 29(1):17–35.
Ghosh, S., Biswas, S., Sarkar, D., and Sarkar, P. P. (2010).
Mining frequent itemsets using genetic algorithm.
arXiv preprint arXiv:1011.0328.
Go, A., Bhayani, R., and Huang, L. (2009). Twitter senti-
ment classification using distant supervision. CS224N
Project Report, Stanford, 1(12).
Griffin, D. R., Webster, F. A., and Michael, C. R. (1960).
The echolocation of flying insects by bats. Animal
behaviour, 8(3-4):141–154.
Hiemstra, D. (2000). A probabilistic justification for using
tf× idf term weighting in information retrieval. Inter-
national Journal on Digital Libraries, 3(2):131–139.
Karabulut, E.,
¨
Ozel, S., and Ibrikci, T. (2012a). Compara-
tive study on the effect of feature selection on classifi-
cation accuracy. Procedia Technology, 1:323 –327.
Karabulut, E. M.,
¨
Ozel, S. A., and Ibrikci, T. (2012b). A
comparative study on the effect of feature selection on
classification accuracy. Procedia Technology, 1:323–
327.
Medhat, W., Hassan, A., and Korashy, H. (2014). Sentiment
analysis algorithms and applications: A survey. Ain
Shams engineering journal, 5(4):1093–1113.
A Comparative Study of Evolutionary Methods for Feature Selection in Sentiment Analysis
137