In Fig.4 and Table 1, the experimental results can
clearly show the relevant results of each model.
Figure 4: Comparison results of three models (Picture credit:
Original)
.
Table 1: Results display of three models.
Accuracy
Recall
F1 score
SVM 0.65 0.65 0.64
Naive Bayes 0.41 0.29 0.27
Random Forest 0.67 0.67 0.67
5 CONCLUSION
This study aims to analyze the white wine dataset and
compare the performance of three models: SVM,
Naive Bayes, and Random Forest on this dataset.
Determine which model best fits the dataset by
comparing the accuracy, F1 score, and recall of the
three models. With an accuracy of almost 67% and
other index values. The researchers determined that
the random forest model exhibited superior
performance. The SVM model has a little inferior
level of accuracy compared to the random forest
model, but it also shows excellent analytical and
predictive capabilities in the white wine data set.
Considering the possibility that the data may not be
linear and the linear kernel function results in
suboptimal accuracy, this research tried to change the
kernel function to a Gaussian kernel function, which
fits the data form better. Hence, the accuracy and
other results were about 65%. The worst-performing
model is the Naive Bayes model. After multiple
optimizations of this model, its accuracy is only 41%.
The reason for this result is that the uneven sample
distribution of the data set affects the model's
accuracy. Since the Naive Bayes method is a white-
box classification method and often needs to be more
accurate for this dataset.
This study fills a research gap regarding white wine
data and the fit between the three models we
employed. This study gives readers a deeper
understanding of the correlation between accuracy
across datasets and different models.
Due to the author's limited academic level and
insufficient optimization capabilities for models such
as Naive Bayes, the expressiveness of the model has
yet to reach a high level. In the future, the author will
continue to delve into machine learning and learn
algorithms for optimizing machine models. Improve
technical level.
REFERENCES
Z. Dong, X. Guo, S. Rajana, and B. Chen, “Understanding
21st Century Bordeaux Wines from Wine Reviews
Using Naïve Bayes Classifier,” Beverages, 2020, pp. 5
B. Shaw, A. K.Suman, and B. Chakraborty, “Wine Quality
Analysis Using Machine Learning,Emerging
Technology in Modelling and Graphics”, Emerging
Technology in Modelling and Graphics, 2019.
M. Nandan, H. Raj Gupta and M. Mondal, "Building a
Classification Model based on Feature Engineering
for the Prediction of Wine Quality by Employing
Supervised Machine Learning and Ensemble
Learning Techniques," International Conference on
Computer, Electrical & Communication Engineering
(ICCECE), 2023, pp. 1-7.
T. H. -Y. Chiu, C. Wu, and C. -H. Chen, “A Generalized
Wine Quality Prediction Framework by Evolutionary
Algorithms,” International Journal of Interactive
Multimedia and Artificial Intelligence, 2021, p. 60.
P. Shruthi, "Wine Quality Prediction Using Data Mining,"
Environment, Computing & Communication
Engineering (ICATIECE), 2019, pp. 23-26.
G. Hu, T. Xi, F. Mohammed and H. Miao, "Classification
of wine quality with imbalanced data," IEEE
International Conference on Industrial Technology,
2016, pp. 1712-1217.
K. Manisha, P. Richa, J. Mayurika and K. Manish,
“Analysis of white wine using machine learning
algorithms,” Materials Today: Proceedings, 2021, pp.
11087-11093.
K. R. Dahal, J.N. Dahal, H. Banjade, S. Gaire, “Prediction
of Wine Quality Using Machine Learning Algo-
rithms,” Open Journal of Statistics, 2021, pp. 278-289.
T. -T. Dai and Y. -S. Dong, “Introduction of SVM Related
Theory and Its Application Research,” 3rd
International Conference on Advanced Electronic
Materials, 2020, pp. 230-233.
F. -J. Yang, “An Implementation of Naive Bayes Classifier,”
International Conference on Computational Science
and Computational Intelligence (CSCI), 2018, pp.
301-306.