characteristics such the nose, eyes, & lips to study and
identify the three emotions joyful, sad, & angry. They
used texture characteristics to train a traditional
neural network classification method, and the
resulting accuracy was 86% for HOG features and
65% for LBP data. In 2019, A. Bhavan et al. (Bhavan
et al. 2019) proposed a method for recognising
emotional states in people's voices by the extraction
of a small number of spectral features that have been
preprocessed (MFCCs and spectral centroids). This
method proposes using a bagged ensemble of SVMs
with a Gaussian kernel as the classification model.
Accuracy of 83.21 percent was found. Separately, the
discriminant temporal pyramid mapping method was
utilised to collect features in (Zhang et al. 2018) a
study using Mel spectrogram and the AlexNet deep
learning network. The gathered data showed that the
pre-trained deep learning model performed
effectively when processing emotional speech.
(Prasomphan 2015) used synthetic neural networks
and the EMO-five DB's emotions to suggest a new
approach to emotion detection using a spectrum
analyzer. Five out of the ten emotions had an 82%
success rate.
The Viola-Jones algorithm has the drawback that
it is difficult to detect emotions when the background
signal is complicated or when there are several noises
present, and it also has a low detection rate. These are
both limitations. Future work has to pay more
attention to a wider range of emotional types. The
system's ability to interpret the relevance of the
speech signal would be an added bonus.
8 CONCLUSION
The model that is being suggested exhibits both the
VJ and the HOG, with the VJ having obtained higher
accuracy values than the HOG as a result of its use.
The HOG has just a 88.65% accurate accuracy rating,
however the VJ has an accuracy rating that is 95.66%
more accurate than that of the HOG in an analysis of
human emotion via voice signal with an enhanced
accuracy rate.
REFERENCES
Adouani, Amal, Wiem Mimoun Ben Henia, and Zied
Lachiri. (2019). “Comparison of Haar-Like, HOG and
LBP Approaches for Face Detection in Video
Sequences.” In 2019 16th International Multi-
Conference on Systems, Signals Devices (SSD), 266–71.
Albornoz, Enrique M., Diego H. Milone, and Hugo L.
Rufiner. (2017). “Feature Extraction Based on Bio-
Inspired Model for Robust Emotion Recognition.” Soft
Computing 21 (17): 5145–58.
Alionte, Elena, and Corneliu Lazar. (2015). “A Practical
Implementation of Face Detection by Using Matlab
Cascade Object Detector.” In 2015 19th International
Conference on System Theory, Control and Computing
(ICSTCC), 785–90.
Bhavan, Anjali, Pankaj Chauhan, Hitkul, and Rajiv Ratn
Shah. (2019). “Bagged Support Vector Machines for
Emotion Recognition from Speech.” Knowledge-Based
Systems 184 (November): 104886.
Dellaert, F., T. Polzin, and A. Waibel. (1996).
“Recognizing Emotion in Speech.” In Proceeding of
Fourth International Conference on Spoken Language
Processing. ICSLP ’96, 3:1970–73 vol.3.
Gao, Yuanbo, Baobin Li, Ning Wang, and Tingshao Zhu.
(2017.) “Speech Emotion Recognition Using Local and
Global Features.” In Brain Informatics, 3–13. Springer
International Publishing.
G. Ramkumar, G. Anitha, P. Nirmala, S. Ramesh and M.
Tamilselvi, "An Effective Copyright Management
Principle using Intelligent Wavelet Transformation
based Water marking Scheme," 2022 International
Conference on Advances in Computing,
Communication and Applied Informatics (ACCAI),
Chennai, India, 2022, pp. 1-7, doi: 10.1109/
ACCAI53970.2022.9752516.
Han, Zhiyan, and Jian Wang. (2017). “Speech Emotion
Recognition Based on Gaussian Kernel Nonlinear
Proximal Support Vector Machine.” In 2017 Chinese
Automation Congress (CAC), 2513–16.
Jason, C. Andy, Sandeep Kumar, and Others. (2020). “An
Appraisal on Speech and Emotion Recognition
Technologies Based on Machine Learning.” Language
67: 68.
Julina, J. Kulandai Josephine, J. Kulandai Josephine Julina,
and T. Sree Sharmila. (2019). “Facial Emotion
Recognition in Videos Using HOG and LBP.” 2019 4th
International Conf. on Recent Trends on Electronics,
Information, Communication & Technology (RTEICT).
https://doi.org/10.1109/rteict46194.2019.9016766.
Kaliouby, R. El, R. El Kaliouby, and P. Robinson. n.d.
“Real-Time Inference of Complex Mental States from
Facial Expressions and Head Gestures.” 2004
Conference on Computer Vision and Pattern
Recognition Workshop. https://doi.org/10.1109/cvpr.
2004.427.
Kerkeni, Leila, Youssef Serrestou, Mohamed Mbarki,
Kosai Raoof, Mohamed Ali Mahjoub, and Catherine
Cleder. (2020). “Automatic Speech Emotion
Recognition Using Machine Learning.” In Social
Media and Machine Learning, edited by Alberto Cano.
London, England: IntechOpen.
Kirana, Kartika Candra, Slamet Wibawanto, and Heru
Wahyu Herwanto. (2018). “Emotion Recognition
Using Fisher Face-Based Viola-Jones Algorithm.” In
2018 5th International Conference on Electrical