loading
Papers Papers/2022 Papers Papers/2022

Research.Publish.Connect.

Paper

Paper Unlock

Authors: Atitaya Yakaew 1 ; Matthew N. Dailey 1 and Teeradaj Racharak 2

Affiliations: 1 Department of Information and Communication Technologies, Asian Institute of Technology, Klong Luang, Pathumthani, Thailand ; 2 School of Information Science, Japan Advanced Institute of Science and Technology, Ishikawa, Japan

Keyword(s): Deep Learning for Multimodal Real-Time Analysis, Emotion Recognition, Video Processing and Analysis, Lightweight Deep Convolutional Neural Networks, Sentiment Classification.

Abstract: Real-time sentiment analysis on video streams involves classifying a subject’s emotional expressions over time based on visual and/or audio information in the data stream. Sentiment can be analyzed using various modalities such as speech, mouth motion, and facial expression. This paper proposes a deep learning approach based on multiple modalities in which extracted features of an audiovisual data stream are fused in real time for sentiment classification. The proposed system comprises four small deep neural network models that analyze visual features and audio features concurrently. We fuse the visual and audio sentiment features into a single stream and accumulate evidence over time using an exponentially-weighted moving average to make a final prediction. Our work provides a promising solution to the problem of building real-time sentiment analysis systems that have constrained software or hardware capabilities. Experiments on the Ryerson audio-video database of emotional speech ( RAVDESS) show that deep audiovisual feature fusion yields substantial improvements over analysis of either single modality. We obtain an accuracy of 90.74%, which is better than baselines of 11.11% – 31.48% on a challenging test dataset. (More)

CC BY-NC-ND 4.0

Sign In Guest: Register as new SciTePress user now for free.

Sign In SciTePress user: please login.

PDF ImageMy Papers

You are not signed in, therefore limits apply to your IP address 18.221.187.121

In the current month:
Recent papers: 100 available of 100 total
2+ years older papers: 200 available of 200 total

Paper citation in several formats:
Yakaew, A.; Dailey, M. and Racharak, T. (2021). Multimodal Sentiment Analysis on Video Streams using Lightweight Deep Neural Networks. In Proceedings of the 10th International Conference on Pattern Recognition Applications and Methods - ICPRAM; ISBN 978-989-758-486-2; ISSN 2184-4313, SciTePress, pages 442-451. DOI: 10.5220/0010304404420451

@conference{icpram21,
author={Atitaya Yakaew. and Matthew N. Dailey. and Teeradaj Racharak.},
title={Multimodal Sentiment Analysis on Video Streams using Lightweight Deep Neural Networks},
booktitle={Proceedings of the 10th International Conference on Pattern Recognition Applications and Methods - ICPRAM},
year={2021},
pages={442-451},
publisher={SciTePress},
organization={INSTICC},
doi={10.5220/0010304404420451},
isbn={978-989-758-486-2},
issn={2184-4313},
}

TY - CONF

JO - Proceedings of the 10th International Conference on Pattern Recognition Applications and Methods - ICPRAM
TI - Multimodal Sentiment Analysis on Video Streams using Lightweight Deep Neural Networks
SN - 978-989-758-486-2
IS - 2184-4313
AU - Yakaew, A.
AU - Dailey, M.
AU - Racharak, T.
PY - 2021
SP - 442
EP - 451
DO - 10.5220/0010304404420451
PB - SciTePress