Show simple item record

dc.contributor.authorIslam, Md. Wahidul
dc.date.accessioned2025-06-25T06:55:02Z
dc.date.available2025-06-25T06:55:02Z
dc.date.issued2025-05-19
dc.identifier.urihttp://suspace.su.edu.bd/handle/123456789/1581
dc.description.abstractEmotion recognition has garnered significant attention in fields such as mental health, human computer interaction, and personalized services. This research explores a multimodal approach to emotion recognition by integrating facial expression analysis and speech prosody to achieve a more accurate and context-sensitive understanding of human emotions. A distinctive aspect of this study is the creation of a custom video dataset designed specifically for facial expression recognition, which captures a wide range of emotional states under various real-world conditions. In parallel, speech emotion detection is performed using publicly available audio datasets, which analyze features such as pitch, tone, and rhythm to discern emotions expressed vocally. The facial expression recognition is based on Convolutional Neural Networks (CNNs), which extract visual features from the video data, while the emotional cues in speech are analyzed using Long Short-Term Memory (LSTM) networks. By combining these modalities, this research addresses the limitations commonly faced by unimodal systems, such as the challenges posed by noisy environments or occluded faces. The findings demonstrate that the integration of facial and auditory data significantly improves emotion classification accuracy, particularly in real-time applications. This research advances the field of affective computing by highlighting the complementary strengths of visual and auditory emotion cues and offers practical implications for applications in customer service, virtual assistants, and mental health diagnostics.en_US
dc.language.isoen_USen_US
dc.publisherSonargaon University (SU)en_US
dc.relation.ispartofseries;CSE-2502332
dc.subjectFacial Landmarks and Speech Prosodyen_US
dc.titleTowards Real-Time Emotion Analytics: Integrating Facial Landmarks and Speech Prosodyen_US
dc.typeThesisen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record