A Multi-Model Framework for Grading of Human Emotion Using CNN and Computer Vision

A Multi-Model Framework for Grading of Human Emotion Using CNN and Computer Vision

Praveen Kulkarni, Rajesh T. M.
Copyright: © 2022 |Pages: 21
DOI: 10.4018/IJCVIP.2022010102
Article PDF Download
Open access articles are freely available for download

Abstract

Emotion analysis is an area which is been widely used in the forensic crime detection domain, a mentoring device for depressed students, psychologically affected patient treatment. The current system helps only in identifying the emotions but not in identifying the level of emotions like whether the individual is truly happy/sad or pretending to be happy /sad. In this proposed work a novel methodology has been introduced. We have rebuilt the Traditional Local Binary Pattern (LBP) feature operator to image the expression and combine the abstract characteristics of facial expression learned from the neural network of deep convolution with the modified features of the texture of the LBP facial expression in the full connection layer. These extracted features have been subjected as input for CNN Alex Net to classify the level of emotions. The results obtained in this phase are used in the confusion matrix for analysis of grading of emotions like Grade-1, Grade-2, and Grade-3 obtained an accuracy of 87.58% in the comparative analysis.
Article Preview
Top

Introduction

People communicate with one another mostly through speech, but they also use bodily gestures to emphasize specific points in their speech and to express feelings. Facial articulations, which are an important aspect of communication, are one of the most important ways people express their feelings. Despite the fact that nothing is said vocally, the signals we transmit and receive via nonverbal contact contain a lot of information. Facial articulations allow for nonverbal communication and play an important role in between-home relationships. In typical human-machine interfaces, programmed facial articulation responses can play an important role. It could be employed in both social research and clinical practice. Despite the fact that people can recognize facial joints nearly instantly, reliable machine recognition remains a challenge. In prior work, a throw analysis was performed using well-known and widely utilized algorithms (Kulkarni et al., 2020), and many studies focused just on emotion recognition rather than grading. Previous work on grading emotion using a combination of LBP and the KNN algorithm (Kulkarni et al., 2021) yielded findings that were not promising, with an accuracy of only 79 percent. This has prompted us to develop the methods proposed in this paper. Decision trees and fuzzy algorithms were used to train the model, which took into account 19 aspects of the human face. Using machine learning methods, the emotion detection accuracy was not up to par.

Despite significant progress in the field of emotion analysis, in areas such as mentoring for depressed students, criminal evaluation systems, psychologically affected patient treatment, and musical therapy for desperate and mentally disabled patients, the assessment of the classification of the person's emotion expressed on his or her face is currently lacking. Grading human emotion can be useful in ensuring that the person remaining in front of the camera isn't just a two-dimensional representation, but is actually expressing a genuine emotion.

The proposed feature points were chosen through a thorough literature review of different feature points that contribute to happy and sad emotions. Emotions are graded by comparing the number of features that match the threshold value. The threshold value is determined by comparing the number of extracted features to the number of matched features. Using a happy emotion as an example, we have subcategorized happy emotions into grade 1 happy, grade 2 happy, and grade 3 happy emotions. Whereas a grade of 1 implies that the person is trying to be happy, and the features matched for this emotion among 19 features are about 6 to 8. That is, if 6 to 8 features are matched, we consider it grade 1. Similarly, grade 2 emotion is nothing, although the person is normally joyful. The threshold value has been set at 12 to 14 features. If more than 12 features match, we consider the person to be normal joyful. Similarly, more than 15 elements should match for the grade 3 feeling, which indicates that the person is extremely delighted. This allows us to discern the actual emotions in a person's face. The same procedure is followed with sad feelings. We proposed a novel methodology for calculating the performance of each model in distinguishing facial expression and rating the level of emotion in happy and sad emotions using LBP inputs by applying CNN and varying its depths. We used the MMI research database, the Japanese Female Facial Expression Database (JAFFE), and some real-time expression movies in our studies. We'll look at how the highlights are separated and changed for decision tree computations in this lesson. We'll look at component extraction calculations and approaches from a variety of sources.

Complete Article List

Search this Journal:
Reset
Volume 14: 1 Issue (2024): Forthcoming, Available for Pre-Order
Volume 13: 1 Issue (2023)
Volume 12: 4 Issues (2022): 1 Released, 3 Forthcoming
Volume 11: 4 Issues (2021)
Volume 10: 4 Issues (2020)
Volume 9: 4 Issues (2019)
Volume 8: 4 Issues (2018)
Volume 7: 4 Issues (2017)
Volume 6: 2 Issues (2016)
Volume 5: 2 Issues (2015)
Volume 4: 2 Issues (2014)
Volume 3: 4 Issues (2013)
Volume 2: 4 Issues (2012)
Volume 1: 4 Issues (2011)
View Complete Journal Contents Listing