Partial face coverings such as sunglasses and face masks unintentionally obscure facial expressions, causing a loss of accuracy when humans and computer systems attempt to categorise emotion. With the rise of soft computing techniques interacting with humans, it is important to know not just their accuracy, but also the confusion errors being made-do humans make less random/damaging errors than soft computing? We analyzed the impact of sunglasses and different face masks on the ability to categorize emotional facial expressions in humans and computer systems. Computer systems, represented by VGG19, ResNet50, and InceptionV3 deep learning algorithms, and humans assessed images of people with varying emotional facial expressions and with four different types of coverings, i.e. unmasked, with a mask covering the lower face, a partial mask with transparent mouth window, and with sunglasses. The first contribution of this work is that computer systems were found to be better classifiers (98.48%) than humans (82.72%) for faces without covering (>15% difference). This difference is due to the significantly lower accuracy in categorizing anger, disgust, and fear expressions by humans (p ′ s < .001). However, the most novel aspect of the work is identifying how soft computing systems make different mistakes 2 to humans on the same data. Humans mainly confuse unclear expressions as neutral emotion, which minimizes affective effects. Conversely, soft techniques often confuse unclear expressions as other emotion categories, which could lead to opposing decisions being made, e.g. a robot categorizing a fearful user as happy. Importantly, the variation in the misclassification can be adjusted by variations in the balance of categories in the training set.Published Version: https://doi.org/10.1016/j.asoc.2022.109701 Reproducible Code Capsule: https://codeocean.com/capsule/2412241/tree/v1
Emotions are considered to convey much meaning in communication. Hence, artificial methods for emotion categorization are being developed to meet the increasing demand to introduce intelligent systems, such as robots, into shared workspaces. Deep learning algorithms have demonstrated limited competency in categorizing images from posed datasets with the main features of the face being visible. However, the use of sunglasses and face masks are common in our daily lives, especially with the outbreak of communicable diseases such as the recent coronavirus. Anecdotally, partial coverings of the face reduce the effectiveness of human communication, so would this have hampering effects on computer vision, and if so, would different emotion categories be affected equally? Here, we analyze the performance of emotion classification systems when faces are partially covered with simulated sunglasses and face masks. Deep neural networks consider all pixels in an image as equally important unlike the neuroscientific findings on how humans recognize emotions. Hence, we propose a method that considers different constituent parts (e.g. mouth, eyes, and jaw) separately, giving more attention to relevant (uncovered) regions of the face. The method is compared with three standard, partial coverings-based and attention-based methods. We found that face coverings worsen emotion categorization by up to 74% for the state-of-the-art methods, whereby emotion categories are affected differently by different coverings, e.g. clear mouth coverings have little effect on categorizing happiness, but sadness is affected badly. The proposed method (on average 60.43%) has significantly improved the performance over the standard deep learning (< 46% on average), partial coverings-based (< 47% on average), as well as the attention-based (< 51% on average) methods for both the CK+, KDEF, and RAF-DB datasets when faces were partially covered with sunglasses or different face masks.