Real-time emotion detection by quantitative facial motion analysis

Jordan R Saadon, Fan Yang, Ryan Burgert, Selma Mohammad, Theresa Gammel, Michael Sepe, Miriam Rafailovich, Charles B Mikell, Pawel Polak, Sima Mofakham, Jordan R Saadon, Fan Yang, Ryan Burgert, Selma Mohammad, Theresa Gammel, Michael Sepe, Miriam Rafailovich, Charles B Mikell, Pawel Polak, Sima Mofakham

Abstract

Background: Research into mood and emotion has often depended on slow and subjective self-report, highlighting a need for rapid, accurate, and objective assessment tools.

Methods: To address this gap, we developed a method using digital image speckle correlation (DISC), which tracks subtle changes in facial expressions invisible to the naked eye, to assess emotions in real-time. We presented ten participants with visual stimuli triggering neutral, happy, and sad emotions and quantified their associated facial responses via detailed DISC analysis.

Results: We identified key alterations in facial expression (facial maps) that reliably signal changes in mood state across all individuals based on these data. Furthermore, principal component analysis of these facial maps identified regions associated with happy and sad emotions. Compared with commercial deep learning solutions that use individual images to detect facial expressions and classify emotions, such as Amazon Rekognition, our DISC-based classifiers utilize frame-to-frame changes. Our data show that DISC-based classifiers deliver substantially better predictions, and they are inherently free of racial or gender bias.

Limitations: Our sample size was limited, and participants were aware their faces were recorded on video. Despite this, our results remained consistent across individuals.

Conclusions: We demonstrate that DISC-based facial analysis can be used to reliably identify an individual's emotion and may provide a robust and economic modality for real-time, noninvasive clinical monitoring in the future.

Conflict of interest statement

The authors have declared that no competing interests exist.

Copyright: © 2023 Saadon et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.

Figures

Fig 1. Heatmaps derived from the results…
Fig 1. Heatmaps derived from the results of DISC analysis of representative frames for each emotion in a single participant.
The top three panels are the original images, whereas the bottom three are the same images with superimposed heatmaps showing magnitude of movement from the baseline (neutral) frames. Units are in pixels.
Fig 2. Heatmaps showing magnitude of facial…
Fig 2. Heatmaps showing magnitude of facial movement in response to happy and sad images for each participant.
Heatmaps were generated from the averaged DISC–calculated displacement across all happy and sad frames for that individual. Numbers represent each participant in the study. Participant 1 declined to have their face included in the publication of this data.
Fig 3. Average magnitude of facial movement…
Fig 3. Average magnitude of facial movement in response to happy and sad images across all participants.
Units are in pixels.
Fig 4. Similarity matrices of DISC results…
Fig 4. Similarity matrices of DISC results from each frame of each participant’s video.
Matrices are numbered 1–10, corresponding to each participant’s ID. The matrix for Participant 1 is enlarged to show matrix organization. A value of 1.0 signifies 100% similarity between two frames and a value of 0 signifies the absence of any similarity.
Fig 5. Plot of first and second…
Fig 5. Plot of first and second principal components of DISC–processed displacement across all participants.
Red squares represent frames from the happy image–viewing period, whereas blue diamonds represent frames from the sad image–viewing period. Black circles signify neutral frames. Large shapes indicate the averages for each participant. Gray lines serve to connect the averages for happy and sad for individual participants.
Fig 6
Fig 6
Temporal changes of average facial movement for each individual (ghosted lines) and across all participants (prominent line) over the duration of (A) happy and (B) sad image presentation. Dashed vertical lines represent the presentation of a new image.
Fig 7. Confusion matrices of the two…
Fig 7. Confusion matrices of the two methods.
(A)–(C): Different classifiers using the DISC method. (D): Amazon Rekognition. On the y–axis are the true emotion labels for the images; on the x–axis are the predicted emotion labels. The numbers in the plots indicate percentages of predicted labels for each true label. Correct predictions are along the diagonal of the matrix, as they are indicated by the boxes at the intersection of the same true and predicted labels on each axis. The Amazon Rekognition software contains seven emotion labels by default and this cannot be modified by the user. Abbreviations: DISC (digital image speckle correlation), SLR (sparse logistic regression), MLP (multi–layer perceptron), CNN (convolutional neural network).

References

    1. Kroenke K, Spitzer RL, Williams JBW. The PHQ-9. J Gen Intern Med. 2001;16: 606–613.
    1. Nease DE Jr, Aikens JE, Klinkman MS, Kroenke K, Sen A. Toward a more comprehensive assessment of depression remission: the Remission Evaluation and Mood Inventory Tool (REMIT). Gen Hosp Psychiatry. 2011;33: 279–286. doi: 10.1016/j.genhosppsych.2011.03.002
    1. Shiffman S, Stone AA, Hufford MR. Ecological momentary assessment. Annu Rev Clin Psychol. 2008;4: 1–32. doi: 10.1146/annurev.clinpsy.3.022806.091415
    1. Hamilton M. A rating scale for depression. J Neurol Neurosurg Psychiatry. 1960;23: 56–62. doi: 10.1136/jnnp.23.1.56
    1. Montgomery SA, Åsberg M. A new depression scale designed to be sensitive to change. Br J Psychiatry. 1979;134: 382–389. doi: 10.1192/bjp.134.4.382
    1. Young RC, Biggs JT, Ziegler VE, Meyer DA. A rating scale for mania: reliability, validity and sensitivity. Br J Psychiatry. 1978;133: 429–435. doi: 10.1192/bjp.133.5.429
    1. Bushey MA, Kroenke K, Baye F, Lourens S. Assessing depression improvement with the remission evaluation and mood inventory tool (REMIT). Gen Hosp Psychiatry. 2019;60: 44–49. doi: 10.1016/j.genhosppsych.2019.07.007
    1. Ebner-Priemer UW, Trull TJ. Ecological momentary assessment of mood disorders and mood dysregulation. Psychol Assess. 2009;21: 463–475. doi: 10.1037/a0017075
    1. Furukawa TA. Assessment of mood: guides for clinicians. J Psychosom Res. 2010;68: 581–589. doi: 10.1016/j.jpsychores.2009.05.003
    1. Wenze SJ, Miller IW. Use of ecological momentary assessment in mood disorders research. Clin Psychol Rev. 2010;30: 794–804. doi: 10.1016/j.cpr.2010.06.007
    1. Wilhelm FH, Pfaltz MC, Grossman P. Continuous electronic data capture of physiology, behavior and experience in real life: towards ecological momentary assessment of emotion. Interact Comput. 2006;18: 171–186.
    1. Nakanishi R, Imai-Matsumura K. Facial skin temperature decreases in infants with joyful expression. Infant Behav Dev. 2008;31: 137–144. doi: 10.1016/j.infbeh.2007.09.001
    1. Benitez-Quiroz CF, Srinivasan R, Martinez AM. Facial color is an efficient mechanism to visually transmit emotion. Proc Natl Acad Sci U S A. 2018;115: 3581–3586. doi: 10.1073/pnas.1716084115
    1. Agrafioti F, Hatzinakos D, Anderson AK. ECG pattern analysis for emotion detection. IEEE Trans Affect Comput. 2012;3: 102–115.
    1. Lin Y-P, Wang C-H, Wu T-L, Jeng S-K, Chen J-H. EEG-based emotion recognition in music listening: A comparison of schemes for multiclass support vector machine. 2009 IEEE International Conference on Acoustics, Speech and Signal Processing. IEEE; 2009. doi: 10.1109/icassp.2009.4959627
    1. Wagner J, Kim J, Andre E. From physiological signals to emotions: Implementing and comparing selected methods for feature extraction and classification. 2005 IEEE International Conference on Multimedia and Expo. IEEE; 2005. doi: 10.1109/icme.2005.1521579
    1. Cheng B, Liu G. Emotion recognition from surface EMG signal using wavelet transform and neural network. 2008 2nd International Conference on Bioinformatics and Biomedical Engineering. IEEE; 2008. doi: 10.1109/icbbe.2008.670
    1. Hänsel K, Alomainy A, Haddadi H. Large scale mood and stress self-assessments on a smartwatch. Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing: Adjunct. New York, NY, USA: ACM; 2016. doi: 10.1145/2968219.2968305
    1. Lee K, Hong H. Designing for self-tracking of emotion and experience with tangible modality. Proceedings of the 2017 Conference on Designing Interactive Systems. New York, NY, USA: ACM; 2017. doi: 10.1145/3064663.3064697
    1. James W. The principles of psychology, Vol I. New York: Henry Holt and Co; 1890.
    1. Hjortsjö C-H. Man’s Face and Mimic Language. Lund, Sweden: Studentlitteratur; 1969.
    1. Ekman P. Emotion in the human face: Guide-lines for research and an integration of findings. Pergamon; 1972.
    1. Ekman P, Friesen WV. Facial Action Coding System: A Technique for the Measurement of Facial Movement. 1978.
    1. Ekman P, Friesen WV. Unmasking the Face. Old Tappan, NJ: Prentice Hall; 1975.
    1. Hamedi M, Salleh S-H, Astaraki M, Noor AM. EMG-based facial gesture recognition through versatile elliptic basis function neural network. Biomed Eng Online. 2013;12: 73. doi: 10.1186/1475-925X-12-73
    1. Ree A, Morrison I, Olausson H, Sailer U, Heilig M, Mayo LM. Using facial electromyography to assess facial muscle reactions to experienced and observed affective touch in humans. J Vis Exp. 2019. doi: 10.3791/59228
    1. Shreve M, Godavarthy S, Goldgof D, Sarkar S. Macro- and micro-expression spotting in long videos using spatio-temporal strain. Face and Gesture 2011. IEEE; 2011. doi: 10.1109/fg.2011.5771451
    1. Ekman P, Friesen WV. Nonverbal leakage and clues to deception. Psychiatry. 1969;32: 88–106. doi: 10.1080/00332747.1969.11023575
    1. Porter S, ten Brinke L. Reading between the lies: identifying concealed and falsified emotions in universal facial expressions. Psychol Sci. 2008;19: 508–514. doi: 10.1111/j.1467-9280.2008.02116.x
    1. Kollias D, Nicolaou MA, Kotsia I, Zhao G, Zafeiriou S. Recognition of affect in the wild using deep neural networks. 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). IEEE; 2017. doi: 10.1109/cvprw.2017.247
    1. Yang H, Ciftci U, Yin L. Facial expression recognition by DE-expression residue learning. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. IEEE; 2018. doi: 10.1109/cvpr.2018.00231
    1. Xie S, Hu H. Facial expression recognition using hierarchical features with deep comprehensive multipatches aggregation convolutional neural networks. IEEE Trans Multimedia. 2019;21: 211–220.
    1. D’mello SK, Kory J. A review and meta-analysis of Multimodal affect detection systems. ACM Comput Surv. 2015;47: 1–36.
    1. Poria S, Cambria E, Bajpai R, Hussain A. A review of affective computing: From unimodal analysis to multimodal fusion. Inf Fusion. 2017;37: 98–125.
    1. Kaushik H, Kumar T, Bhalla K. iSecureHome: A deep fusion framework for surveillance of smart homes using real-time emotion recognition. Appl Soft Comput. 2022;122: 108788.
    1. Peters WH, Ranson WF. Digital Imaging Techniques In Experimental Stress Analysis. Opt Eng. 1982;21. doi: 10.1117/12.7972925
    1. Guan E, Smilow S, Rafailovich M, Sokolov J. Determining the mechanical properties of rat skin with digital image speckle correlation. Dermatology. 2004;208: 112–119. doi: 10.1159/000076483
    1. Pamudurthy S, Guan E, Mueller K, Rafailovich M. Dynamic approach for face recognition using digital image skin correlation. Lecture Notes in Computer Science. Berlin, Heidelberg: Springer Berlin Heidelberg; 2005. pp. 1010–1018.
    1. Staloff IA, Guan E, Katz S, Rafailovitch M, Sokolov A, Sokolov S. An in vivo study of the mechanical properties of facial skin and influence of aging using digital image speckle correlation. Skin Res Technol. 2008;14: 127–134. doi: 10.1111/j.1600-0846.2007.00266.x
    1. Bhatnagar D, Conkling N, Rafailovich M, Phillips BT, Bui DT, Khan SU, et al.. An in vivo analysis of the effect and duration of treatment with botulinum toxin type A using digital image speckle correlation. Skin Res Technol. 2013;19: 220–229. doi: 10.1111/srt.12010
    1. Verma R, Klein G, Xu Y, Rafailovich M, Gilbert Fernandez JJ, Khan SU, et al.. Digital image speckle correlation to optimize botulinum toxin type A injection: A prospective, randomized, crossover trial. Plast Reconstr Surg. 2019;143: 1614–1618. doi: 10.1097/PRS.0000000000005637
    1. Bhatnagar D, Fiore SM, Rafailovich M, Davis RP. An Analysis of Facial Nerve Function in Patients with Vestibular Schwannomas Using Digital Image Speckle Correlation. Journal of Neuroscience and NeuroEngi. 2014;3: 62–71.
    1. Bradley MM, Lang PJ. Measuring emotion: the Self-Assessment Manikin and the Semantic Differential. J Behav Ther Exp Psychiatry. 1994;25: 49–59. doi: 10.1016/0005-7916(94)90063-9
    1. Lang PJ, Bradley MM, Cuthbert BN. International Affective Picture System (IAPS): Affective Ratings of Pictures and Instruction Manual. Gainesville, FL: University Press of Florida; 2008.
    1. Pearson K. LIII. On lines and planes of closest fit to systems of points in space. Lond Edinb Dublin Philos Mag J Sci. 1901;2: 559–572.
    1. Chouinard B, Scott K, Cusack R. Using automatic face analysis to score infant behaviour from video collected online. Infant Behav Dev. 2019;54: 1–12. doi: 10.1016/j.infbeh.2018.11.004
    1. Kim Y, Kwon S, Heun Song S. Multiclass sparse logistic regression for classification of multiple cancer types using gene expression data. Comput Stat Data Anal. 2006;51: 1643–1655.
    1. Pedregosa F, Varoquaux G, Gramfort A, Michel V, Thirion B, Grisel O, et al.. Scikit-learn: Machine Learning in Python. J Mach Learn Res. 2011;12: 2825–2830.
    1. LeCun Y, Boser B, Denker JS, Henderson D, Howard RE, Hubbard W, et al.. Backpropagation applied to handwritten zip code recognition. Neural Comput. 1989;1: 541–551.
    1. Tran D, Bourdev L, Fergus R, Torresani L, Paluri M. Learning Spatiotemporal Features with 3D Convolutional Networks. 2015 IEEE International Conference on Computer Vision (ICCV). IEEE; 2015. doi: 10.1109/iccv.2015.510
    1. Turk MA, Pentland AP. Face recognition using eigenfaces. Proceedings 1991 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. IEEE Comput. Sco. Press; 1991. doi: 10.1109/cvpr.1991.139758
    1. Kim B-K, Roh J, Dong S-Y, Lee S-Y. Hierarchical committee of deep convolutional neural networks for robust facial expression recognition. J Multimodal User Interfaces. 2016;10: 173–189.
    1. Darwin C. The expression of the emotions in man and animals. London: John Murray; 1872.
    1. Haggard EA, Isaacs KS. Micromomentary facial expressions as indicators of ego mechanisms in psychotherapy. Methods of Research in Psychotherapy. Boston, MA: Springer US; 1966. pp. 154–165.
    1. Wang S-J, Chen H-L, Yan W-J, Chen Y-H, Fu X. Face recognition and micro-expression recognition based on discriminant tensor subspace analysis plus extreme learning machine. Neural Process Lett. 2014;39: 25–43.
    1. Merrett F. Reflections on the Hawthorne effect. Educ Psychol (Lond). 2006;26: 143–146.

Source: PubMed

3
구독하다