Message Boards Message Boards

[WSSA16] Facial Emotion Recognition

The big image on the left side is the Input, and on the right side is the Output

The big image on the left side is the Input, and on the right side is the Output

The aim of my project for Wolfram Summer School in Armenia is to recognize emotions of a person, using a neural network with input of a picture of the face. The mentioned network is based on LeNet network. In this paper, I propose an approach to "Facial emotion recognition" in a fully automated way. Though, only four databases were used during the research, manual operations were applied to generate different and practical data as an addition to the ready ones.

The big image on the left side is the Input, and on the right side is the Output

Collecting Data

Four different databases were used during this research:

  1. Cohn-Kanade AU-Coded Expression Database (+)
  2. Japanese Female Facial Expression (+)
  3. FaceTales (+)
  4. Yale face database (+)

Every database has it's own emotions, so there is a need to filter emotions we are interested in. We came up with 7 emotion plus one Neutral face. Here is the share of each emotion in all 4211 labeled pictures that we had:

  1. Happy(844)
  2. Sad (412)
  3. Angry (659)
  4. Disgusted (592)
  5. Afraid (408)
  6. Ashamed (81)
  7. Surprised (775)
  8. Neutral (440)

Samples

Some emotions are taken from different datasets.

Tolerance problem

In the real world there is no guarantee that one stands strictly straight in front of the camera, so some processes were applied to data to match more to the real situations. As a result, all images were rotated at 5, 10 and 15 in both clockwise and counter clockwise directions (appropriate cut-offs were performed on newly generated artiacts) and even been reflected from left to right, in order to generate more data. The manual generation increased overall number of the images to be used in training by 14 times.

The functions below describe the process:

reflectData := Join[KeyMap[(ImageReflect[#, Left -> Right]) &, #], #] &;
rotateDatas[x_] := Join[
                      KeyMap[
                         Function[u, ImageRotate[u, -5Degree, 128, Background->White]]
                         , 
                         x
                       ]
                       ,
                       KeyMap[
                         Function[u, ImageRotate[u, -10Degree, 128, Background->White]]
                         ,
                         x
                       ]
                       ,
                       KeyMap[
                         Function[u, ImageRotate[u, -15Degree, 128, Background->White]]
                         ,
                         x
                       ]
                       ,
                       KeyMap[
                         Function[u, ImageRotate[u, +5Degree, 128, Background->White]]
                         ,
                         x
                       ]
                       ,
                       KeyMap[
                         Function[u, ImageRotate[u, +10Degree, 128, Background->White]]
                         ,
                         x
                       ]
                       ,
                       KeyMap[
                         Function[u, ImageRotate[u, +15Degree, 128, Background->White]]
                         ,
                         x
                       ]
                      ,
                      x
                   ]

Preprocess

All pictures have to be rescaled (128x128) so that they have identical size of face and same color properties (grayscale), this is done to perform the biggest face detection (the default FindFaces function is used), and omit the margin around it.

The big image on the left side is the Input, and on the right side is the Output

Neural network

The used network was based on the LeNet neural network that was used to recognize digits. Of course the encoder and decoder layers are changed with respect to the requirements of the present problem. As a result the final network layers look like:

LeNet Neural Network

This code below describes the creation of neural network:

classes = {"Neutral", "Angry", "Happy", "Sad", "Afraid", "Disgusted", "Ashamed", "Surprise"};
decoder = NetDecoder["Class", classes];
encoder = NetEncoder["Image", {128,128}, "Grayscale"];
loss = CrossEntropyLossLayer["Target" -> NetEncoder["Class", classes]];
lenet = NetChain[{
                  ConvolutionLayer[20, {5,5}], 
                  ElementwiseLayer[Ramp], 
                  PoolingLayer[{2, 2}, {2, 2}],
                  ConvolutionLayer[50, {5, 5}],
                  ElementwiseLayer[Ramp],
                  PoolingLayer[{2, 2} ,{2, 2}],
                  FlattenLayer[],
                  DotPlusLayer[500],
                  ElementwiseLayer[Ramp],
                  DotPlusLayer[8],
                  SoftmaxLayer[]
                 },
                 "Output" -> decoder,
                 "Input" -> encoder
           ];

The big image on the left side is the Input, and on the right side is the Output

Training

Data was split into three parts: Training data (54954 Images), Validation data (3000 Images), Test data (1000 Images). It took 14 hours on "Intel Core i5 4210U CPU, Windows 10 x64" to train for 10 rounds, and the accuracy on test data was 98.5%, but as there is a great chance for a test data to be a rotated or a reflected version of a image in train data or vice versa, which means that this number is not pretty reliable, though the result during dynamic test using the laptop camera is still good.

Block[{tempData},
   {testData, tempData} = TakeDrop[RandomSample[complexData], 1000];
   {validationData, trainData} = TakeDrop[RandomSample[tempData], 3000]
 ];
 classComplex = NetTrain[classGray, Normal[trainData], loss, ValidationSet -> Normal[validationData]];

The big image on the left side is the Input, and on the right side is the Output

Issues and Further Improvement

One of elements in building this application is a Wolfram Language built-in function: FindFace. This function is still in improvement phase and has several issues, from showing non-face examples and having problem with rotated pictures. Due to this problem there is one additional step added to the user-interface. If some parts of the image are false recognized as faces, one can omit them by clicking on the image, but this problem still doesn’t really solve the dynamic recognition issues.

Even in dynamic mode, all processes are mapped into many single frames; so some wrong results may be shown when picture is changing rapidly. It is possible to improve the functionality by saving the probabilities of every emotions for several frames, and estimate emotion of a period, by using some simple formulas.

As the experiment showed that it was more wise approach to split images to training data, validation data and test data, and only then start the manual transformations. Of course the result may not vary much, but the accuracy may be more reliable.

All training data were pictures of random people, standing exactly in camera direction. Having even more databases for the training may solve this problem in the future. The emotions, the neural network came up recognizing, were the ones present in the databases. It was possible to have some more emotions like Contempt or Tired. Still more databases may help.

Iman Nazari And Mikayel Egibyan

Conclusion

Applying machine learning methods to emotion recognition was successful. Our results appear to be at the state of art in Facial Emotion Recognition. The network is designed to perform in under-control situation (stable picture, enough light, standing straight and into camera direction). More data and maybe more complex network are required to improve the functionality, so it can be applied to more practical/real-life situations. Because of the selected network architecture, the network could be trained on a low-level representation of data that had minimal processing (as apposed to elaborated feature extraction). Because of redundant nature of data and because of the constraints imposed on the network, the learning time was relatively short considering the size of the training set (one gigabyte of grayscaled 128x128 images).

One can test the application, by downloading this network file (Mirror) and putting it in $HomeDirectory (usually Documents folder for Windows users) and then only evaluate last two parts of the code in the attached notebook.

Happy to supply more information on request.

Some more examples

Iman Nazari is testing his project

The big image on the left side is the Input, and on the right side is the Output

The big image on the left side is the Input, and on the right side is the Output

The big image on the left side is the Input, and on the right side is the Output

The big image on the left side is the Input, and on the right side is the Output

The big image on the left side is the Input, and on the right side is the Output

The big image on the left side is the Input, and on the right side is the Output

(With some special thanks to Brad Pitt, Natalie Portman, Marlon Brando, Edward Norton, Shrek and others for all special things they have done.)

POSTED BY: Iman Nazari
4 Replies

Very interesting project!!!

Kristina

POSTED BY: Kristina Akopyan

Thank you Kristina ^_^

POSTED BY: Iman Nazari

enter image description here - you earned "Featured Contributor" badge, congratulations !

This is a great post and it has been selected for the curated Staff Picks group. Your profile is now distinguished by a "Featured Contributor" badge and displayed on the "Featured Contributor" board.

POSTED BY: Moderation Team

This is a wonderful illustration of ML with facial recognition. My version is V11.1. I ran your breast cancer analysis model successfully after some minor revision on the original codes. But this model gave some errors in my version. Can you also post these mx files, so that I may run over again? complexData.mx, dataSet1.mx, dataSet2.mx, dataSet3.mx, dataSet4.mx

Thanks so much Peter

POSTED BY: Teck Boon Lim
Reply to this discussion
Community posts can be styled and formatted using the Markdown syntax.
Reply Preview
Attachments
Remove
or Discard

Group Abstract Group Abstract