Learning via social awareness: improving sketch representations with facial feedback

02/13/2018 ∙ by Natasha Jaques, et al. ∙ Google MIT 0

In the quest towards general artificial intelligence (AI), researchers have explored developing loss functions that act as intrinsic motivators in the absence of external rewards. This paper argues that such research has overlooked an important and useful intrinsic motivator: social interaction. We posit that making an AI agent aware of implicit social feedback from humans can allow for faster learning of more generalizable and useful representations, and could potentially impact AI safety. We collect social feedback in the form of facial expression reactions to samples from Sketch RNN, an LSTM-based variational autoencoder (VAE) designed to produce sketch drawings. We use a Latent Constraints GAN (LC-GAN) to learn from the facial feedback of a small group of viewers, and then show in an independent evaluation with 76 users that this model produced sketches that lead to significantly more positive facial expressions. Thus, we establish that implicit social feedback can improve the output of a deep learning model.



There are no comments yet.


page 1

page 2

page 3

page 4

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

1 Introduction

Despite the recent rapid and compelling progress in ML and DL, modern AI is still remarkably far from approximating the intelligence of even simple animals. A notable deficit is the degree of explicit supervision required in order to learn, either through labeled samples or well-defined external rewards such as points in a game. The limited scope of such supervision will not enable the development of a generally intelligent AI.

For this reason, some researchers have focused on intrinsic motivators, inherent drives that cause the agent to learn representations that are useful across a variety of tasks and environments. Examples include curiosity (a drive for novelty) (Pathak et al., 2017), and empowerment (a drive for the ability to manipulate the environment) (Capdepuy et al., 2007). However, so far this research has overlooked an important intrinsic motivator for humans: the drive for positive social interactions.

We argue that making an AI agent intrinsically motivated to obtain a positive social reaction from humans in its environment is an important new research direction. Specifically, the agent should be able to recognize implicit feedback from humans in the form of facial expressions, body language, or tone in voice and text, and optimize for actions that appear to please humans as measured through these signals.

The representations learned by such an agent are more likely to capture dimensions of the task that are relevant to human satisfaction. This has meaningful implications for questions of AI safety; an AI agent motivated by satisfaction expressed by humans will be less likely to take actions against human interest. Such an agent will also be better suited to perform tasks which already involve AI. Imagine if a home assistant could sense when a user responds with an angry or frustrated tone and this acted as a negative incentive, training the algorithm not to repeat the action that led to the user’s frustration? Rather than requiring the user to manually train the device, it could learn quickly through passive sensing of the user’s emotional state, leading to a more immediately satisfying experience for the user. Finally, some machine learning problems — including the one under investigation in this paper — cannot be solved without human feedback; when the objective function is human aesthetic preference, it cannot be approximated without human input.

Social awareness may be a key component of AGI. There is substantial evidence that emotion recognition, which is critical for empathy and successful social interaction, plays an influential role in cognitive development in humans (Kujawa et al., 2014). According to Social Learning Theory (Bandura and Walters, 1977), observing the attitudes and behaviors of others is a central component of how humans learn both intelligent behavior and how to adapt to new situations. It has been argued that social learning is responsible for the rapid cultural evolution of the human species (van Schaik and Burkart, 2011). Given the importance of cultural evolution to humans’ technological success, endowing a deep learning agent with the ability to perceive and benefit from this socially exchanged cultural knowledge could allow it to rapidly develop more generalizable knowledge representations.

In this work we demonstrate the utility of learning through implicit social feedback via an experiment in which samples generated by a deep learning model are presented to people, and their facial expression response is detected. The model is Sketch RNN (Ha and Eck, 2017), an LSTM-based VAE with a Mixture Density Network output, designed to produce sketch drawings. Using a newly developed technique known as Latent Constraints (Engel et al., 2017)

, we train a Generative Adversarial Network (GAN) to produce VAE embedding vectors that, when decoded by Sketch RNN, are more likely to produce drawings that lead to positive facial expressions such as smiling. In a rigorous, double-blind evaluation, we show that samples from the social feedback model generate statistically significantly better affective responses than the prior, and are consistently rated as more preferred by human judges. Thus, this experiment is a first step in demonstrating that deep learning models are able to improve in quality as a result of learning from implicit social feedback.

2 Related work

Many affective computing papers have addressed how to automatically detect facial expressions (e.g. Senechal et al. (2015)). A comprehensive review of this work is out of scope for this paper. We instead build on this work by assuming that an accurate facial expression detector is already available, and asking what can be learned using this facial feedback.

Previous work has attempted to train ML and DL models to approximate human preferences. For example, Knox and Stone (2009)

ask users to press a button to teach a reinforcement learning (RL) model to play

Mountain Car

, and model human reaction latencies as a Gamma distribution in order to distribute the reward appropriately over past time steps. A more recent work attempts to train a deep learning model from human preferences, by first training an approximator of human button presses using supervised learning, and then using this to train an RL model

(Christiano et al., 2017). However, both of these approaches require the human to provide explicit supervision by manually entering feedback. In contrast, our approach enables learning from implicit social cues that can be obtained ubiquitously, through awareness of the non-verbal reactions people naturally provide. Essentially, we obtain human-in-the-loop training without additional human effort.

The closest work to our own of which we are aware is an approach that used valence and engagement, detected via facial expressions, as a reward function in a Q-learning framework (Gordon et al., 2016). The goal of the project was to allow an intelligent tutoring system to adapt its behavior so that children would remain engaged while using the system. While this is an excellent example of learning from implicit social feedback, the goals of this paper are quite distinct from our own. We believe we are the first authors to use implicit social feedback to improve a generative deep learning model. Our model attempts to learn to improve its ability to produce creative content by observing the implicit responses it receives from human judges. This process could be considered analogous to a human artist fine-tuning her work after she observes critics’ reactions.

3 Methods

3.1 Study design

To gather social feedback, we focused on facial expression recognition, since this is currently one of the most reliable and accurate ways to detect social signals (Senechal et al., 2015). The facial expression detector employed for this project is a pre-trained convolutional network trained to detect common facial expressions.

To obtain facial feedback at scale, we built a web app that serves samples from a deep learning model while recording the user’s facial expressions with a webcam. The webcam images were fed into the facial-expression detection network and used to compute the intensities of common expressions, including amusement, contentment, surprise, sadness, and concentration. Due to the degree of inter-individual variation in users’ resting facial expression, these intensities were normalized against each user’s average expression to produce value vectors . The app is also capable of collecting Likert-scale ratings of sketch quality and asking users to choose which of two sketches they prefer. These mechanisms were used to collect evaluation data. The app can be viewed at https://facial-feedback-for-ai.appspot.com/.

To test the hypothesis that facial feedback can improve the outputs of a deep learning model, we sought a model for which the outputs were likely to generate a natural facial expression response. We chose Sketch RNN (Ha and Eck, 2017), a model which generates sequences of strokes that form a sketched image of a common object, vehicle, or animal (see Figure 5). Such sketches were determined to elicit facial responses in initial tests.

3.2 Machine learning techniques

Sketch RNN is a VAE that was trained in an unsupervised manner on a large corpus on human sketch data collected via Quick, Draw!111https://quickdraw.withgoogle.com/. The sketches are represented as sequences of coordinates that represent the points where the pen is placed during sketching. The architecture of Sketch RNN comprises: a) a bidirectional LSTM encoder that projects each input sketch into a latent embedding vector , b) an LSTM decoder which takes

as input and generates a sequence of parameters for c) a Gaussian Mixture Model that generates the

coordinates of the tip of the pen during each stroke. This Mixture Density Network (MDN) approach is similar to prior work on handwriting generation (Graves, 2013).

The design of Sketch RNN provides important benefits that facilitate optimizing the model with facial feedback. First, due to the variational constraint, it is straightforward to sample a latent vector and feed this into the Sketch RNN decoder to produce a recognizable sketch. Second, the latent embeddings learned by Sketch RNN provide a clean, compressed representation of sketch drawings. These features allowed us to apply a newly developed technique known as Latent Constraints (Engel et al., 2017) in order to learn to produce sketches likely to lead to positive facial expressions.

Figure 1: Steps involved in training the LC-GAN facial feedback model.

The latent constraints GAN (LC-GAN) is a GAN applied to the latent embedding space of a VAE. The steps of training this model to use facial feedback are shown in Figure 1. We first sample a number of vectors from the VAE prior (), and feed these into the Sketch RNN decoder to obtain sketches. These sketches are shown to users, and the intensity of their facial expression responses is recorded; we refer to these intensities as the value of a sketch, . The vectors are then used as input to a discriminator

, which is trained to estimate the value

of different regions of the latent space; for example, which regions decode to sketches that produced the highest intensity of smiles. A generator is then trained to convert a randomly sampled into a modified that produces a higher . In fact, the generator uses a gating mechanism to control how heavily the original is modified. The generator loss is . Because the Sketch RNN latent space is a compressed, 128-dimensional, robust representation, the discriminator is able to learn a value function on even with relatively small sample sizes.

4 Experiments

Data collection for the experiments was conducted in four phases. In the initial phase, we ran a pilot study on users who viewed a total of sketches, in which we collected both facial expressions and Likert-scale ratings of sketch quality. Then, we used the webapp to collect facial reactions from 28 users to a total of 334 sketches, recording the embedding vector for each sketch. These pairs were used to train the LC-GAN. Finally, two phases of data collection were used to evaluate the model. Both involved rigorous, double-blind experiments in which we randomly generated hundreds of samples from both the facial feedback and baseline models, and displayed them in random order to users “in the wild”, using their personal webcams, without experimenter supervision. The first evaluation sought to establish that sketches from the LC-GAN are able to elicit more positive facial expressions than the original Sketch RNN. For this test, we obtained evaluation data from users, spanning 536 sketches. The second evaluation asked users to rate which of two sketches they preferred; we collected 4,692 ratings from 79 users.

(a) Contentment, =-.58, .001
(b) Concentration, =.58, .001
Figure 2: Correlation between quality ratings and expressions.

5 Results

5.1 Facial expression analysis

Optimizing for user preference requires knowing which facial expressions indicate that the user likes a sketch. Therefore, we used the pilot study data to assess how users’ ratings of sketch quality related to their facial expressions. We found significant positive correlations with contentment and amusement (smiling), and significantly negative correlations with sadness and concentration (frowning), as shown in Table 1; examples are shown in Figure 2. Notably, these results indicate that implicit facial feedback carries an informative signal about the user’s preferences.

Emotion metric
Contentment .582 .001
Amusement .546 .002
Concentration -.576 .001
Sadness -.405 .026
Table 1: Correlations between expressions and quality ratings.

However, learning from facial feedback still represents a challenging problem; there is a high degree of inter-individual variability, the meaning of facial expressions may be extremely context dependent, and the data can be remarkably noisy. In addition to noise introduced through inaccuracies in the detector, there are many confounding reasons that may cause a person to make a given facial expression.

Figure 3: The user expresses contentment and amusement at a scribbled sketch.

For example, Hoque and Picard (2011) found that users tend to smile when they are frustrated. In our case, we noticed that users tend to smile simply at the concept of an AI making drawings, or even at their own face as shown to them via the webcam feed. This can lead to highly misleading interactions; Figure 3 shows an example in which the user smiles profusely at a drawing that is no better than a scribble.

Finally, the difficulty of modeling user preference through facial expressions is enhanced by the non-stationarity of the data. We found that users’ facial expressions tended to change over repeated interactions with the system. There were significant relationships between the number of previous sketches viewed by the user and the user’s average sadness () and concentration ). Thus, the meaning of an intense expression of concentration may change depending on when it occurs in the user’s interaction with the app.

5.2 Machine learning results

Figure 4: Sketches from the LC-GAN generate more positive expressions.

In spite of the noise and non-stationarity inherent in the data, we found that the LC-GAN was able to use the facial feedback to learn produce significantly higher quality sketches. Given the direction of the relationships discovered between facial expressions and quality, we trained the LC-GAN to maximize amusement and contentment, and minimize concentration and sadness. Although relatively little data was collected (63-69 samples per sketch class), the LC-GAN was able to effectively optimize for more pleasing sketches. Figure 5 shows the difference between samples produced with the Sketch RNN prior and the LC-GAN. The LC-GAN appears to have learned that people smile more and frown less for cats with larger, smiling faces with whiskers. Similarly, the quality of crab and rhinoceros sketches generated by the LC-GAN appears to be consistently higher. For example, the original rhinoceros model often produced sketches that did not resemble a rhinoceros, or were no better than scribbles. After training with a small amount of facial feedback, the LC-GAN model consistently produces more realistic drawings.

Figure 5: Samples drawn randomly from the cat, crab, and rhinoceros sketch classes, produced by (a) the original Sketch RNN, and (b) the LC-GAN trained on a small amount of social feedback.

The evaluation data revealed that humans found sketches from the LC-GAN model to be significantly better. In the first experiment, we were able to support the hypothesis that sketches from the LC-GAN model generate significantly more positive facial expressions than the original Sketch RNN. Figure 4 shows the results of this evaluation, indicating that all of the facial expression metrics improved in the expected direction under the LC-GAN. Two of the metrics reached statistical significance: mean amusement, , and mean sadness, . In the second part of the evaluation, we tested the hypothesis that humans actually rate the quality of the LC-GAN sketches as higher. Users reported preferred the LC-GAN 2843 times, as opposed to 1770 for the original Sketch RNN, a significant improvement as shown in a Binomial test, .

6 Conclusions and Future Work

We have demonstrated that implicit social feedback in the form of facial expressions not only can reflect user preference, but also can significantly improve the performance of a deep learning model.

There are many ways to enhance and extend this work. For example, we could use an RL framework to improve the model’s ability to draw based on facial feedback. Further, our current model makes no use of the evolving temporal dynamics of the collected facial expressions, instead relying on an average intensity over viewing the sketch. A more sophisticated system could use the alignment between the process of drawing the sketch and the user’s expressions to gain better temporal credit assignment in an RL framework.

We would like to thank James Tolentino, Ira Blossom, Adrien Baranes, Katherine Lee, Chris Han, Curtis Hawthorne, Rebecca Salois, Josh Lovejoy, Sherol Chen, and Mike Dory for their contributions to this project.