Understanding Grounded Language Learning Agents

10/26/2017 ∙ by Felix Hill, et al. ∙ Google 0

Neural network-based systems can now learn to locate the referents of words and phrases in images, answer questions about visual scenes, and even execute symbolic instructions as first-person actors in partially-observable worlds. To achieve this so-called grounded language learning, models must overcome certain well-studied learning challenges that are also fundamental to infants learning their first words. While it is notable that models with no meaningful prior knowledge overcome these learning obstacles, AI researchers and practitioners currently lack a clear understanding of exactly how they do so. Here we address this question as a way of achieving a clearer general understanding of grounded language learning, both to inform future research and to improve confidence in model predictions. For maximum control and generality, we focus on a simple neural network-based language learning agent trained via policy-gradient methods to interpret synthetic linguistic instructions in a simulated 3D world. We apply experimental paradigms from developmental psychology to this agent, exploring the conditions under which established human biases and learning effects emerge. We further propose a novel way to visualise and analyse semantic representation in grounded language learning agents that yields a plausible computational account of the observed effects.

READ FULL TEXT VIEW PDF

Authors

page 3

page 9

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

1 Introduction

The learning challenge faced by children acquiring their first words has long fascinated cognitive scientists and philosophers (Quine, 1960; Brown, 1973). To start making sense of language, an infant must induce structure in a constant stream of continuous visual input, slowly reconcile this structure with consistencies in the available linguistic observations, store this knowledge in memory, and apply it to inform decisions about how best to respond.

Many neural network models also overcome a learning task that is – to varying degrees – analogous to early human word learning. Image classification tasks such as the ImageNet Challenge 

(Deng et al., 2009) require models to induce discrete semantic classes, in many cases aligned to words, from unstructured pixel representations of large quantities of photographs (Krizhevsky et al., 2012). Visual question answering (VQA) systems (Antol et al., 2015; Xiong et al., 2016; Xu & Saenko, 2016) must reconcile raw images with (arbitrary-length) sequences of symbols, in the form of natural language questions, in order to predict lexical or phrasal answers. Recently, situated language learning agents have been developed that learn to understand sequences of linguistic symbols not only in terms of the contemporaneous raw visual input, but also in terms of past visual input and the actions required to execute an appropriate motor response (Oh et al., 2017; Chaplot et al., 2017; Hermann et al., 2017; Misra et al., 2017). The most advanced such agents learn to execute a range of phrasal and multi-task instructions, such as find the green object in the red roompick up the pencil in the third room on the right or  

go to the small green torch

, in a continous, simulated 3D world. To solve these tasks, an agent must execute sequences of hundreds of fine-grained actions, conditioned on the available sequence of language symbols and active (first-person) visual perception of the surroundings. Importantly, the knowledge acquired by such agents while mastering these tasks also permits the interpretation of familiar language in entirely novel surroundings, and the execution of novel instructions composed of combinations of familiar words (Chaplot et al., 2017; Hermann et al., 2017).

The potential impact of situated linguistic agents, VQA models and other grounded language learning systems is vast, as a basis for human users to interact with situated learning applications such as self-driving cars and domestic robotic tools. However, our understanding of how

these agents learn and behave is limited. The challenges of interpreting the factors or reasoning behind the decisions and predictions of neural networks are well known. Indeed, a concerted body of research in both computer vision 

(Zeiler & Fergus, 2014; Simonyan et al., 2014; Yosinski et al., 2015)

and natural language processing 

(Linzen et al., 2016; Strobelt et al., 2016) has focused on addressing this uncertainty. As grounded language learning agents become more prevalent, then, understanding their learning dynamics, representation and decision-making will become increasingly important, both to inform future research and to build confidence in users who interact with such models.

We therefore aim to establish a better understanding of neural network-based models of grounded language learning, noting the parallels with research in neuroscience and psychology that aims to understand human language acquisition. Extending the approach of Ritter et al. (2017), we adapt various experimental techniques initially developed by experimental psychologists (Landau et al., 1988; Markman, 1990; Hollich et al., 2000). In line with typical experiments on humans, our experimental simulations are conducted in a highly controlled environment: a simulated 3D world with a limited set of objects and properties, and corresponding unambiguous, symbolic linguistic stimuli (Figure 1). However, the simplicity and generality of our architecture and the form of the inputs to the model (continuous visual plus symbolic linguistic) make the proposed methods and approach directly applicable to VQA and other tasks that combine linguistic and visual data. Using these methods, we explore how the training environment of our agent affects its learning outcomes and speed, measure the generality and robustness of its understanding of certain fundamental linguistic concepts, and test for biases in the decisions it takes once trained. Further, by applying layerwise attention, a novel tool for visualising computation in grounded language learning models, we obtain a plausible computational account of some of the effects in terms of representation and processing. Our principal findings about this canonical grounded language learning architecture are the following:

Shape / colour biases

If the agent is exposed to more shape words than colour words during training, it can develop a human-like propensity to assume that new words refer to the shape rather than the colour of objects. However, when the training distribution is balanced between shape and colour terms, it develops an analagous bias in favour of colours.

The problem of learning negation

The agent learns to execute negated instructions, but if trained on small amounts of data it tends to represent negation in an ad hoc way that does not generalise.

Curriculum effects for vocabulary growth

The agent learns words more quickly if the range of words to which it is exposed is limited at first and expanded gradually as its vocabulary develops.

Semantic processing and representation differences

The agent learns words of different semantic classes at different speeds and represents them with features that require different degrees of visual processing depth (or abstraction) to compute.

Before describing the experiments that reveal these effects, we briefly outline details of the environment and agent used for the simulations.

2 A 3D world for language learning

Our experiments take place in the DeepMind Lab simulated world (Beattie et al., 2016), modified to include a language channel. An agent in this environment receives textual instructions, such as find the pencil

, and is rewarded for satisfying the instruction, in this case by executing movement actions (move-left, turn right etc.) that allow it to locate a (3D, rotating) pencil and move into the space that it occupies. At each timestep in such an episode, the agent receives a continuous (RGB) pixel tensor of visual input and a symbolic (word-level) textual instruction,

111In all simulations in this paper, the textual instruction remains constant for the duration of each episode. and must execute a movement action. To solve tasks and receive rewards, the agent must therefore first learn to perceive this environment, actively controlling what it sees via movement of its head (turning actions), and to navigate its surroundings via meaningful sequences of actions.

A typical simulation involves specifying certain aspects of the environment while leaving others to be determined randomly. For instance, in an object identification task, we might wish to specify the overall layout of the world, the range of positions in which objects can appear, a list of objects that can appear in each position, a probability of appearance and rewards associated with selecting each object. The environment engine is then responsible for randomly instantiating episodes that satisfy these constraints together with corresponding language instructions. Even with a detailed specification and a finite inventory of objects, properties and instruction words, there are tens of millions of unique episodes that the agent can encounter during training, each involving different object shapes, colours, patterns, shades, sizes and/or relative positions. With respect to the goal of understanding models of grounded language learning, this simulated environment and synthetic language is a useful asset: we can straightforwardly apply the methods of behavioural psychologists, testing how agents respond to precisely crafted training and test stimuli.

3 A situated language learning agent

Figure 1: Left: Schematic agent architecture. Right: An example of the word learning environment common to all experiments in this paper. The agent observes two 3D rotating objects and a language instruction and must select the object that matches the instruction. In this case the instruction is a shape word (chair). The confounding object (a refrigerator) and the colours of both objects are selected at random and will vary across the agent’s experience of the word chair.

For maximum generality, our simulations involve an agent that combines standard modules for processing sequential symbolic input (a recurrent network) and visual input (a convolutional network). At each time step , the visual input is encoded by the convolutional vision module and a recurrent (LSTM, Hochreiter & Schmidhuber (1997)language module encodes the instruction string . A mixing module determines how these signals are combined before they are passed to a LSTM action module : here is simply a feedforward linear layer operating on the concatenation of the output from and . The hidden state of

is fed to a policy function, which computes a probability distribution over possible motor actions

, and a state-value function approximator

, which computes a scalar estimate of the agent value function for optimisation.

Val estimates the expected discounted future return, by approximating the state-value function where is the state of the environment at time when following policy and is the reward received following the action performed at time . represents a discount parameter. Note that this architecture is a simplified version of that proposed by Hermann et al. (2017), without auxiliary learning components.

Weight updates are computed according to the asynchronous advantage actor-critic (A3C) algorithm (Mnih et al., 2016)

, in conjunction with the RMSProp update rule

(Tieleman & Hinton, 2012)

. During training, a single parameter vector is shared across

CPU cores, which offers a suitable tradeoff between training time and loss of accuracy due to the asynchronous updates.

4 Experiments

4.1 Word learning biases

One effect that is considered instrumental in allowing children to overcome the challenges of early word learning is the human shape bias (Landau et al., 1988), whereby infants tend to to presume that novel words refer to the shape of an unfamiliar object rather than, for instance, its colour, size or texture.

Our simulated environment permits the replication of the original experiment by Landau et al. (1988) designed to demonstrate the shape bias in humans. During training, the agent learns word meanings in a room containing two objects, one that matches the instruction word (positive reward) and a confounding object that does not (negative reward). Using this method, the agent is taught the meaning of a set of colour terms, of shape terms and of ambiguous terms (in the original experiment, the terms were the nonsense terms ‘dax’ and ‘riff’). The target referent for a shape term can be of any colour and, similarly, the target referent when learning the colours in can be of any shape. In contrast, the ambiguous terms in always correspond to objects with a specific colour and shape (e.g. ‘dax’ always referred to a black pencil, and neither black nor pencils were observed in any other context) . Note also that colour terms refer to a range of RGB space through the application of Gaussian noise to prototypical RGB codes, so that two instances of red objects will have subtly different colours.

As the agent learns, we periodically measure its bias by means of test episodes for which no learning takes place. In a test episode, the agent receives an instruction (‘dax’) and must decide between two objects, , whose shape is and whose colour is (a blue pencil), and , whose shape is and whose colour is (a black fork). Note that in the present example neither the colour blue nor the shape fork are observed by the agent during training. As with the original human experiment, the degree of shape bias in the agent can be measured, as the agent is learning, by its propensity to select in preference to . Moreover, by varying the size of sets and , we can examine the effect of different training regimes on this bias exhibited by the agent.

Figure 2 illustrates how a shape/colour bias develops in agents exposed to three different training regimes. An agent that is taught only shape terms and no colour terms (, ) exhibits a shape bias, whereas a colour bias (of larger magnitude) develops in an agent that is taught only colour words (, ). An agent that is taught an equal number of shape and colour terms (, ) develops a moderate colour bias.

This effect provides a possible explanatory account of the rapid acceleration of word learning observed by Hermann et al. (2017); it is precisely the progressive specialisation of the agent’s object recognition and labelling mechanisms (towards shapes, colours or both, as determined by the training regime) that narrows the space of possible referents, permitting faster word learning as training progresses.222We analyse this specialisation at a computational level in Section 4.4.

These conclusions can be incorporated with those of Ritter et al. (2017), who observe a shape bias in convolutional networks trained on the ImageNet Challenge training set. Our experiments with a single canonical architecture exposed to different training stimuli indicate the cause of this effect to be the training data distribution (the ImageNet data indeed contains many more shape-based than colour-based categories) rather than the convolutional architecture itself. Indeed, our findings suggest that a feed-forward convolutional architecture operating (bottom-up) on image pixels promotes a colour rather than shape bias. On the other hand, a typical linguistic environment (for American children at least333As verified by simple analysis of the child-directed language corpus Wordbank (Frank et al., 2017).

) and, perhaps by extension, most broad-coverage machine-learning datasets, contains many more instances of shape categories than colour categories.

Figure 2: Degrees of shape bias for different training regimes: An agent that is trained only on shape words (right) more readily presumes that ambiguous words refer to object shape than to object colour. This tendency is measured across all combinations of known and confounding objects and labels and represented by the blue line. The magnitude of the bias on the scale is the mean ‘score’ ( for the object matching the instruction in shape and

for the object matching in colour) over 1000 random test episodes. In contrast, an agent trained only on colour words (left) exhibits a colour bias. Interestingly, an agent trained on 8 colour and 8 shape words (middle) also exhibits a colour bias. Data (in this and proceeding figures) show mean and standard error across five fastest-learning agents of 16 different hyperparameter settings, sampled at random from ranges specified in in supplementary material 

6.1.

4.2 The problem of learning negation

The interpretation of negated sentences such as tell me a joke that is not offensive is a fundamental facet of natural language understanding, and potentially critical for artificial agents receiving instructions from human users. Despite its communicative importance, negation can be challenging to acquire for human language learners. For instance, negated utterances pose greater production and comprehension difficulties than the equivalent non-negated language (Nordmeyer & Frank, 2014; Pea, 1980).

To explore the acquisition of negation in grounded language learning models, we designed a simulation in which, as before, our agent was placed in a single room and required to select one of two objects matching an instruction. From a full set of training words (e.g. red or ball), a subset, , was sampled and presented to the agent in both positive and negative forms (ball, not ball) and a disjoint subset, , was provided only in positive forms (pencil). To test whether the agent could learn a generally applicable notion of negation, we periodically measured its ability to interpret negated versions of the instructions in .

As illustrated in Figure 3, the agent learned to follow both positive and negative instructions, for various sets of instruction words . However, unlike other linguistic operations such as adjective-noun modifiation (Hermann et al., 2017), the agent exhibited difficulty generalising the notion of negation acquired in the context of to the held-out items . This difficulty was most acute when consisted of 12 colour terms split evenly into and . Indeed, the ability to generalise negation improved as the size of increased to include 40 shapes, from just above chance (a small positive average reward) to 75% (yielding an average reward of ). There was also a small but interesting difference in generalisation when negating shape terms vs. colour terms, which is consistent with the processing differences discussed above and in more detail in Section 4.4.

We conjecture that negation does not generalise easily because, for a word and corresponding extension set of objects , the agent can perform perfectly on the training set by simply associating instructions of the form ‘not w’ with the extension . This understanding of negation would generalise much worse than an interpretation of ‘not w’ that involved identifying and avoiding an object of type w. For small training sets, the results suggest that the model prefers the former interpretation, but also that its tendency to discover the latter more generalisable understanding increases as the set of negated concepts to which it is exposed grows. Thus, with appropriately broad exposure to instances of negation during training, neural networks without bespoke biases or regularization can learn to respond effectively to negated stimuli pertaining to their perceptible surroundings. However, tailored architectures and computational biases may be required in cases where agents learn from, and act on, constrained sets of linguistic stimuli.

Figure 3: The problem of learning negation in language learning agents: The agent must be exposed to negative instructions in a sufficiently diverse range of contexts in order to learn a useful generalisable notion of negation. If trained to interpret positive commands involving 12 terms and negative commands involving 6 of those 12 terms (left, colour terms, middle, shape terms), the agent does not effectively interpret negative commands involving the remaining 6 terms. When exposed to 40 shape terms and trained to interpret negative commands involving 20 of those terms, the agent generalises the negation operation more effectively, but still not perfectly.

4.3 Curriculum learning

The idea that learning is more successful if simpler things are studied before more complex things is a basic tenet of human education. There is some consensus that early exposure to simple, clear linguistic input helps child language acquisition (Fernald et al., 2010), although this is not unanimous (Shore, 1997)

. Deep learning researchers have also reported faster or more effective learning when training examples are ordered by some metric of complexity, an approach referred to as curriculum learning 

(Bengio et al., 2009). However, robust improvements due to curriculum learning can be difficult to achieve in the context of text-based learning (Mikolov et al., 2011; Graves et al., 2017), and curriculum learning is not ordinarily applied when training text-based neural language models. Recent evidence suggests that the importance of curriculum training may be far greater for agents learning to act conditioned on language than those learning to map between linguistic inputs and outputs. Both Hermann et al. (2017) and Oh et al. (2017) observed that curricula were essential for agents learning to execute linguistic instructions that require both resolving of referring expressions (get the red ball..) and non-trivial action policies such as exploration (..in the green room).

Here, we chose to explore curriculum learning in a more controlled way in grounded language learning agents. To do so, we trained our agent to learn the meaning of 40 shape words (as exhibited by its ability to respond appropriately) under two conditions. In one condition, the agent was presented with the 40 words (together with corresponding target and confounding objects) sampled randomly throughout training. In another condition, the agent was only presented with a subset of the 40 words (selected at random) until these were mastered (as indicated by an average reward of over 1000 consecutive trials), at which point this subset was expanded to include more words. This process was iterated for subsets of size 2, 5, 10 and eventually 40 words.

Figure 4: Curriculum training expedites vocabulary growth: An agent that is presented with stimuli sampled uniformly from a set of 40 shape words (red line) learns more slowly than one whose stimuli are constrained to a two-word subset , until the agent learns both words, then extended to a 5-word subset , then a 10-word subset . This strong effect of ‘curriculum learning’ can be observed both when comparing average reward when agents in the two conditions are learning words sampled from (left - note that the agent in the curriculum condition begins reporting performance on after prior training on the restricted subsets) and by measuring vocabulary size as a function of training episodes (right).555Here, a word is considered to be ‘known’, and thus part of the agent’s vocabulary, if the agent responds correctly to that word for 50 consecutive exposures.

As shown in Figure 4, an agent that followed the curriculum (i.e. in the second condition) learned notably faster overall than one presented directly with a large group of new words. This result further corroborates the importance of training curricula for grounded language learning agents. Moreover, unlike the effect observed by Hermann et al. (2017), which focused on tasks requiring the agent to explore a large maze, the present simulation demonstrates strong curriculum effects simply when learning to associate objects with words. Thus, the development of core linguistic and semantic knowledge in situated or grounded agents can be clearly expedited by starting small and easy and slowly increasing the language learning challenge faced by the agent.

4.4 Processing and representation differences

Many studies aiming to understand linguistic processing in humans do so by uncovering connections between different semantic or conceptual domains and distinct processing or patterns of representation. These effects emerge via both behavioural methods (measuring differences in how subjects learn, use or even forget concepts of different types (Bowerman & Choi, 2001; Paivio et al., 1994)) and neuroimaging (associating words of different types with spatially and structurally distinct brain regions (Huth et al., 2016; Patterson et al., 2007)). Neuroscientific theories of memory, representation and learning have been developed to account for these effects (Rogers & McClelland, 2004; Binder & Desai, 2011). In the pursuit of a better understanding of artificial agents, we can similarly explore links between word classes, semantic domains and patterns of learning, behaviour, processing or representation. This knowledge could ultimately be essential for informing the process of designing architectures capable of learning not just the simple language studied here, but also the full range of abstract semantic phenomena inherent in adult language.

Word learning speeds

The order in which words of different types are acquired has been used to inform theories of child language acquisition (Gentner, 1982) and human semantic processing more generally (Ghyselinck et al., 2004). To explore the order of word learning in our agent, we exposed it to randomly interleaved training instances for words of six different classes (shapes, colours, patterns, sizes, shades and superordinate category terms, such as  furniture), for multiple shapes. We compared the rates of word learning in two conditions. In the controlled condition, each class was restricted to two exemplars. In the naturalistic condition, each class was represented by a different number of members, a more faithful reflection of natural language data.666See supplementary material 6.2 for details. In both conditions, the training stimuli were sampled random uniformly from all word types (not word classes), so that an agent in the naturalistic condition received approximately four times as much exposure to shape words as to colour words, for instance.

As illustrated in Figure 5, there were clear differences in the speed with which the agent learned words of different classes. In the controlled condition, the first words to be learned were blue (a colour word) and diagonal-striped (a pattern), with the second colour word, brown, learned around the same time as the two shapes chair and suitcase and the relative size terms larger and smaller. Category terms were learned after shape terms.777Humans also learn basic level categories before superordinate category terms (Horton & Markman, 1980). In contrast, in the naturalistic condition the variable exposure to different word classes seems to cause a degree of specialisation in shape words, so that the agent learns all 40 shape words well before it acquires the 12 colour words.

Figure 5: Words from different semantic classes are learned at different speeds: In the controlled condition (left), the agent learns two words from each class. In the realistic condition (right) each class has a different number of members, as per supplementary material 6.2.
Layer-wise attention

Figure 6: Representation and processing differences between colour and shape words. ‘Dashboards’ for interpreting processing in an agent with layerwise attention. The large pane at the top left of the dashboard shows the input to the agent. The bar chart on the bottom left shows the attention distribution over all 520 ‘locations’ from the agent’s visual representations. 400 red bars show the attention on the () locations output from the lowest layer of the convnet, 81 green bars show the attention on the () locations from the middle layer and 49 blue bars show the attention on the () locations from the top layer. The small windows on the right side illustrate these attention weights (grouped by layer) propagated back to and superimposed over a greyscale copy of the input image, as described by Simonyan et al. (2014). An agent trained exclusively on colour words (A) relies more on the first and second layers of the convnet than an agent trained exclusively on shape words (B), which uses second and upper layer visual features. C: A schematic of layerwise attention in the agent architecture. D: A 2D (t-SNE) visualisation of the space of the word embeddings weights in the language module (L) of an agent trained on different word types, illustrating that words cluster naturally according to semantic classes in the linguistic memory of the agent.

To complement our behavioural analysis we developed a method for better understanding semantic processing and representation in the agent at a computational level. The method, which we call layer-wise attention

, involves modifying the agent architecture to expose processing differences in the visual features that are most pertinent to each lexical concept. In the standard agent, a distributed representation of the linguistic input is concatenated with the output from the top layer of a 3-layer convolutional visual module at each timestep, fed through a multi-layer perceptron and then passed to the agent’s (recurrent) core. We modify this agent so that it can learn to attend to the output from different layers of its visual processing module, conditioned on the linguistic input available at a particular moment.

Let be the representation of a language instruction and be the output of layer of the visual module with dimension , where is the number of feature maps. In the layerwise attention module, the are first passed through 3 independent linear layers to with common final dimension , such that is also the length of . The are then stacked into a single tensor of dimension , where . is then multiplied by

and passed through a softmax layer to yield a

dimensional discrete probability distribution over all (pixel-like) locations represented in each layer of the visual module . These values are applied multiplicatively to each of the (-length) representations returned by , before a pooling step (mirroring that of the final layer in the original agent) and then concatenation.

Layerwise attention provides a measure not only of which image locations contain the most important information for the agent when choosing actions at a given timestep, but also at what level of (visual) abstraction that information is most useful. This insight can be visualised by applying the method of Simonyan et al. (2014), propagating the probability mass from the attention distribution back onto the input image. Figure 6

illustrates the effect of backpropagating the attention probabilities corresponding to each layer of the convnet onto (grayscale copies of) the visual input.

888Best viewed in video format at https://www.youtube.com/watch?v=9vY8D4wuEV0.

As is clear from these visualisations, an agent that is exposed only to shape words will learn to rely on features from the upper-most layers of its visual module when considering objects in its surroundings. In contrast, an agent trained to interpret only colour terms focuses with feature detectors from the lower layers of its visual module in order to distinguish between objects of interest. It is well established that convolutional networks trained to classify images also exhibit differential specialisation of feature detectors between layers (see e.g.  

LeCun et al. (2010)). Layerwise attention provides a means to quantify the magnitude of this specialisation, and to measure the importance of each layer with respect to particular linguistic stimuli. It is also notable that a more conventional 2D (T-SNE) visualisation of the word embeddings in the input layer of provides further evidence of of word-class-specific processing, as illustrated in Figure 6.

5 Conclusion

Models that are capable of grounded language learning promise to significantly advance the ways in which humans and intelligent technology can interact. In this study, we have explored how a situated language learning agent built from canonical neural-network components overcomes the challenge of early language learning. We measured the behaviour exhibited once the first words and simple phrases are acquired, tested factors that speed up this learning, explored aspects of language that pose particular problems and presented a technique, layerwise attention, for better understanding semantic and visual processing in such agents.

The application of experimental paradigms from cognitive psychology to better understand deep neural nets was proposed by Ritter et al. (2017), who observed that convolutional architectures exhibit a shape bias when trained on the ImageNet Challenge data. The ability to control precisely both training and test stimuli in our simulated environment allowed us to isolate this effect as deriving from the training data, and indeed to reach the opposite conclusion about the architecture itself. This study also goes beyond that of Ritter et al. (2017) in exploring more abstract linguistic operations (negation, abstract category terms) and studying curriculum effects on the dynamics of word learning. Further, we complement these behavioural observations with computatoinal analysis of representation and processing, via layerwise attention.

While the control and precision afforded by the simulated environment in the present study has made these analyses and conclusions possible, in future, as our understanding of language learning agents develops, it will be essential to verify conclusions on agents trained on more naturalistic data. At first, this might involve curated sets of images, videos and naturally-occurring text etc, and, ultimately, experiments on robots trained to communicate about perceptible surroundings with human interlocutors. In a world with agents capable of learning such advanced linguistic behaviour, it would certainly be more challenging, but also even more crucial, to understand not just what they can do, but also how they learn to do it.

Acknowledgments

We would like to thank Michael C Frank, Noah Goodman, Jay McClelland, Sam Ritter, Hinrich Schütze and all attendees of the MIC3 workshop for many helpful ideas and suggestions.

References

  • Antol et al. (2015) Stanislaw Antol, Aishwarya Agrawal, Jiasen Lu, Margaret Mitchell, Dhruv Batra, C. Lawrence Zitnick, and Devi Parikh. VQA: Visual Question Answering. In International Conference on Computer Vision (ICCV), 2015.
  • Beattie et al. (2016) Charles Beattie, Joel Z. Leibo, Denis Teplyashin, Tom Ward, Marcus Wainwright, Heinrich Küttler, Andrew Lefrancq, Simon Green, Víctor Valdés, Amir Sadik, Julian Schrittwieser, Keith Anderson, Sarah York, Max Cant, Adam Cain, Adrian Bolton, Stephen Gaffney, Helen King, Demis Hassabis, Shane Legg, and Stig Petersen. Deepmind lab. CoRR, abs/1612.03801, 2016. URL http://arxiv.org/abs/1612.03801.
  • Bengio et al. (2009) Yoshua Bengio, Jérôme Louradour, Ronan Collobert, and Jason Weston. Curriculum learning. In Proceedings of the 26th annual international conference on machine learning, pp. 41–48. ACM, 2009.
  • Binder & Desai (2011) Jeffrey R Binder and Rutvik H Desai. The neurobiology of semantic memory. Trends in cognitive sciences, 15(11):527–536, 2011.
  • Bowerman & Choi (2001) Melissa Bowerman and Soonja Choi. Shaping meanings for language: universal and language-specific in the acquisition of semantic categories. In Language acquisition and conceptual development, pp. 475–511. Cambridge University Press, 2001.
  • Brown (1973) Roger Brown. A first language: The early stages. Harvard U. Press, 1973.
  • Chaplot et al. (2017) Devendra Singh Chaplot, Kanthashree Mysore Sathyendra, Rama Kumar Pasumarthi, Dheeraj Rajagopal, and Ruslan Salakhutdinov. Gated-attention architectures for task-oriented language grounding. arXiv preprint arXiv:1706.07230, 2017.
  • Deng et al. (2009) Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In

    Computer Vision and Pattern Recognition, 2009. CVPR 2009. IEEE Conference on

    , pp. 248–255. IEEE, 2009.
  • Fernald et al. (2010) Anne Fernald, Kirsten Thorpe, and Virginia A. Marchman. Blue car, red car: Developing efficiency in online interpretation of adjective–noun phrases. Cognitive Psychology, 60(3):190 – 217, 2010.
  • Frank et al. (2017) Michael C Frank, Mika Braginsky, Daniel Yurovsky, and Virginia A Marchman. Wordbank: An open repository for developmental vocabulary data. Journal of child language, 44(3):677–694, 2017.
  • Gentner (1982) Dedre Gentner. Why nouns are learned before verbs: Linguistic relativity versus natural partitioning. Center for the Study of Reading Technical Report; no. 257, 1982.
  • Ghyselinck et al. (2004) Mandy Ghyselinck, Michael B Lewis, and Marc Brysbaert. Age of acquisition and the cumulative-frequency hypothesis: A review of the literature and a new multi-task investigation. Acta psychologica, 115(1):43–67, 2004.
  • Graves et al. (2017) Alex Graves, Marc G Bellemare, Jacob Menick, Remi Munos, and Koray Kavukcuoglu. Automated curriculum learning for neural networks. Proceedings of ICML, 2017.
  • Hermann et al. (2017) Karl Moritz Hermann, Felix Hill, Simon Green, Fumin Wang, Ryan Faulkner, Hubert Soyer, David Szepesvari, Wojtek Czarnecki, Max Jaderberg, Denis Teplyashin, et al. Grounded language learning in a simulated 3d world. arXiv preprint arXiv:1706.06551, 2017.
  • Hochreiter & Schmidhuber (1997) Sepp Hochreiter and Jürgen Schmidhuber. Long short-term memory. Neural computation, 9(8):1735–1780, 1997.
  • Hollich et al. (2000) George J Hollich, Kathy Hirsh-Pasek, Roberta Michnick Golinkoff, Rebecca J Brand, Ellie Brown, He Len Chung, Elizabeth Hennon, Camille Rocroi, and Lois Bloom. Breaking the language barrier: An emergentist coalition model for the origins of word learning. Monographs of the society for research in child development, pp. i–135, 2000.
  • Horton & Markman (1980) Marjorie S Horton and Ellen M Markman. Developmental differences in the acquisition of basic and superordinate categories. Child development, pp. 708–719, 1980.
  • Huth et al. (2016) Alexander G Huth, Wendy A de Heer, Thomas L Griffiths, Frédéric E Theunissen, and Jack L Gallant. Natural speech reveals the semantic maps that tile human cerebral cortex. Nature, 532(7600):453–458, 2016.
  • Krizhevsky et al. (2012) Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton.

    Imagenet classification with deep convolutional neural networks.

    In Advances in neural information processing systems, pp. 1097–1105, 2012.
  • Landau et al. (1988) Barbara Landau, Linda B Smith, and Susan S Jones. The importance of shape in early lexical learning. Cognitive development, 3(3):299–321, 1988.
  • LeCun et al. (2010) Yann LeCun, Koray Kavukcuoglu, and Clément Farabet. Convolutional networks and applications in vision. In Circuits and Systems (ISCAS), Proceedings of 2010 IEEE International Symposium on, pp. 253–256. IEEE, 2010.
  • Linzen et al. (2016) Tal Linzen, Emmanuel Dupoux, and Yoav Goldberg. Assessing the ability of lstms to learn syntax-sensitive dependencies. arXiv preprint arXiv:1611.01368, 2016.
  • Markman (1990) Ellen M Markman. Constraints children place on word meanings. Cognitive Science, 14(1):57–77, 1990.
  • Mikolov et al. (2011) Tomáš Mikolov, Anoop Deoras, Daniel Povey, Lukáš Burget, and Jan Černockỳ. Strategies for training large scale neural network language models. In Automatic Speech Recognition and Understanding (ASRU), 2011 IEEE Workshop on, pp. 196–201. IEEE, 2011.
  • Misra et al. (2017) Dipendra K Misra, John Langford, and Yoav Artzi.

    Mapping instructions and visual observations to actions with reinforcement learning.

    Proceedings of EMNLP, 2017.
  • Mnih et al. (2016) Volodymyr Mnih, Adria Puigdomenech Badia, Mehdi Mirza, Alex Graves, Timothy P Lillicrap, Tim Harley, David Silver, and Koray Kavukcuoglu. Asynchronous methods for deep reinforcement learning. In International Conference on Machine Learning, 2016.
  • Nordmeyer & Frank (2014) Ann E Nordmeyer and Michael C Frank. The role of context in young children’s comprehension of negation. Journal of Memory and Language, 77:25–39, 2014.
  • Oh et al. (2017) Junhyuk Oh, Satinder Singh, Honglak Lee, and Pushmeet Kohli. Zero-shot task generalization with multi-task deep reinforcement learning. Proceedings of ICML, 2017.
  • Paivio et al. (1994) Allan Paivio, Mary Walsh, and Trudy Bons. Concreteness effects on memory: When and why? Journal of Experimental Psychology: Learning, Memory, and Cognition, 20(5):1196, 1994.
  • Patterson et al. (2007) Karalyn Patterson, Peter J Nestor, and Timothy T Rogers. Where do you know what you know? the representation of semantic knowledge in the human brain. Nature reviews. Neuroscience, 8(12):976, 2007.
  • Pea (1980) Roy D Pea. The development of negation in early child language. The social foundations of language and thought, pp. 156–186, 1980.
  • Quine (1960) W. V. O. Quine. Word & Object. MIT Press, 1960.
  • Ritter et al. (2017) Samuel Ritter, David GT Barrett, Adam Santoro, and Matt M Botvinick. Cognitive psychology for deep neural networks: A shape bias case study. Proceedings of ICML, 2017.
  • Rogers & McClelland (2004) Timothy T Rogers and James L McClelland. Semantic cognition: A parallel distributed processing approach. MIT press, 2004.
  • Shore (1997) Rima Shore. Rethinking the brain: New insights into early development. ERIC, 1997.
  • Simonyan et al. (2014) K. Simonyan, A. Vedaldi, and A. Zisserman. Deep inside convolutional networks: Visualising image classification models and saliency maps. In Workshop at International Conference on Learning Representations, 2014.
  • Strobelt et al. (2016) Hendrik Strobelt, Sebastian Gehrmann, Bernd Huber, Hanspeter Pfister, and Alexander M Rush. Visual analysis of hidden state dynamics in recurrent neural networks. arXiv preprint arXiv:1606.07461, 2016.
  • Tieleman & Hinton (2012) Tijmen Tieleman and Geoffrey Hinton. Lecture 6.5-rmsprop: Divide the gradient by a running average of its recent magnitude. COURSERA: Neural networks for machine learning, 4(2), 2012.
  • Xiong et al. (2016) Caiming Xiong, Stephen Merity, and Richard Socher. Dynamic memory networks for visual and textual question answering. In International Conference on Machine Learning, pp. 2397–2406, 2016.
  • Xu & Saenko (2016) Huijuan Xu and Kate Saenko. Ask, attend and answer: Exploring question-guided spatial attention for visual question answering. In European Conference on Computer Vision, pp. 451–466. Springer, 2016.
  • Yosinski et al. (2015) Jason Yosinski, Jeff Clune, Anh Nguyen, Thomas Fuchs, and Hod Lipson. Understanding neural networks through deep visualization. ICML Deep Learning workshop, 2015.
  • Zeiler & Fergus (2014) Matthew D Zeiler and Rob Fergus. Visualizing and understanding convolutional networks. In European conference on computer vision, pp. 818–833. Springer, 2014.

6 Supplementary Material

6.1 Agent details

Hyperparameter Value Description
train_steps 640m Theoretical maximum number of time steps (across all episodes) for which the agent will be trained.
env_steps_per_core_step 4 Number of time steps between each action decision (action smoothing)
num_workers 32 Number of independent workers running replicas of the environment with asynchronous updating.
unroll_length 50 Number of time steps through which error is backpropagated in the core LSTM action module
language encoder
encoder_type BOW Whether the language encoder uses an additive bag-of-words (BOW) or an LSTM architecture.
cost calculation
additional_discounting 0.99 Discount used to compute the long-term return R_t in the A3C objective
cost_base 0.5 Multiplicative scaling of all computed gradients on the backward pass in the network
optimisation
clip_grad_norm 100 Limit on the norm of the gradient across all agent network parameters (if above, scale down)
decay 0.99 Decay term in RMSprop gradient averaging function
epsilon 0.1 Epsilon term in RMSprop gradient averaging function
learning_rate_finish 0 Learning rate at the end of training, based on which linear annealing of is applied.
momentum 0 Momentum parameter in RMSprop gradient averaging function
Table 1: Agent hyperparameters that are fixed throughout our experimentation but otherwise not specified in the text.
Hyperparameter Value Description
language encoder
embed_init uniform(0.5, 1) Standard deviation of normal distribution (mean = 0) for sampling
initial values of word-embedding weights in L.
optimisation
entropy_cost uniform(0.0005, 0.005) Strength of the (additive) entropy regularisation term in the A3C cost function.
learning_rate_start loguniform(0.0001, 0.002) Learning rate at the beginning of training
annealed linearly to reach learning_rate_finish at the end of train_steps.
Table 2: Agent hyperparameters that randomly sampled in order to yield different replicas of our agents for training. uniform() indicates that values are sampled uniformly from the range . loguniform()

indicates that values are sampled from a uniform distribution in log-space (favouring lower values) on the range

.

6.2 Experiment details

Word class Words
Shapes (40) chair, suitcase, tv, ball, balloon, cow1, zebra1
cake, can5, cassette, chair, guitar, hair-brush2,
hat4, ice-lolly, ladder, mug5, pencil2, suitcase,
toothbrush2, key2, bottle5, car, cherries3, fork2,
fridge, hammer2, knife2, spoon2, apple3, banana3
flower, jug5, pig1, pincer2, plant3, saxophone,
shoe4, tennis-racket, tomato1, tree3, wine-glass5
Colours (12) blue, brown, pink, yellow, red, green, cyan,
magenta, white, grey, purple
Categories (5) 1:animals, 2:tools, 3:plants,
4:clothing, 5:containers
Patterns (3) diagonal-striped, chequered, spotted
Shades (2) lighter, darker
Sizes (2) larger, smaller
Table 3: Word classes (class size) used in the word learning speed experiment in Section 4.4. Superscript indicates is the shape a word refers to is also in the extension of a category word.