Reinforcement learning (RL) is increasingly being used as a tool to study language emergence (Mordatch and Abbeel, 2017; Chaabouni et al., 2020; Lazaridou and Baroni, 2020). By allowing multiple agents to communicate with each other while solving a common task, a communication protocol needs to be established. The resulting protocol can be studied to see if it adheres to properties of human language, such as compositionality (Kirby, 2001; Geffen Lan et al., 2020). The tasks and environments themselves can also be studied, to see what types of constraints are necessary for human-like language to emerge (Steels, 1997). Referential games are often used for this purpose (Kajic et al., 2020; Havrylov and Titov, 2017; Yuan et al., 2020). While these studies open up the possibility of using computational models to investigate how language emerged and how language is acquired through interaction with an environment and other agents, most RL studies consider communication using discrete symbols.
Spoken language instead operates and presumably emerged over a continuous acoustic channel. Human infants acquire their native language by being exposed to speech audio in their environments (Kuhl, 2005); by interacting and communicating with their caregivers using continuous signals, infants can observe the consequences of their communicative attempts (e.g. through parental responses) that may guide the process of language acquisition (see e.g. Howard and Messum (2014) for discussion). Continuous signalling is challenging since an agent needs to be able to deal with different acoustic environments and noise introduced by the lossy channel. These intricacies are lost when agents communicate directly with discrete symbols. This raises the question: Are we able to observe emergent language between agents with a continuous communication channel, trained through RL? This paper is our first step towards answering this larger research question.
. But these models often rely on heuristic approaches and older neural modelling techniques, making them difficult to extend; e.g. it isn’t easy to directly incorporate other environmental rewards or interactions between multiple agents. More recent RL approaches would make this possible, but as noted, has mainly focused on discrete communication. Our work here tries to bridge the disconnect between recent contributions in multi-agent reinforcement learning (MARL) and earlier literature in language acquisition and modelling(Moulin-Frier and Oudeyer, 2021).
One recent exception which do use continuous signalling within a modern RL framework is the work of Gao et al. (2020)
. In their setup, a Student agent is exposed to a large collection of unlabelled speech audio, from which it builds up a dictionary of possible spoken words. The Student can then select segmented words from its dictionary to play back to a Teacher, which uses a trained automatic speech recognition (ASR) model to classify the words and execute a movement command in a discrete environment. The Student is then awarded for moving towards a goal position. We also propose a Student-Teacher setup, but importantly, our agents can generate their own unique audio waveforms rather than just segmenting and repeating words exactly from past observations. Moreover, in our setup an agent is not required to use a pretrained ASR system for “listening”.
Concretely, we propose the environment illustrated in Figure 1, which is an extension of the referential signalling game of Chaabouni et al. (2020) and Rita et al. (2020). Here represents one out of a set of possible concepts the Speaker must communicate to a Listener agent. Taking this concept as input, the Speaker produces a waveform as output, which passes over a (potentially lossy) acoustic channel. The Listener “hears” the utterance from the speaker. Taking the waveform as input, the Speaker produces output . This output is the Listener’s interpretation of the concept that the Speaker agent tried to communicate. The agents must develop a common communication protocol such that . This process encapsulates one of the core goals of human language: conveying meaning through communication (Dor, 2014). To train the agents, we use deep Q-learning (Mnih et al., 2013).
Our bigger goal is to explore the question of whether and how language emerges when using RL to train agents that communicate via continuous acoustic signals. Our proposed environment and training methodology serves as a means to perform such an exploration, and the goal of the paper is to showcase the capabilities of the platform. Concretely, we illustrate that a valid protocol is established between agents communicating freely, that basic compositionality emerges when agents need to communicate a combination of two concepts, that channel noise affects generalisation, and that one agent will act accordingly when the other is made to “hear” or “speak” English. At the end of the paper, we also discuss questions that can be tackled in the future using the groundwork laid here.
We base our environment on the referential signaling game from Chaabouni et al. (2020) and Rita et al. (2020)—which itself is based on Lewis (1969)—where a sender must convey a message to a receiver. In our case, communication takes place between a Speaker and a Listener over a continuous acoustic channel, instead of sending symbols directly (Figure 1). In each game round, a Speaker agent is tasked with conveying a single concept. The Speaker needs to explain this concept using a speech waveform which is transmitted over a noisy communication channel, and then received by a Listener agent. The Listener agent then classifies its understanding of the Speaker’s concept. If the Speaker’s target concept matches the classified concept from the Listener, the agents are rewarded. The Speaker is then presented with another concept and the cycle repeats.
Formally, in each episode, the environment generatestarget concepts from a set . The Speaker receives and generates a sequence of phones , each representing a phone from a predefined phonetic alphabet . The phone sequence is then converted into a waveform , an audio signal sampled at . For this we use a trained text-to-speech model (Black and Lenzo, 2000; Duddington, 2006). A channel noise function is then applied to the generated waveform, and the result is presented as input to the Listener. The Listener converts the input waveform to a mel-scale spectrogram: a sequence of vectors over time representing the frequency content of an audio signal scaled to mimic human frequency perception (Davis and Mermelstein, 1980). Taking the mel-spectrogram sequence as input, the Listener agent outputs a vector representing its predicted concept. The agents are both rewarded if the predicted word is equal to the target word .
To make the environment a bit more concrete, we present a brief example in Figure 2. For illustrative purposes, consider a set of concepts . The state representation for down would be . A possible phone sequence generated by the Speaker would be .111<s> and </s> respectively represent the start-of-sequence and end-of-sequence tokens. This would be synthesised, passed through the channel, and then be interpreted by the Listener agent. If the Listener’s prediction is , then it selected the correct concept of down. The environment would then reward the agents accordingly:
In our environment we have modelled the task of the Speaker agent as a discrete problem. Despite this, the combination of both agents and their environment is a continuous communication task; in our communication channel, we apply continuous signal transforms which can be motivated by real acoustic environments. The Listener also needs to take in and process a noisy acoustic signal. It is true that the Speaker outputs a discrete sequence; what we have done here is to equip the Speaker with articulatory capabilities so that these do not need to be learned by the model. There are studies that consider how articulation can be learned (Howard and Messum, 2014; Asada, 2016; Rasilo and Räsänen, 2017)
, but none of these do so in an RL environment, rather using a form of imitation learning. In Section5 we discuss how future work could consider learning the articulation process itself within our environment, and the challenges involved in doing so.
3 Learning to Speak and Hear Using RL
To train our agents, we use deep Q-learning (Mnih et al., 2013). For the Speaker agent, this means predicting the action-value of phone sequences. The Listener agent predicts the value of selecting each classification target .
3.1 Speaker Model
The Speaker agent is tasked with generating a sequence of phones describing a concept or idea. The model architecture is shown in Figure 3. The target concept is represented by the one-hot input state
. We use gated recurrent unit (GRU) based sequence generation as the core of the Speaker agent, which generates a sequence of Q-values, a distribution over phonesper output-step from 1 to . The input state is embedded as the initial hidden state of the GRU. The output phone of each GRU layer is embedded as input to the next GRU layer.222No gradients flow through the argmax: this connection indicates to the network which phone was selected at the previous GRU step. We also make use of start-of-sequence (SOS) and end-of-sequence (EOS) tokens,<s> and </s> respectively, appended to the phone-set. These allow the Speaker to generate arbitrary length phone sequences up to a maximum length of .
3.2 Listener Model
The Listener agent may be viewed as a classification task with the full model architecture illustrated in Figure 4. The model is roughly based on (Amodei et al., 2016). Given an input mel-spectrogram , the Listener generates a set of state-action values. These action-values represent the expected reward for each classification vector .
We first apply a set of convolutional layers over the input mel-spectrogram, keeping the size of the time-axis consistent throughout. We then flatten the convolution outputs over the filters and feature axis, resulting in a single vector per time step. We process each vector through a bidirectional GRU, feeding the final hidden state through a linear layer to arrive at our final action-value predictions. An argmax of these action-values gives us a greedy prediction for .
3.3 Deep Q-Learning
The Q-network of the Speaker agent generates a sequence of phones in every communication round until the EOS token is reached. The sequence of phones may be seen as predicting an action sequence per environment step, while standard RL generally only predicts a single action per step. To train such a Q-network, we therefore modify the general gradient-descent update equation from Sutton and Barto (1998). Since we only have a single communication round, we update the model parameters as follows:
where the reward is given in (1), is the environment state, is the action, is the learning rate, and . For the Speaker, is the value of performing the action at output . For the Speaker, the environment state would be the desired concept and the actions would be , the output of the network in Figure 3.
The Listener is also trained using (2), but here this corresponds to the more standard case where the agent produces a single action, i.e. . Concretely, for the Listener this action is , the output of the network in Figure 4. The Listener’s environment is the mel-spectrogram . The Speaker and Listener each have their own independent learner and replay buffer (Mnih et al., 2013). A replay buffer is a storage buffer that keeps track of the observed environment states, actions and rewards. The replay buffer is then sampled when updating the agent’s Q-networks through gradient descent with (2). We may see this two-agent environment as multi-agent deep Q-learning (Tampuu et al., 2017), and therefore have to take careful consideration of the non-stationary replay buffer: we limit the maximum replay buffer size to twice the batch size. This ensures that the agent learns only from its most recent experiences.
, with varianceand , respectively. The channel also masks up to of the mel-spectrogram time-axis during training. We train our agents with an -greedy exploration, where is decayed exponentially from to over the training steps.
We use eSpeak (Duddington, 2006) as our speech synthesiser. eSpeak is a parametric text-to-speech software package that uses formant synthesis to generate audio from phone sequences. Festival (Black and Lenzo, 2000) was also tested, although eSpeak is favoured for its simpler phone scheme and multi-language support. We use eSpeak’s full English phone-set of 164 unique phones and phonetic modifiers. The standard maximum number of phones the Speaker is allowed to generate in each communication round is , including the EOS token. All GRUs have 2 layers with a hidden layer size of 256. All Speaker agent embeddings (Section 3.1) are also 256-dimensional. The Listener (Section 3.2
) uses 4 convolutional layers, each with 64 filters and a kernel width and height of 3. The input to the first convolutional layer is a sequence of 128-dimensional mel-spectrogram vectors extracted every 32 ms. We apply zero padding of size 1 at each layer to retain the input dimensions. Additional experimental details are given in AppendixA.
4.2 Unconstrained Communication of Single Concepts
Motivation We first verify that the environment works as expected and that a valid communication protocol emerges when no constraints are applied to the agents.
Setup The Speaker and Listener agents are trained simultaneously here, as described in Section 3.3. The agents are tasked with communicating 16 unique concepts.
Findings Figure 4(a) shows the mean evaluation reward of the Listener agent over training steps. (This is also an indication of the Speaker’s performance, since without successful coordination between the two agents, no reward is given to either.) The agents achieve a final mean reward of 0.917 after 5000 training episodes, successfully developing a valid communication protocol for roughly 15 out of the total of 16 concepts.333The maximum evaluation reward in all experiments is . What does the communication sound like? Since there are no constraints placed on communication, the agents can easily coordinate to use arbitrary phone sequences to communicate distinct concepts. The interested reader can listen to generated samples.444Audio samples for all experiments are available at https://iclr2022-1504.github.io/samples/. We next consider a more involved setting in order to study composition and generalisation.
Results for unconstrained communication. The agents are evaluated every 100 training episodes over 20 runs. Shading indicates the bootstrapped 95% confidence interval.
4.3 Unconstrained Communication Generalising to Multiple Concepts
Motivation To study composition and generalisation, we perform an experiment based on (Kirby, 2001). They used an iterative language model (ILM) to convey two separate meanings ( and ) in a single string. This ILM was able to generate structured compositional mappings from meaning to strings. For example, in one result they found and . The combination of the two meanings was therefore . Similarly, with . Motivated by this, we try to test the generalisation capabilities in continuous signalling in our environment.
Setup Rather than conveying a single concept in each episode, we now ask the agents to convey two concepts. The target concept and predicted concept now become and , respectively. We also make sure that some concept combinations are never seen during training. We then see if the agents are still able to convey these concept combinations at test time, indicating how well the agents generalise to novel inputs. The reward model is adjusted accordingly, with the agents receiving for each concept correctly identified by the Listener. Here can take on 4 distinct concepts while can take on another 4 concepts. Out of the 16 total combinations, we make sure that 4 are never seen during training. The unseen combinations are chosen such that there remains an even distribution of individual unseen concepts. We also increase the maximum phone length to .
As an example, you can think of as indicating an item from the set of concepts while indicates and item from and we want the agents to communicate concept combinations such as up+fast. Some combinations such as right+slow is never given as the target concept combination during training (but e.g. right+fast and left+slow would be), and we see if the agents can generalise to these unseen combinations at test time and how they do it.
Findings The results are shown in Figure 4(b). We see the mean evaluation reward of the Listener agent reaches on the training concept combinations. The agents achieve a mean evaluation reward of on the unseen combinations, indicating that they are usually able to successfully communicate at least one of the two concepts. The chance-level baseline for this task would receive a mean reward of 0.25. The performance on the unseen combinations is thus better than random.
|jLLeee||@@ R Ree||wwwxxx||sss@@@|
|Average SNR (dB)||Training Codes||Unseen Codes|
Table 2 shows examples of the sequences produced by a trained Speaker agent for each concept combination, with the phone units written using the international phonetic alphabet. Ideally, we would want each row and each column to affect the phonetic sequence in a unique way. This would indicate that the agents have learnt a compositional language protocol, combining phonetic segments together to create a sequence in which the Listener can distinguish the individual component concepts. We see this type of behaviour to some degree in our Speaker samples, such as the [x] phones for or the repeated [s] sound when . This indicates at least some level of compositionality in the learned communication. More qualitatively, the realisation from eSpeak of [L] sounds very similar to [n] for = 0. (We refer the reader to the sample page, linked in Section 4.2.)
The bold phone sequences in Table 2 were unseen during training. The agents correctly classified one combination () out of the 4 unseen combinations. For the other 3 unseen combinations, the agents correctly predicted at least or correctly. These sequences also show some degree of compositionality, such as the [jL] sequence where . We should note that the agents are never specifically encouraged to develop any sort of compositionality in this experiment. They could, for example, use a unique single phone for each of the 16 concept combinations.
Table 2 shows the mean evaluation reward of the same two-concept experiments, but now with varying degrees of channel noise expressed in SNR.555The SNR is calculated based on the average energy in a signal generated by eSpeak. The goal here is to evaluate how the channel influences the generalisation of the agents to unseen input combinations. In the no-channel case, the Speaker output is directly input to the Listener agent, without any time stretching or pitch shifting. The no channel case does best on the training codes as expected, but does not generalise as well to unseen input combinations. We find that increasing channel noise decreases the performance of the training codes and increases generalisation performance on unseen codes, up to a point where both decrease. This is an early indication that the channel specifically influences generalisation.
4.4 Grounding Emergent Communication
Motivation Although the Speaker uses an English phone-set, up to this point there has been no reason for the agents to actually learn to use English words to convey the concepts. In this subsection, either the Speaker or Listener is predisposed to speak or hear English words, and the other agent needs to act accordingly. One scientific motivation for this setting is that it can be used to study how an infant learns language from a caregiver (Kuhl, 2005). To study this computationally, several studies have looked at cognitive models of early vocal development through infant-caregiver interaction; Asada (2016) provides a comprehensive review. Most of these studies, however, considered the problem of learning to vocalise (Howard and Messum, 2014; Moulin-Frier et al., 2015; Rasilo and Räsänen, 2017), which limits the types of interactions and environmental rewards that can be incorporated into the model. We instead simplify the vocalisation process by using an existing synthesiser, but this allows us to use modern MARL techniques to study continuous signalling.
We first give the Listener agent the infant role, and the Speaker will be the caregiver. This mimics the setting where an infant learns to identify words spoken by a caregiver. Later, we reverse the roles, having the Speaker agent assume the infant role. This represents an infant learning to speak their first words and their caregiver responds to recognised words. Since here one agent (the caregiver) has an explicit notion of the meaning of a word, this process can be described as “grounding” from the other agent’s perspective (the infant).
Setup We first consider a setting where we have a single set of 4 concepts . While this is similar to the examples given in preceding sections, here the agents will be required to use actual English words to convey these concepts. In the setting where the Listener acts as an infant, the caregiver Speaker agent speaks English words; the Speaker consists simply of a dictionary lookup for the pronunciation of the word, which is then generated by eSpeak. In the setting where the Speaker takes on the role of the infant, the Listener is now a static entity that can recognise English words; we make use of a dynamic time warping (DTW) system that matches the incoming waveform to a set of reference words and selects the closest one as its output label. 50 reference words are generated by eSpeak. The action-space of the Speaker agent is very large (
), and would be near impossible to explore entirely. Therefore, we provide guidance: with probability(Section 4.1), choose the correct ground truth phonetic sequence for . We also consider the two-concept combination setting of Section 4.3 where either the Speaker or Listener now hears or speaks actual English words; DTW is too slow for the static Listener in this case, so here we first train the Listener in the infant role and then fix it as the caregiver when training the Speaker.
Findings: Grounding the Listener Here the Listener is trained while the Speaker is a fixed caregiver. The Listener agent reached a mean evaluation reward of 1.0, indicating the agent learnt to correctly classify all 4 target words 100% of the time (full graphs given in Appendix B.1). The Listener agent was also tested with a vocabulary size of 50, consisting of the 50 most common English words including the original up, down, left, and right. With this setup, the Listener still reached a mean evaluation reward of 0.934.
Findings: Grounding the Speaker We now ground the Speaker agent by swapping its role to that of the infant. The Speaker agent reaches a mean evaluation reward of 0.983 over 20 runs, indicating it is generally able to articulate all of the 4 target words. Table 3 gives samples of one of the experiment runs and compares them to the eSpeak ground truth phonetic descriptions. Although appearing very different to the ground truth, the audio generated by eSpeak of the phone sequences qualitatively similar. The reader can confirm this for themselves by listening to the generated samples (again we refer the reader to the sample page, linked in Section 4.2.)
|Target word||Ground truth||Predicted phones|
Findings: Grounding generalisation in communicating two concepts Analogous to Section 4.3, we now have infant and caregiver agents in a setting with two concepts, specifically and . Here, these sets don’t simply serve as an example as in Section 4.3, but the Speaker would now actually say “up” when it is the caregiver and the Listener will now actually be pretrained to recognise the word “up” when it is the caregiver. 4 combinations are unseen during training: up-slow, down-regular, left-medium, and right-fast. Again we consider both role combinations of infant and caregiver. Figure 5(a) shows the results when training a two-word Listener agent. The agent reaches a mean evaluation reward of 1.0 for the training codes and 0.952 for the unseen code combinations. This indicates that the Listener agent learns near-optimal generalisation. As mentioned above, for the case where the Speaker is the infant, the DTW-based fixed Listener was found to be impractical. Thus, we use a static Listener agent pre-trained to classify 50 concepts for each and . This totals to 2500 unique input combinations. The results of the two-word Speaker agent are shown in Figure 5(b). The Speaker agent does not perform as well as the Listener agent, reaching a mean evaluation reward of 0.719 for the training word combinations and 0.425 for the unseen.
We have replicated the experiments in this subsection using the Afrikaans version of eSpeak, reaching similar performance to English. This shows our results are not language specific.
The work we have presented here has gone further than Gao et al. (2020), which only allowed segmented template words to be generated: our Speaker agent has the ability to generate unique audio waveforms. On the other hand, our Speaker can only generate sequences based on a fixed phone-set (which is then passed over a continuous acoustic channel). This is in contrast to earlier work (Howard and Messum, 2014; Asada, 2016; Rasilo and Räsänen, 2017) that considered a Speaker that learns a full articulation model in an effort to come as close as possible in imitating an utterance from a caregiver; this allows a Speaker to generate arbitrary learnt units. We have thus gone further than Gao et al. (2020) but not as far as these older studies. Nevertheless, our approach has the benefit that it is formulated in a modern MARL setting: it can therefore be easily extended. Future work can therefore consider whether articulation can be learnt as part of our model – possibly using imitation learning to guide the agent’s exploration of the very large action-space of articulatory movements.
In the experiments carried out in this study, we only considered a single communication round. We also referred to our setup as multi-agent, which is accurate but could be extended even further where a single agent has both a speaking and listening module, and these composed agents then communicate with one another. Future work could therefore consider multi-round communication games between 2 or more agents. Such games would extend our work to the full MARL problem, where agents would need to “speak” to and “hear” each other to solve a common task.
Finally, in terms of future work, we saw in Section 4.3 the importance of the channel for generalisation. Adding white noise is, however, not a good enough simulation of real-life channel acoustic channels. But our approach could be extended with real background noise and more accurate models of environmental dynamics. This could form the basis for a computational investigation of the effect of real acoustic channels in language learning and emergence.
We reflect on our initial research question: Are we able to observe emergent language between agents with a continuous acoustic communication channel trained through RL? This work has laid only a first foundation for answering this larger question. We have showcased the capability of a environment and training approach which will serve as a means of further exploration in answering the question.
This work is supported in part by the National Research Foundation of South Africa (grant no. 120409).
We currently do not identify any obvious reasons to have ethical concerns about this work. Ethical considerations will be made taken into account in the future if some of the models are compared to data from human studies or trials.
- Deep Speech 2: end-to-end speech recognition in English and Mandarin. In Proc. ICML, pp. 173–182. Cited by: §3.2.
- Modeling early vocal development through infant–caregiver interaction: a review. IEEE Transactions on Cognitive and Developmental Systems, pp. 128–138. Cited by: §2, §4.4, §5.
- Building voices in the festival speech synthesis system. unpublished document. External Links: Cited by: §2, §4.1.
- Compositionality and generalization in emergent languages. In Proc. ACL, pp. 4427–4442. Cited by: §1, §1, §2.
- Comparison of parametric representations for monosyllabic word recognition in continuously spoken sentences. IEEE Transactions on Acoustics, Speech, and Signal Processing, pp. 357–366. Cited by: §2.
- The instruction of imagination: language and its evolution as a communication technology. pp. 105–125. Cited by: §1.
- eSpeak text to speech. External Links: Cited by: §2, §4.1.
- Spoken language acquisition based on reinforcement learning and word unit segmentation. In Proc. ICASSP, pp. 6149–6153. Cited by: §1, §5.
- On the Spontaneous Emergence of Discrete and Compositional Signals. In Proc. ACL, pp. 4794–4800. Cited by: §1.
- Emergence of language with multi-agent games: learning to communicate with sequences of symbols. In Proc. NeurIPS, pp. . Cited by: §1.
- Learning to pronounce first words in three languages: an investigation of caregiver and infant behavior using a computational model of an infant. PLOS ONE, pp. 1–21. Cited by: §1, §2, §4.4, §5.
- Learning to cooperate: emergent communication in multi-agent navigation. arXiv e-prints, pp. . Cited by: §1.
Spontaneous evolution of linguistic structure: an iterated learning model of the emergence of regularity and irregularity.
IEEE Transactions on Evolutionary Computation, pp. 102–110. Cited by: §1, §4.3.
- Early language acquisition: cracking the speech code. Nature Reviews Neuroscience, pp. 831–843. Cited by: §1, §4.4.
Emergent multi-agent communication in the deep learning era. CoRR. External Links: Cited by: §1.
- Convention. Blackwell. Cited by: §2.
- Librosa/librosa: 0.8.1rc2. Zenodo. External Links: Cited by: §4.1.
- Playing Atari with deep reinforcement learning. In NIPS Deep Learning Workshop, pp. . External Links: Cited by: §1, §3.3, §3.
- Emergence of grounded compositional language in multi-agent populations. In Proc. AAAI, pp. . Cited by: §1.
- COSMO (“communicating about objects using sensory–motor operations”): a bayesian modeling framework for studying speech communication and the emergence of phonological systems. Journal of Phonetics, pp. 5–41. Cited by: §4.4.
- Multi-Agent Reinforcement Learning as a Computational Tool for Language Evolution Research: Historical Context and Future Challenges. In Proc. AAAI, Cited by: §1.
- The self-organization of speech sounds. Journal of Theoretical Biology, pp. 435–449. Cited by: §1.
- An online model for vowel imitation learning. Speech Communication, pp. 1–23. Cited by: §2, §4.4, §5.
- “LazImpa”: lazy and impatient neural agents learn to communicate efficiently. In Proc. ACL, pp. 335–343. Cited by: §1, §2.
- Coordinating perceptually grounded categories through language: a case study for colour. Behavioral and Brain Sciences, pp. 469–489. Cited by: §1.
- The synthetic modeling of language origins. Evolution of Communication, pp. 1–34. Cited by: §1.
- Reinforcement learning: an introduction. MIT Press. Cited by: §3.3.
- Multiagent cooperation and competition with deep reinforcement learning. PLOS ONE, pp. 1–15. Cited by: §3.3.
- Emergence of pragmatics from referential game between theory of mind agents. In Proc. NeurIPS, pp. . Cited by: §1.
Appendix A Experiment Details
a.1 General experimental setup
Here we provide the general setup for all experimentation.
|Runs (varying seed)||20|
|GPU||Nvidia RTX 2080 Super|
|Time (per run)||minutes|
a.2 Experiment parameters
Here we provide specific details on a per-experiment basis. The phone sequence length in the grounded experiments is chosen such that the full ground truth phonetic pronunciation could be made by the speaker agent.
|Experiment||Agent||Learning Rate||Phone length ()||GRU hidden size|