Learning Calibratable Policies using Programmatic Style-Consistency

10/02/2019 ∙ by Eric Zhan, et al. ∙ Microsoft California Institute of Technology 21

We study the important and challenging problem of controllable generation of long-term sequential behaviors. Solutions to this problem would impact many applications, such as calibrating behaviors of AI agents in games or predicting player trajectories in sports. In contrast to the well-studied areas of controllable generation of images, text, and speech, there are significant challenges that are unique to or exacerbated by generating long-term behaviors: how should we specify the factors of variation to control, and how can we ensure that the generated temporal behavior faithfully demonstrates diverse styles? In this paper, we leverage large amounts of raw behavioral data to learn policies that can be calibrated to generate a diverse range of behavior styles (e.g., aggressive versus passive play in sports). Inspired by recent work on leveraging programmatic labeling functions, we present a novel framework that combines imitation learning with data programming to learn style-calibratable policies. Our primary technical contribution is a formal notion of style-consistency as a learning objective, and its integration with conventional imitation learning approaches. We evaluate our framework using demonstrations from professional basketball players and agents in the MuJoCo physics environment, and show that our learned policies can be accurately calibrated to generate interesting behavior styles in both domains.



There are no comments yet.


page 1

page 2

page 3

page 4

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

1 Introduction

The widespread availability of recorded tracking data is enabling the study of complex behaviors in many domains, including sports (Chen et al., 2016a; Le et al., 2017; Zhan et al., 2019), video games (Kurin et al., 2017; Broll et al., 2019), laboratory animals (Eyjolfsdottir et al., 2014, 2017; Johnson et al., 2016), facial expressions (Suwajanakorn et al., 2017; Taylor et al., 2017), commonplace activities such as cooking (Nishimura et al., 2019), and driving (Bojarski et al., 2016; Chang et al., 2019). The tracking data is often obtained from multiple experts and can exhibit very diverse styles (e.g., aggressive versus passive play in sports). Our work is motivated by the opportunity to maximally leverage these datasets by cleanly extracting such styles in addition to modeling the raw behaviors. Our goal is to train policies that can be controlled, or calibrated, to produce different behavioral styles inherent in the demonstration data. For example, Figure 1a depicts demonstrations from real basketball players with variations of many types, including movement speed, desired destinations, tendencies for long versus short passes, and curvature of movement routes, amongst many others. A calibratable policy would be able to generate trajectories consistent with various styles, such as low movement speed as in Figure 1b, or approach the basket as in Figure 1c, or to both styles simultaneously as in Figure 1d. Importantly, we aim to train a single policy that can generate behaviors calibrated across multiple styles. Having such policies would empower many downstream tasks, including behavior discovery (Eyjolfsdottir et al., 2014), realistic simulations (Le et al., 2017), virtual agent design (Broll et al., 2019), and counterfactual behavioral reasoning (Zhan et al., 2019). We focus on three research questions. The first question is strategic: what systematic form of domain knowledge can we leverage to quickly and cleanly extract style information from raw behavioral data? The second question is formulaic: how can we formalize the learning objective to encourage learning style-calibratable policies? The third question is algorithmic: how do we design practical learning approaches that reliably optimize the learning objective? To address these challenges, we present a novel framework inspired by data programming (Ratner et al., 2016), a paradigm in weak supervision that utilizes automated labeling procedures, called labeling functions, to learn without ground-truth labels. In our setting, labeling functions enable domain experts to quickly translate domain knowledge of diverse styles into programmatically generated style annotations. For instance, it is trivial to write programmatic labeling functions for the two styles—speed and destination—depicted in Figure 1. Labeling functions also motivate a metric for learning, which we call programmatic style-consistency, to evaluate calibration of policies: rollouts generated for a specific style should return the same style label when fed to the labeling function. Finally, our framework is generic and is easily integrated into conventional imitation learning approaches. To summarize, our contributions are:

  • We propose a novel framework for learning policies calibrated to diverse behavior styles.

  • Our framework allows users to express styles as labeling functions, which can be quickly applied to programmatically produce a weak signal of style labels.

  • Our framework introduces style-consistency as a metric to evaluate calibration to styles.

  • We present an algorithm to learn calibratable policies that maximize style-consistency of the generated behaviors, and validate it in basketball and simulated physics environments.

(a) Expert demonstrations
(b) Style: SPEED
(d) Both styles
Figure 1: Basketball trajectories from policies that are: (a) the expert; (b) calibrated to move at low speeds; (c) calibrated to terminate near the basket (within green boundary); and (d) calibrated for both (b) & (c) simultaneously. Diamonds () and dots () indicate initial and final positions.

2 Background: Imitation Learning using Trajectory VAEs

Since our focus is on learning style-calibratable generative policies, for simplicity we develop our approach with the basic imitation learning paradigm of behavioral cloning using trajectory variational autoencoders, which we describe here. Interesting future directions include composing our approach with more advanced imitation learning approaches as well as with reinforcement learning.

Notation. Let and denote the environment state and action spaces. At each timestep , an agent observes state and executes action using a policy . The environment then transitions to the next state according to a (typically unknown) dynamics function . For the rest of this paper, we assume is deterministic; a modification of our approach for stochastic is included in Appendix B. A trajectory is a sequence of state-action pairs and the last state: . Let be a set of trajectories collected from expert demonstrations. In our experiments, each trajectory in has the same length , but in general this does not need to be the case. Learning objective. We begin with the basic imitation learning paradigm of behavioral cloning (Syed and Schapire, 2008). The goal is to learn a policy that behaves like the pre-collected demonstrations:



is a loss function that quantifies the mismatch between the actions chosen by

and those in the demonstrations. Since we are primarily interested in probabilistic or generative policies, we typically use (variations of) negative log-likelihood: , where

is the probability of

choosing action in state . Trajectory Variational Autoencoders. A common model choice for instantiating is the trajectory variational autoencoder (TVAE), which is a sequential generative model built on top of variational autoencoders (Kingma and Welling, 2014), and have been shown to work well in a range of generative policy learning settings (Wang et al., 2017; Ha and Eck, 2018; Co-Reyes et al., 2018). In its simplest form, a TVAE introduces a latent variable z (also called a trajectory embedding) with prior distribution , an encoder network , and a policy decoder . Its imitation learning objective is:


The main shortcoming of TVAEs and related approaches, which we address in Sections 3 & 4, is that the resulting policies cannot be easily calibrated to generate specific styles of behavior. For instance, the goal of the trajectory embedding z

is to capture all the styles that exist in the expert demonstrations, but there is no guarantee that the embeddings cleanly encode the desired styles in a calibrated way. Previous work has largely relied on unsupervised learning techniques that either require significant domain knowledge

(Le et al., 2017), or have trouble scaling to complex styles commonly found in real-world applications (Wang et al., 2017; Li et al., 2017).

3 Programmatic Style-consistency

Building upon the basic setup in Section 2, we focus on the setting where the demonstrations contain diverse behavior styles. To start, let denote a single style label (e.g., speed or destination, as shown in Figure 1). Our goal is to learn a policy that can be explicitly calibrated to y, i.e., trajectories generated by should match the demonstrations in that exhibit style y. Obtaining style labels can be expensive using conventional annotation methods, and unreliable using unsupervised approaches. We instead utilize easily programmable labeling functions that automatically produce style labels, described next. We then formalize a notion of style-consistency as a learning objective, and in Section 4 describe a practical learning approach. Labeling functions. Introduced in the data programming paradigm (Ratner et al., 2016), labeling functions programmatically produce weak and noisy labels to learn models on otherwise unlabeled datasets. A significant benefit is that labeling functions are often simple scripts that can be quickly applied to the dataset, which is much cheaper than manual annotations and more reliable than unsupervised methods. In our framework, we study behavior styles that can be represented as labeling functions, which we denote , that map trajectories to style labels y. A simple example is:


which distinguishes between trajectories with large (greater than a threshold ) versus small total displacement. We experiment with a range of labeling functions, as described in Section 6. Multiple labeling functions can be provided at once, possibly from multiple users. Many behavior styles used in previous work can be represented as labeling functions, e.g., agent speed (Wang et al., 2017). We use trajectory-level labels in our experiments, but in general labeling functions can be applied on subsequences to obtain per-timestep labels. We can efficiently annotate datasets using labeling functions, which we denote as . Our goal can now be phrased as: given , train a policy such that is calibrated to styles y found in . Style-consistency. A key insight in our work is that labeling functions naturally induce a metric for calibration. If a policy is calibrated to , we would expect the generated behaviors to be consistent with the label. So, we expect the following loss to be small:


where is a prior over the style labels, and is obtained by executing the style-conditioned policy in the environment. is thus a disagreement loss over labels that is minimized at , e.g., for categorical labels. We refer to (4) as the style-consistency loss, and say that is maximally calibrated to when (4) is minimized. Our full learning objective incorporating (4) with (1) is:


The simplest choice for the prior distribution is the marginal distribution of styles in . The first term in (5

) is a standard imitation learning objective and can be tractably estimated using

. To enforce style-consistency with the second term, conceptually we need to sample several , then several rollouts from the current policy, and query the labeling function for each of them. Furthermore, if is a non-differentiable function defined over the entire trajectory, as is the case in (3

), then we cannot simply backpropagate the style-consistency loss. In Section

4, we introduce differentiable approximations to more easily optimize the challenging objective in (5). Multiple styles. Our notion of style-consistency can be easily extended to simultaneously optimize for multiple styles. Suppose we have labeling functions and corresponding label spaces . Let denote and y denote . Then style-consistency becomes:


Note that style-consistency is optimized when the generated trajectory agrees with all labeling functions. Although this can be very challenging to achieve, it describes the most desirable outcome, i.e. is a policy that can be calibrated to all styles simultaneously.

4 Learning Approach

1:Input: demonstrations , labeling functions
2:construct by applying on trajectories in
3:optimize (7) to convergence to learn
4:optimize (8) to convergence to learn
Algorithm 1 Generic recipe for optimizing (5)

Optimizing (5) is challenging due to the long-time horizon and non-differentiability of the labeling functions .111This issue is not encountered in previous work on style-dependent imitation learning (Li et al., 2017; Hausman et al., 2017), since they use purely unsupervised methods such as maximizing mutual information. Given unlimited queries to the environment, one could naively employ model-free reinforcement learning, e.g., estimating (4) using rollouts and optimizing using policy gradient approaches. We instead take a model-based approach, described generically in Algorithm 1, that is more computationally-efficient and decomposable. The advantages of our approach are that it is compatible with batch or offline learning, and enables easier diagnosis of deficiencies in the algorithmic framework. To develop our approach, we first introduce a label approximator for , and then show how to optimize through the environmental dynamics using a differentiable model-based learning approach. Approximating labeling functions. To deal with non-differentiability of , we approximate it with a differentiable function parameterized by :


Here, is a differentiable loss that approximates , such as cross-entropy loss when is the loss. In our experiments we use a recurrent neural net to represent . We then modify the style-consistency term in (5) with and optimize:


Optimizing over trajectories. The next challenge to be addressed is one of credit assignment over time steps. For instance, consider the labeling function in (3) that computes the difference between the first and last states. Our label approximator may converge to a solution that ignores all inputs except for and . In this case, gradient descent through provides no information about intermediate timesteps. In other words, effective optimization of style-consistency in (8) requires informative learning signals on all actions taken by the policy. In general, there are two types of approaches to address this challenge: model-free and model-based. A model-free solution views this credit assignment challenge as analogous to that faced by RL, and repurposes generic reinforcement learning algorithms. We instead choose a model-based approach for two reasons: (a) we found it to be compositionally simpler and easier to debug; and (b) we can use the learned model to obtain hallucinated rollouts of the current policy efficiently during training. Modeling dynamics for credit assignment. Our model-based approach utilizes a dynamics model to approximate the environment’s dynamics by predicting the change in state given the current state and action:


where is often or squared- loss (Nagabandi et al., 2018; Luo et al., 2019). This allows us to generate trajectories by rolling out: . Then optimizing for style-consistency in (8) would backpropagate through our dynamics model and provide informative learning signals to the policy at every timestep. We outline our model-based approach in Algorithm 2. Lines 10-12 describe an optional step to fine-tune the dynamics model by querying the environment for trajectories of the current policy (similar to Luo et al. (2019)); we found that this can help improve style-consistency in some experiments.

1:Input: demonstrations , labeling function , label approximator , dynamics model
3:for  iterations do
4:     optimize (9) with batch from Train dynamics model
5:for  iterations do
6:     optimize (7) with batch from Train label approximator
7:for  iterations do
8:      { collect trajectories with and current policy }
9:     optimize (8) with batch from and Train policy
10:     for  iterations do
11:           collect 1 trajectory from environment with
12:          optimize (9) with Fine-tune dynamics model      
Algorithm 2 Model-based approach for optimizing style-consistency

5 Related Work

Our work combines ideas from imitation learning and data programming, developing a weakly supervised approach for more explicit and fine-grained calibration. This is related to learning disentangled representations and controllable generative modeling, reviewed below. Imitation learning of diverse behaviors has focused on unsupervised approaches to infer latent variables/codes that capture behavior styles (Li et al., 2017; Hausman et al., 2017; Wang et al., 2017). Similar approaches have also been studied for generating text conditioned on attributes such as sentiment or tense (Hu et al., 2017). A typical strategy is to maximize the mutual information between the latent codes and trajectories, in contrast to our notion of programmatic style-consistency. Disentangled representation learning aims to learn representations where each latent dimension corresponds to exactly one desired factor of variation (Bengio et al., 2012). Recent studies (Locatello et al., 2019) have noted that popular techniques (Chen et al., 2016b; Higgins et al., 2017; Kim and Mnih, 2018; Chen et al., 2018)

can be sensitive to hyperparameters and that evaluation metrics can be correlated with certain model classes and datasets, which suggests that unsupervised learning approaches may, in general, be unreliable for discovering cleanly calibratable representations.

Conditional generation for images has recently focused on attribute manipulation (Bao et al., 2017; Creswell et al., 2017; Klys et al., 2018)

, which aims to enforce that changing a label affects only one aspect of the image while keeping everything else the same (similar to disentangled representation learning). We extend these models and compare with our approach in Section 

6. Our experimental results suggest that these algorithms do not necessarily scale well into sequential domains. Enforcing consistency in generative modeling, such as cycle-consistency in image generation (Zhu et al., 2017), and self-consistency in hierarchical reinforcement learning (Co-Reyes et al., 2018) has proved beneficial. The former minimizes a discriminative disagreement, whereas the latter minimizes a distributional disagreement between two sets of generated behaviors (e.g., KL-divergence). From this perspective, our style-consistency notion is more similar to the former; however we also enforce consistency over multiple time-steps, which is more similar to the latter.

6 Experiments

We first briefly describe our experimental setup and choice of baselines, and then discuss our main experimental results. A full description of the experiments is available in Appendix C. Data. We validate our framework on two datasets: 1) a collection of professional basketball player trajectories with the goal of learning a policy that generates realistic player-movement, and 2) a Cheetah agent running horizontally in MuJoCo with the goal of learning a policy with calibrated gaits. The former has a known dynamics function: , where and are the player’s position and velocity on the court respectively; we expect the dynamics model to easily recover this function. The latter has an unknown dynamics function (which we learn a model of when approximating style-consistency). We obtain Cheetah demonstrations from a collection of policies that we trained using pytorch-a2c-ppo-acktr (Kostrikov, 2018) to interface with the DeepMind Control Suite’s Cheetah domain (Tassa et al., 2018)—see Appendix C for details. Labeling functions. Labeling functions for Basketball include: 1) average SPEED of the player, 2) DISPLACEMENT from initial to final position, 3) distance from final position to a fixed DESTINATION on the court (e.g. the basket), 4) mean DIRECTION of travel, and 5) CURVATURE of the trajectory, which measures the player’s propensity to change directions. For Cheetah, we have labeling functions for the agent’s 1) SPEED, 2) TORSO HEIGHT, 3) BACK-FOOT HEIGHT, and 4) FRONT-FOOT HEIGHT that can be trivially extracted from the environment. We threshold the aforementioned labeling functions into categorical labels (leaving real-valued labels for future work) and use (4) for style-consistency with as the loss. We use cross-entropy for and list all other hyperparameters in Appendix C. Whenever we report style-consistency results, we use in (4) so that all results are easily interpreted as accuracies. Baselines. We compare our approach, CTVAE-style, with 3 baseline policy models:

  1. CTVAE: The conditional version of TVAEs (Wang et al., 2017).

  2. CTVAE-info: CTVAE with information factorization (Creswell et al., 2017) that implicitly maximizes style-consistency by removing all information correlated with y from z.

  3. CTVAE-mi: CTVAE with mutual information maximization between style labels and trajectories. This is a supervised variant of unsupervised models (Chen et al., 2016b; Li et al., 2017), and also requires learning a dynamics model for sampling policy rollouts.

Detailed descriptions of baselines are in Appendix A, and model parameters are in Appendix C. Although all models build upon TVAEs, we emphasize that the underlying model choice is orthogonal to our contributions; our framework is compatible with any imitation learning algorithm.

6.1 How well can we calibrate policies for individual styles?

We first threshold labeling functions into 3 classes for Basketball and 2 classes for Cheetah; the marginal distribution of styles in is roughly uniform over these classes. Then we learn a policy calibrated to each of these styles. Finally, we generate rollouts from each of the learned policies to measure style-consistency. Table 1 compares the median style-consistency (over 5 seeds) of learned policies. For Basketball, CTVAE-style significantly outperforms baselines and achieves almost perfect style-consistency for 4 of the 5 styles (the best style-consistency over 5 seeds outperforms all baselines, shown in Tables 7(a) and 8(a) in Appendix C). For Cheetah, CTVAE-style outperforms all baselines, but the absolute performance is lower than for Basketball (mostly due to the more complex environment dynamics). We visualize our CTVAE-style policy calibrated for DESTINATION(net) (with style-consistency of 0.97) in Figure 2. The green boundaries divide the court into 3 regions, one for each label class. Policy rollouts almost always terminate in the corresponding region of the label class. Note that although the policy is calibrated for one style, rollouts still exhibit diverse behaviors (i.e. distribution of trajectories did not collapse into a single mode), which suggests that there are other styles being imitated. Section 6.2

examines this further by testing calibration to multiple styles simultaneously. We also consider cases in which labeling functions can have several classes and non-uniform distributions (i.e. some styles are more/less common than others). We threshold

DESTINATION(net) into 6 classes for Basketball and SPEED into 4 classes for Cheetah and compare the policies in Table 2. In general, we observe degradation in overall style-consistency accuracies as the number of classes increase. However, CTVAE-style policies still consistently achieve better style-consistency than baselines in this setting as well. In the appendix, we visualize all 6 classes of DESTINATION(net) in Figure 4 and include another experiment with up to 8 classes of DISPLACEMENT in Table 7(c). These results suggest that incorporating programmatic style-consistency while training via (8) can yield good qualitative and quantitative calibration results.

Basketball Cheetah
Model Speed Disp. Dest. Dir. Curve Speed Torso BFoot FFoot
CTVAE 83 72 82 77 61 59 63 68 68
CTVAE-info 84 71 79 72 60 57 63 65 66
CTVAE-mi 86 74 82 77 72 60 65 65 70
CTVAE-style 95 96 97 97 68 79 80 80 77
Table 1: Individual Style Calibration: Style-consistency (, median over 5 seeds) of policies evaluated with 4,000 Basketball and 500 Cheetah rollouts. Trained separately for each style, CTVAE-style policies outperform baselines for all styles in Cheetah and 4/5 styles in Basketball.
(a) Label class 0 (close)
(b) Label class 1 (mid)
(c) Label class 2 (far)
Figure 2: CTVAE-style rollouts calibrated for DESTINATION(net), 0.97 style-consistency. Diamonds () and dots () indicate initial and final positions. Regions divided by green lines represent label classes.
Basketball - DESTINATION(net) Cheetah - SPEED
Model 2 classes 3 classes 4 classes 6 classes 3 classes 4 classes
CTVAE 87 82 78 74 45 37
CTVAE-info 87 81 75 77 49 39
CTVAE-mi 88 81 74 76 48 37
CTVAE-style 98 97 89 84 59 51
Table 2: Fine-grained Style-consistency: (, median over 5 seeds) Training on labeling functions with more classes yields increasingly fine-grained calibration of behavior. Although CTVAE-style degrades as the number of classes increases, it outperforms baselines for all styles.
Basketball Cheetah
Model 2 styles 3 styles 4 styles 5 styles 2 styles 3 styles
CTVAE 71 58 50 37 41 28
CTVAE-info 69 58 51 32 41 27
CTVAE-mi 72 56 51 30 40 28
CTVAE-style 93 88 88 75 54 40
Table 3: Multi Style-consistency: (, median over 5 seeds) Simultaneously calibrated to multiple styles, CTVAE-style policies outperform baselines for all styles in Cheetah and in Basketball.

6.2 Can we calibrate policies for multiple styles simultaneously?

We now consider multiple style-consistency as in (6), which measures the total accuracy with all labeling functions simultaneously. For instance, in addition to terminating close to the net in Figure 2, a user may also want to control the speed at which the agent moves towards the target destination.

(a) Label class 0 (slow)
(b) Label class 1 (mid)
(c) Label class 2 (fast)
Figure 3: CTVAE-style rollouts calibrated for 2 styles: label class 1 of DESTINATION(net) (see Figure 2) and each class for SPEED, with 0.93 style-consistency. Diamonds () and dots () indicate initial and final positions.

Table 3 compares the style-consistency of policies calibrated for up to 5 styles for Basketball and 3 styles for Cheetah. Calibrating for multiple styles simultaneously is a very difficult task for baselines, as their style-consistency degrades significantly as the number of styles increases. On the other hand, CTVAE-style sees a modest decrease in style-consistency but is still significantly better calibrated (0.75 style-consistency for all 5 styles vs. only 0.30 for the best baseline in Basketball). We visualize CTVAE-style calibrated for two styles with style-consistency 0.93 in Basketball in Figure 3, while Figure 1(d) shows another example of calibration to multiple styles. CTVAE-style outperforms baselines in Cheetah as well, but there is still room for improvement to reach maximal style-consistency in future work.

6.3 What is the trade-off between style-consistency and imitation quality?

Basketball Cheetah
TVAE 2.5 -190 29 -119
CTVAE 2.5 -191 29 -118
CTVAE-info 2.3 -190 29 -115
CTVAE-mi 2.6 -191 29 -114
CTVAE-style 2.3 -188 30 -55
Table 4: Imitation loss for TVAE models (lower is better). CTVAE-style is comparable to baselines for Basketball, but is slightly worse for Cheetah.

In Table 4, we investigate whether CTVAE-style’s superior style-consistency is attained at a significant cost to imitation quality, since we jointly optimize both in (5). For Basketball, high style-consistency is achieved without any degradation in imitation quality. For Cheetah, negative log-likelihood is slightly worse; a followup experiment in Table 10 of the appendix shows that we can improve imitation quality with further training, which can sometimes modestly decrease style-consistency.

7 Conclusion and Future Work

We propose a novel framework for imitating diverse behavior styles while also calibrating to desired styles. Our framework leverages labeling functions to tractably represent styles and introduces programmatic style-consistency, a metric that allows for fair comparison between calibrated policies. Our experiments demonstrate strong empirical calibration results. We believe that our framework lays the foundation for many directions of future research. First, can one model more complex styles not easily captured with a single labeling function (e.g. aggressive vs. passive play in sports) by composing simpler labeling functions (e.g. max speed, distance to closest opponent, number of fouls committed, etc.), similar to (Ratner et al., 2016; Bach et al., 2017)? Second, can we use these per-timestep labels to model transient styles, or simplify the credit assignment problem when learning to calibrate? Third, can we blend our programmatic supervision with unsupervised learning approaches to arrive at effective semi-supervised solutions? Fourth, can we use leverage model-free approaches to further optimize self-consistency, e.g., to fine-tune from our model-based approach? Finally, can we integrate our framework with reinforcement learning to also optimize for environmental rewards?


  • S. H. Bach, B. D. He, A. Ratner, and C. Ré (2017) Learning the structure of generative models without labeled data. In

    International Conference on Machine Learning (ICML)

    Cited by: §7.
  • J. Bao, D. Chen, F. Wen, H. Li, and G. Hua (2017) CVAE-GAN: fine-grained image generation through asymmetric training. In

    IEEE International Conference on Computer Vision (ICCV)

    Cited by: §5.
  • Y. Bengio, A. C. Courville, and P. Vincent (2012)

    Unsupervised feature learning and deep learning: A review and new perspectives

    arXiv preprint arXiv:1206.5538. Cited by: §5.
  • M. Bojarski, D. Del Testa, D. Dworakowski, B. Firner, B. Flepp, P. Goyal, L. D. Jackel, M. Monfort, U. Muller, J. Zhang, et al. (2016) End to end learning for self-driving cars. arXiv preprint arXiv:1604.07316. Cited by: §1.
  • B. Broll, M. Hausknecht, D. Bignell, and A. Swaminathan (2019) Customizing scripted bots: sample efficient imitation learning for human-like behavior in minecraft. AAMAS Workshop on Adaptive and Learning Agents. Cited by: §1.
  • M. Chang, J. Lambert, P. Sangkloy, J. Singh, S. Bak, A. Hartnett, D. Wang, P. Carr, S. Lucey, D. Ramanan, et al. (2019) Argoverse: 3d tracking and forecasting with rich maps. In

    IEEE Conference on Computer Vision and Pattern Recognition (CVPR)

    Cited by: §1.
  • J. Chen, H. M. Le, P. Carr, Y. Yue, and J. J. Little (2016a)

    Learning online smooth predictors for realtime camera planning using recurrent decision trees

    In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4688–4696. Cited by: §1.
  • T. Q. Chen, X. Li, R. B. Grosse, and D. K. Duvenaud (2018) Isolating sources of disentanglement in variational autoencoders. In Neural Information Processing Systems (NeurIPS), Cited by: §5.
  • X. Chen, Y. Duan, R. Houthooft, J. Schulman, I. Sutskever, and P. Abbeel (2016b) InfoGAN: interpretable representation learning by information maximizing generative adversarial nets. In Neural Information Processing Systems (NeurIPS), Cited by: §5, item 3.
  • K. Cho, B. van Merrienboer, Ç. Gülçehre, F. Bougares, H. Schwenk, and Y. Bengio (2014) Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv preprint arXiv:1406.1078. Cited by: Appendix C.
  • J. D. Co-Reyes, Y. Liu, A. Gupta, B. Eysenbach, P. Abbeel, and S. Levine (2018) Self-consistent trajectory autoencoder: hierarchical reinforcement learning with trajectory embeddings. In International Conference on Machine Learning (ICML), Cited by: §2, §5.
  • A. Creswell, A. A. Bharath, and B. Sengupta (2017) Conditional autoencoders with adversarial information factorization. arXiv preprint arXiv:1711.05175. Cited by: Appendix A, §5, item 2.
  • E. Eyjolfsdottir, K. Branson, Y. Yue, and P. Perona (2017) Learning recurrent representations for hierarchical behavior modeling. In International Conference on Learning Representations (ICLR), Cited by: §1.
  • E. Eyjolfsdottir, S. Branson, X. P. Burgos-Artizzu, E. D. Hoopfer, J. Schor, D. J. Anderson, and P. Perona (2014) Detecting social actions of fruit flies. In European Conference on Computer Vision, pp. 772–787. Cited by: §1.
  • D. Ha and D. Eck (2018) A neural representation of sketch drawings. In International Conference on Learning Representations (ICLR), Cited by: §2.
  • K. Hausman, Y. Chebotar, S. Schaal, G. S. Sukhatme, and J. J. Lim (2017) Multi-modal imitation learning from unstructured demonstrations using generative adversarial nets. In Neural Information Processing Systems (NeurIPS), Cited by: Appendix A, §5, footnote 1.
  • I. Higgins, L. Matthey, A. Pal, C. Burgess, X. Glorot, M. Botvinick, S. Mohamed, and A. Lerchner (2017) Beta-vae: learning basic visual concepts with a constrained variational framework. In ICLR, Cited by: §5.
  • Z. Hu, Z. Yang, X. Liang, R. Salakhutdinov, and E. P. Xing (2017) Toward controlled generation of text. In International Conference on Machine Learning (ICML), Cited by: §5.
  • M. Johnson, D. K. Duvenaud, A. Wiltschko, R. P. Adams, and S. R. Datta (2016)

    Composing graphical models with neural networks for structured representations and fast inference

    In Advances in neural information processing systems, Cited by: §1.
  • H. Kim and A. Mnih (2018) Disentangling by factorising. In International Conference on Machine Learning (ICML), Cited by: §5.
  • D. P. Kingma and M. Welling (2014) Auto-encoding variational bayes. In International Conference on Learning Representations (ICLR), Cited by: Appendix B, §2.
  • J. Klys, J. Snell, and R. S. Zemel (2018) Learning latent subspaces in variational autoencoders. In Neural Information Processing Systems (NeurIPS), Cited by: Appendix A, §5.
  • I. Kostrikov (2018) PyTorch implementations of reinforcement learning algorithms. GitHub. Note: https://github.com/ikostrikov/pytorch-a2c-ppo-acktr-gail Cited by: Appendix C, §6.
  • V. Kurin, S. Nowozin, K. Hofmann, L. Beyer, and B. Leibe (2017) The atari grand challenge dataset. arXiv preprint arXiv:1705.10998. Cited by: §1.
  • H. M. Le, Y. Yue, P. Carr, and P. Lucey (2017) Coordinated multi-agent imitation learning. In International Conference on Machine Learning (ICML), Cited by: §1, §2.
  • Y. Li, J. Song, and S. Ermon (2017) InfoGAIL: interpretable imitation learning from visual demonstrations. In Neural Information Processing Systems (NeurIPS), Cited by: Appendix A, §2, §5, item 3, footnote 1.
  • F. Locatello, S. Bauer, M. Lucic, S. Gelly, B. Schölkopf, and O. Bachem (2019) Challenging common assumptions in the unsupervised learning of disentangled representations. In International Conference on Machine Learning (ICML), Cited by: §5.
  • Y. Luo, H. Xu, Y. Li, Y. Tian, T. Darrell, and T. Ma (2019) Algorithmic framework for model-based deep reinforcement learning with theoretical guarantees. In International Conference on Learning Representations (ICLR), Cited by: §4.
  • A. Nagabandi, G. Kahn, R. S. Fearing, and S. Levine (2018) Neural network dynamics for model-based deep reinforcement learning with model-free fine-tuning. In International Conference on Robotics and Automation (ICRA), Cited by: §4.
  • T. Nishimura, A. Hashimoto, Y. Yamakata, and S. Mori (2019) Frame selection for producing recipe with pictures from an execution video of a recipe. In Proceedings of the 11th Workshop on Multimedia for Cooking and Eating Activities, pp. 9–16. Cited by: §1.
  • A. Ratner, C. D. Sa, S. Wu, D. Selsam, and C. Ré (2016) Data programming: creating large training sets, quickly. In Neural Information Processing Systems (NeurIPS), Cited by: §1, §3, §7.
  • J. Schulman, F. Wolski, P. Dhariwal, A. Radford, and O. Klimov (2017) Proximal policy optimization algorithms. arXiv preprint arXiv:1707.06347. Cited by: Appendix C.
  • S. Suwajanakorn, S. M. Seitz, and I. Kemelmacher-Shlizerman (2017) Synthesizing obama: learning lip sync from audio. ACM Transactions on Graphics (TOG) 36 (4), pp. 95. Cited by: §1.
  • U. Syed and R. E. Schapire (2008) A game-theoretic approach to apprenticeship learning. In Advances in neural information processing systems, pp. 1449–1456. Cited by: §2.
  • Y. Tassa, Y. Doron, A. Muldal, T. Erez, Y. Li, D. de Las Casas, D. Budden, A. Abdolmaleki, J. Merel, A. Lefrancq, T. P. Lillicrap, and M. A. Riedmiller (2018) DeepMind control suite. arXiv preprint arXiv:1801.00690. Cited by: Appendix C, §6.
  • S. Taylor, T. Kim, Y. Yue, M. Mahler, J. Krahe, A. G. Rodriguez, J. Hodgins, and I. Matthews (2017) A deep learning approach for generalized speech animation. ACM Transactions on Graphics (TOG) 36 (4), pp. 93. Cited by: §1.
  • Z. Wang, J. Merel, S. Reed, G. Wayne, N. de Freitas, and N. Heess (2017) Robust imitation of diverse behaviors. In Neural Information Processing Systems (NeurIPS), Cited by: §2, §3, §5, item 1.
  • E. Zhan, S. Zheng, Y. Yue, L. Sha, and P. Lucey (2019) Generating multi-agent trajectories using programmatic weak supervision. In International Conference on Learning Representations (ICLR), Cited by: §1.
  • J. Zhu, T. Park, P. Isola, and A. A. Efros (2017)

    Unpaired image-to-image translation using cycle-consistent adversarial networks

    In Proceedings of the IEEE international conference on computer vision, pp. 2223–2232. Cited by: §5.

Appendix A Baseline Policy Models

1) Conditional-TVAE (CTVAE).

The conditional version of TVAEs optimizes:


2) CTVAE with information factorization (CTVAE-info).

(Creswell et al., 2017; Klys et al., 2018) augment conditional-VAE models with an auxiliary network which is trained to predict the label y from z, while the encoder is also trained to minimize the accuracy of . This model implicitly maximizes self-consistency by removing the information correlated with y from z, so that any information pertaining to y that the decoder needs for reconstruction must all come from y. While this model was previously used for image generation, we extend it into the sequential domain:


Note that the encoder in (10) and (11) differ in that is no longer conditioned on the label y.

3) CTVAE with mutual information maximization (CTVAE-mi).

In addition to (10), we can also maximize the mutual information between labels and trajectories . This quantity is hard to maximize directly, so instead we maximize the variational lower bound:


where approximates the true posterior . In our setting, the prior over labels is known, so is a constant. Thus, the learning objective is:


Optimizing (13) also requires collecting rollouts with the current policy, so similarly we also pretrain and fine-tune a dynamics model . This baseline can be interpreted as a supervised analogue of unsupervised models that maximize mutual information in (Li et al., 2017; Hausman et al., 2017).

Appendix B Stochastic Dynamics Function

If the dynamics function of the environment is stochastic, we modify our approach in Algorithm 2

by changing the form of our dynamics model. We can model the change in state as a Gaussian distribution and minimize the negative log-likelihood:


where , , , and , are neural networks that can share weights. We can sample a change in state during rollouts using the reparametrization trick (Kingma and Welling, 2014), which allows us to backpropagate through the dynamics model during training.

Appendix C Experiment Details

Dataset details.

See Table 5. For Cheetah, we train 125 policies using PPO (Schulman et al., 2017) to run forwards at speeds ranging from 0 to 4 (m/s). We collect 25 trajectories per policy by sampling actions from the policy. We use (Kostrikov, 2018) to interface with (Tassa et al., 2018).

Training hyperparameters.

See Table 6.

Model parameters.

We model all trajectory embeddings z as a diagonal Gaussian with a standard normal prior. Encoder and label approximators are bi-directional GRUs (Cho et al., 2014) followed by linear layers. Policy is recurrent for basketball, but not for Cheetah. The Gaussian log sigma returned by is state-dependent for basketball, but state-independent for Cheetah. For Cheetah, we made these choices based on prior work in Mujoco for training gait policies. For Basketball, we observed a lot more variation in the 500k demonstrations so we experimented with more flexible model classes. See Table 7 for more model details.

frequency (Hz)
Basketball 2 2 25 520,015 67,320 3
Cheetah 18 6 200 2,500 625 40
Table 5: Dataset parameters for basketball and Cheetah environments.
batch size # batch learning rate
Basketball 128 4,063 128 0
Cheetah 16 157 16 1
Table 6: Hyperparameters for Algorithm 2. is the number of batches to see all trajectories in the dataset once. We also use regularization of for training the dynamics model .
z-dim GRU GRU GRU sizes sizes
Basketball 4 128 128 128 (128,128) (128,128)
Cheetah 8 200 200 - (200,200) (500,500)
Table 7: Model parameters for basketball and Cheetah environments.
(a) Label class 0 (closest)
(b) Label class 1
(c) Label class 2
(d) Label class 3
(e) Label class 4
(f) Label class 5 (farthest)
Figure 4: Rollouts from our policy calibrated to DESTINATION(net) with 6 classes. The 5 green boundaries divide the court into 6 regions, each corresponding to a label class. The label indicates the target region of a trajectory’s final position (). This policy achieves a style-consistency of 0.93, as indicated in Table 7(b). Note that the initial position () is the same as in Figures 2 and 3 for comparison, but in general we sample an initial position from the prior to compute style-consistency.
Model Speed Displacement Destination Direction Curvature
CTVAE 82 83 85 71 72 74 81 82 82 76 77 80 60 61 62
CTVAE-info 84 84 87 69 71 74 78 79 83 71 72 74 60 60 62
CTVAE-mi 84 86 87 71 74 74 80 82 84 75 77 78 58 72 74
CTVAE-style 34 95 97 89 96 97 91 97 98 96 97 98 52 68 83
(a) Style-consistency wrt. single styles of 3 classes (roughly uniform distributions).
Model 2 classes 3 classes 4 classes 6 classes
CTVAE 86 87 87 80 82 83 76 78 79 70 74 77
CTVAE-info 83 87 88 79 81 83 73 75 78 71 77 78
CTVAE-mi 86 88 88 80 81 84 71 74 79 73 76 78
CTVAE-style 97 98 99 68 97 98 35 89 95 67 84 93
(b) Style-consistency wrt. DESTINATION(net) with up to 6 classes (non-uniform distributions).
Model 2 classes 3 classes 4 classes 6 classes 8 classes
CTVAE 91 92 93 79 83 84 76 79 79 68 70 72 64 66 69
CTVAE-info 90 90 92 83 83 85 75 76 77 68 70 72 60 63 67
CTVAE-mi 90 92 93 81 84 86 75 77 80 66 70 72 62 62 67
CTVAE-style 98 99 99 15 98 99 15 96 96 02 92 94 80 90 93
(c) Style-consistency wrt. DISPLACEMENT of up to 8 classes (roughly uniform distributions).
Model 2 styles 3 styles 4 styles 5 styles
CTVAE 67 71 73 58 58 62 49 50 52 27 37 35
CTVAE-info 68 69 70 54 58 59 48 51 54 28 32 35
CTVAE-mi 71 72 73 48 56 61 45 51 52 16 30 31
CTVAE-style 92 93 94 86 88 90 62 88 88 66 75 80
(d) Style-consistency wrt. multiple styles simultaneously.
Table 8: [min, median, max] style-consistency (, 5 seeds) of policies evaluated with 4,000 basketball rollouts each. CTVAE-style policies significantly outperform baselines in all experiments and are calibrated at almost maximal style-consistency for 4/5 labeling functions. We note some rare failure cases with our approach, which we leave as a direction for improvement for future work.
Model Speed Torso Height B-Foot Height F-Foot Height
CTVAE 53 59 62 62 63 70 61 68 73 63 68 72
CTVAE-info 56 57 61 62 63 72 58 65 72 63 66 69
CTVAE-mi 53 60 62 62 65 70 60 65 70 66 70 73
CTVAE-style 68 79 81 79 80 84 77 80 88 74 77 80
(a) Style-consistency wrt. single styles of 2 classes (roughly uniform distributions).
Model 3 classes 4 classes
CTVAE 41 45 49 35 37 41
CTVAE-info 47 49 52 36 39 42
CTVAE-mi 47 48 53 36 37 38
CTVAE-style 59 59 65 42 51 60
(b) Style-consistency wrt. SPEED with varying # of classes (non-uniform distributions).
Model 2 styles 3 styles
CTVAE 39 41 43 25 28 29
CTVAE-info 39 41 46 25 27 30
CTVAE-mi 34 40 48 27 28 31
CTVAE-style 43 54 60 38 40 52
(c) Style-consistency wrt. multiple styles simultaneously.
Table 9: [min, median, max] style-consistency (, 5 seeds) of policies evaluated with 500 Cheetah rollouts each. CTVAE-style policies consistently outperform all baselines, but we note that there is still room for improvement (to reach 100% style-consistency).
Speed Torso Height B-Foot Height F-Foot Height
CTVAE-style 30 -55 79 29 -47 80 28 -31 80 29 -43 77
CTVAE-style+ 30 -97 70 30 -84 83 29 -72 80 29 -84 74
Table 10: We report the median imitation loss (KL, NLL, lower is better) and style-consistency (higher is better) of CTVAE-style policies for Cheetah (5 seeds). The first row corresponds to experiments in Tables 1 and 8(a), and the second row corresponds to same experiments with 50% more training iterations. Although imitation quality improves, style-consistency can sometimes degrade (e.g. SPEED, FRONT-FOOT HEIGHT), indicating a possible trade-off between imitation quality and style-consistency.