1 Introduction
Neural networks are bringing incredible performance gains on text classification tasks (Howard and Ruder, 2018; Peters et al., 2018; Devlin et al., 2019). However, this power comes hand in hand with a desire for more interpretability, even though its definition may differ (Lipton, 2016). While it is useful to obtain high classification accuracy, with more data available than ever before it also becomes increasingly important to justify predictions. Imagine having to classify a large collection of documents, while verifying that the classifications make sense. It would be extremely timeconsuming to read each document to evaluate the results. Moreover, if we do not know why a prediction was made, we do not know if we can trust it.
What if the model could provide us the most important parts of the document, as a justification for its prediction? That is exactly the focus of this paper. We use a setting that was pioneered by Lei et al. (2016). A rationale is defined to be a short yet sufficient part of the input text; short so that it makes clear what is most important, and sufficient so that a correct prediction can be made from the rationale alone. One neural network learns to extract the rationale, while another neural network, with separate parameters, learns to make a prediction from just the rationale. Lei et al. model this by assigning a binary Bernoulli variable to each input word. The rationale then consists of all the words for which a 1 was sampled. Because gradients do not flow through discrete samples, the rationale extractor is optimized using REINFORCE (Williams, 1992). An regularizer is used to make sure the rationale is short.
We propose an alternative to purely discrete selectors for which gradient estimation is possible without REINFORCE, instead relying on a reparameterization of a random variable that exhibits both continuous and discrete behavior
(Louizos et al., 2017). To promote compact rationales, we employ a relaxed form of regularization (Louizos et al., 2017), penalizing the objective as a function of the expected proportion of selected text. We also propose the use of Lagrangian relaxation to target a specific rate of selected input text.Our contributions are summarized as follows:^{1}^{1}1Code available at https://github.com/bastings/interpretable_predictions.
2 Latent Rationale
We are interested in making NNbased text classifiers interpretable by (i) uncovering which parts of the input text contribute features for classification, and (ii) basing decisions on only a fraction of the input text (a rationale). Lei et al. (2016) approached (i) by inducing binary latent selectors that control which input positions are available to an NN encoder that learns features for classification/regression, and (ii) by regularising their architectures using sparsityinducing penalties on latent assignments. In this section we put their approach under a probabilistic light, and this will then more naturally lead to our proposed method.
In text classification, an input is mapped to a distribution over target labels:
(1) 
where we have a neural network architecture parameterize the model— collectively denotes the parameters of the NN layers in
. That is, an NN maps from data space (e.g. sentences, short paragraphs, or premisehypothesis pairs) to the categorical parameter space (i.e. a vector of class probabilities). For the sake of concreteness, consider the input a sequence
. A target is typically a categorical outcome, such as a sentiment class or an entailment decision, but with an appropriate choice of likelihood it could also be a numerical score (continuous or integer).Lei et al. (2016) augment this model with a collection of latent variables which we denote by . These variables are responsible for regulating which portions of the input contribute with predictors (i.e. features) to the classifier. The model formulation changes as follows:
(2)  
where an NN predicts a sequence of Bernoulli parameters—one per latent variable—and the classifier is modified such that indicates whether or not is available for encoding. We can think of the sequence as a binary gating mechanism used to select a rationale, which with some abuse of notation we denote by . Figure 1 illustrates the approach.
Parameter estimation for this model can be done by maximizing a lower bound on the loglikelihood of the data derived by application of Jensen’s inequality:^{2}^{2}2This can be seen as variational inference (Jordan et al., 1999) where we perform approximate inference using a datadependent prior .
(3)  
These latent rationales approach the first objective, namely, uncovering which parts of the input text contribute towards a decision. However note that an NN controls the Bernoulli parameters, thus nothing prevents this NN from selecting the whole of the input, thus defaulting to a standard text classifier. To promote compact rationales, Lei et al. (2016) impose sparsityinducing penalties on latent selectors. They penalise for the total number of selected words, in (4), as well as, for the total number of transitions, fused lasso in (4), and approach the following optimization problem
(4) 
via gradientbased optimisation, where and
are fixed hyperparameters. The objective is however intractable to compute, the lowerbound, in particular, requires marginalization of
binary sequences. For that reason, Lei et al. sample latent assignments and work with gradient estimates using REINFORCE (Williams, 1992).The key ingredients are, therefore, binary latent variables and sparsityinducing regularization, and therefore the solution is marked by nondifferentiability. We propose to replace Bernoulli variables by rectified continuous random variables
(Socci et al., 1998), for they exhibit both discrete and continuous behaviour. Moreover, they are amenable to reparameterization in terms of a fixed random source (Kingma and Welling, 2014), in which case gradient estimation is possible without REINFORCE. Following Louizos et al. (2017), we exploit one such distribution to relax regularization and thus promote compact rationales with a differentiable objective. In section 3, we introduce this distribution and present its properties. In section 4, we employ a Lagrangian relaxation to automatically target a prespecified selection rate. And finally, in section 5 we present an example for sentiment classification.3 Hard Kumaraswamy Distribution
Key to our model is a novel distribution that exhibits both continuous and discrete behaviour, in this section we introduce it. With nonnegligible probability, samples from this distribution evaluate to exactly or exactly . In a nutshell: i) we start from a distribution over the open interval (see dashed curve in Figure 2); ii) we then stretch its support from to in order to include and (see solid curve in Figure 2); finally, iii) we collapse the probability mass over the interval to , and similarly, the probability mass over the interval to (shaded areas in Figure 2). This stretchandrectify technique was proposed by Louizos et al. (2017), who rectified samples from the BinaryConcrete (or GumbelSoftmax) distribution (Maddison et al., 2017; Jang et al., 2017)
. We adapted their technique to the Kumaraswamy distribution motivated by its close resemblance to a Beta distribution, for which we have stronger intuitions (for example, its two shape parameters transit rather naturally from unimodal to bimodal configurations of the distribution). In the following, we introduce this new distribution formally.
^{3}^{3}3We use uppercase letters for random variables (e.g. , , and ) and lowercase for assignments (e.g. , , ). For a random variable ,is the probability density function (pdf), conditioned on parameters
, andis the cumulative distribution function (cdf).
3.1 Kumaraswamy distribution
The Kumaraswamy distribution (Kumaraswamy, 1980) is a twoparameters distribution over the open interval , we denote a Kumaraswamydistributed variable by , where and control the distribution’s shape. The dashed curve in Figure 2 illustrates the density of . For more details including its pdf and cdf, consult Appendix A.
The Kumaraswamy is a close relative of the Beta distribution, though not itself an exponential family, with a simple cdf whose inverse
(5) 
for , can be used to obtain samples
(6) 
by transformation of a uniform random source . We can use this fact to reparameterize expectations (Nalisnick and Smyth, 2016).
3.2 Rectified Kumaraswamy
We stretch the support of the Kumaraswamy distribution to include and . The resulting variable takes on values in the open interval where and , with cdf
(7) 
We now define a rectified random variable, denoted by , by passing a sample through a hardsigmoid, i.e. . The resulting variable is defined over the closed interval . Note that while there is probability of sampling , sampling corresponds to sampling any , a set whose mass under is available in closed form:
(8) 
That is because all negative values of are deterministically mapped to zero. Similarly, samples are all deterministically mapped to , whose total mass amounts to
(9) 
See Figure 2 for an illustration, and Appendix A for the complete derivations.
3.3 Reparameterization and gradients
Because this rectified variable is built upon a Kumaraswamy, it admits a reparameterisation in terms of a uniform variable . We need to first sample a uniform variable in the open interval and transform the result to a Kumaraswamy variable via the inverse cdf (10a), then shift and scale the result to cover the stretched support (10b), and finally, apply the rectifier in order to get a sample in the closed interval (10c).
(10a)  
(10b)  
(10c) 
We denote this for short. Note that this transformation has two discontinuity points, namely, and . Though recall, the probability of sampling exactly or exactly is zero, which essentially means stochasticity circumvents points of nondifferentiability of the rectifier (see Appendix A.3).
4 Controlled Sparsity
Following Louizos et al. (2017), we relax nondifferentiable penalties by computing them on expectation under our latent model . In addition, we propose the use of Lagrangian relaxation to target specific values for the penalties. Thanks to the tractable Kumaraswamy cdf, the expected value of is known in closed form
(11)  
where . This quantity is a tractable and differentiable function of the parameters of the latent model. We can also compute a relaxation of fused lasso by computing the expected number of zerotononzero and nonzerotozero changes:
(12)  
In both cases, we make the assumption that latent variables are independent given , in Appendix B.1.2 we discuss how to estimate the regularizers for a model that conditions on the prefix of sampled assignments.
We can use regularizers to promote sparsity, but just how much text will our final model select? Ideally, we would target specific values and solve a constrained optimization problem. In practice, constrained optimisation is very challenging, thus we employ Lagrangian relaxation instead:
(13) 
where is a vector of regularisers, e.g. expected and expected fused lasso, and is a vector of Lagrangian multipliers . Note how this differs from the treatment of Lei et al. (2016) shown in (4) where regularizers are computed for assignments, rather than on expectation, and where are fixed hyperparameters.
5 Sentiment Classification
As a concrete example, consider the case of sentiment classification where is a sentence and is a way sentiment class (from very negative to very positive). The model consists of
(14)  
where the shape parameters , i.e. two sequences of strictly positive scalars, are predicted by a NN, and the support boundaries are fixed hyperparameters.
We first specify an architecture that parameterizes latent selectors and then use a reparameterized sample to restrict which parts of the input contribute encodings for classification:^{4}^{4}4We describe architectures using blocks denoted by , boldface letters for vectors, and the shorthand for a sequence .
where is an embedding layer, is a bidirectional encoder, and are feedforward transformations with outputs, and turns the uniform sample into the latent selector (see §3). We then use the sampled to modulate inputs to the classifier:
where and are recurrent cells such as LSTMs (Hochreiter and Schmidhuber, 1997) that process the sequence in different directions, and is a feedforward transformation with output. Note how modulates features of the input that are available to the recurrent composition function.
We then obtain gradient estimates of via Monte Carlo (MC) sampling from
(15) 
where is a shorthand for elementwise application of the transformation from uniform samples to HardKuma samples. This reparameterisation is the key to gradient estimation through stochastic computation graphs (Kingma and Welling, 2014; Rezende et al., 2014).
Deterministic predictions.
At test time we make predictions based on what is the most likely assignment for each . We across configurations of the distribution, namely, , , or . When the continuous interval is more likely, we take the expected value of the underlying Kumaraswamy variable.
6 Experiments
We perform experiments on multiaspect sentiment analysis to compare with previous work, as well as experiments on sentiment classification and natural language inference. All models were implemented in PyTorch, and Appendix
B provides implementation details.Goal.
When rationalizing predictions, our goal is to perform as well as systems using the full input text, while using only a subset of the input text, leaving unnecessary words out for interpretability.
6.1 Multiaspect Sentiment Analysis
Method  Look  Smell  Taste  

% Precision  % Selected  % Precision  % Selected  % Precision  % Selected  
Attention (Lei et al.)  80.6  13  88.4  7  65.3  7 
Bernoulli (Lei et al.)  96.3  14  95.1  7  80.2  7 
Bernoulli (reimpl.)  94.8  13  95.1  7  80.5  7 
HardKuma  98.1  13  96.8  7  89.8  7 
In our first experiment we compare directly with previous work on rationalizing predictions (Lei et al., 2016). We replicate their setting.
Data.
A preprocessed subset of the BeerAdvocate^{5}^{5}5https://www.beeradvocate.com/ data set is used (McAuley et al., 2012). It consists of 220,000 beer reviews, where multiple aspects (e.g. look, smell, taste) are rated. As shown in Figure 1, a review typically consists of multiple sentences, and contains a 05 star rating (e.g. 3.5 stars) for each aspect. Lei et al. mapped the ratings to scalars in .
Model.
We use the models described in §5 with two small modifications: 1) since this is a regression task, we use a sigmoid activation in the output layer of the classifier rather than a softmax,^{6}^{6}6
From a likelihood learning point of view, we would have assumed a LogitNormal likelihood, however, to stay closer to
Lei et al. (2016), we employ mean squared error. and 2) we use an extra RNN to condition on :(16a)  
(16b)  
(16c) 
For a fair comparison we follow Lei et al. by using RCNN^{7}^{7}7An RCNN cell can replace any LSTM cell and works well on text classification problems. See appendix B. cells rather than LSTM cells for encoding sentences on this task. Since this cell is not widely used, we verified its performance in Table 1. We observe that the BiRCNN performs on par with the BiLSTM (while using 50% fewer parameters), and similarly to previous results.
Evaluation.
A test set with sentencelevel rationale annotations is available. The precision of a rationale is defined as the percentage of words with that is part of the annotation. We also evaluate the predictions made from the rationale using mean squared error (MSE).
Baselines.
For our baseline we reimplemented the approach of Lei et al. (2016) which we call Bernoulli after the distribution they use to sample from. We also report their attention baseline, in which an attention score is computed for each word, after which it is simply thresholded to select the topk percent as the rationale.
Results.
Table 2 shows the precision and the percentage of selected words for the first three aspects. The models here have been selected based on validation MSE and were tuned to select a similar percentage of words (‘selected’). We observe that our Bernoulli reimplementation reaches the precision similar to previous work, doing a little bit worse for the ‘look’ aspect. Our HardKuma managed to get even higher precision, and it extracted exactly the percentage of text that we specified (see §4).^{8}^{8}8We tried to use Lagrangian relaxation for the Bernoulli model, but this led to instabilities (e.g. all words selected). Figure 3 shows the MSE for all aspects for various percentages of extracted text. We observe that HardKuma does better with a smaller percentage of text selected. The performance becomes more similar as more text is selected.
6.2 Sentiment Classification
We also experiment on the Stanford Sentiment Treebank (SST) (Socher et al., 2013). There are 5 sentiment classes: very negative, negative, neutral, positive, and very positive. Here we use the HardKuma model described in §5, a Bernoulli model trained with REINFORCE, as well as a BiLSTM.
Results.
Figure 4 shows the classification accuracy for various percentages of selected text. We observe that HardKuma outperforms the Bernoulli model at each percentage of selected text. HardKuma reaches fulltext baseline performance already around 40% extracted text. At that point, it obtains a test score of 45.84, versus 42.22 for Bernoulli and 47.4 for the fulltext baseline.
Analysis.
We wonder what kind of words are dropped when we select smaller amounts of text. For this analysis we exploit the wordlevel sentiment annotations in SST, which allows us to track the sentiment of words in the rationale. Figure 5 shows that a large portion of dropped words have neutral sentiment, and it seems plausible that exactly those words are not important features for classification. We also see that HardKuma drops (relatively) more neutral words than Bernoulli.
6.3 Natural Language Inference
In Natural language inference (NLI), given a premise sentence and a hypothesis sentence , the goal is to predict their relation which can be contradiction, entailment, or neutral. As our dataset we use the Stanford Natural Language Inference (SNLI) corpus (Bowman et al., 2015).
Baseline.
We use the Decomposable Attention model (DA) of
Parikh et al. (2016).^{9}^{9}9Better results e.g. Chen et al. (2017) and data sets for NLI exist, but are not the focus of this paper. DA does not make use of LSTMs, but rather uses attention to find connections between the premise and the hypothesis that are predictive of the relation. Each word in the premise attends to each word in the hypothesis, and vice versa, resulting in a set of comparison vectors which are then aggregated for a final prediction. If there is no link between a word pair, it is not considered for prediction.Model.
Because the premise and hypothesis interact, it does not make sense to extract a rationale for the premise and hypothesis independently. Instead, we replace the attention between premise and hypothesis with HardKuma attention. Whereas in the baseline a similarity matrix is normalized across rows (premise to hypothesis) and columns (hypothesis to premise) to produce attention matrices, in our model each cell in the attention matrix is sampled from a HardKuma parameterized by . To promote sparsity, we use the relaxed to specify the desired percentage of nonzero attention cells. The resulting matrix does not need further normalization.
Results.
With a target rate of 10%, the HardKuma model achieved 8.5% nonzero attention. Table 3 shows that, even with so many zeros in the attention matrices, it only does about 1% worse compared to the DA baseline. Figure 6 shows an example of HardKuma attention, with additional examples in Appendix B. We leave further explorations with HardKuma attention for future work.
7 Related Work
This work has connections with work on interpretability, learning from rationales, sparse structures, and rectified distributions. We discuss each of those areas.
Interpretability.
Machine learning research has been focusing more and more on interpretability (Gilpin et al., 2018). However, there are many nuances to interpretability (Lipton, 2016), and amongst them we focus on model transparency.
One strategy is to extract a simpler, interpretable model from a neural network, though this comes at the cost of performance. For example, Thrun (1995) extract ifthen rules, while Craven and Shavlik (1996)
extract decision trees.
There is also work on making word vectors more interpretable. Faruqui et al. (2015) make word vectors more sparse, and Herbelot and Vecchi (2015) learn to map distributional word vectors to modeltheoretic semantic vectors.
Similarly to Lei et al. (2016), Titov and McDonald (2008) extract informative fragments of text by jointly training a classifier and a model predicting a stochastic mask, while relying on Gibbs sampling to do so. Their focus is on using the sentiment labels as a weak supervision signal for opinion summarization rather than on rationalizing classifier predictions.
There are also related approaches that aim to interpret an alreadytrained model, in contrast to Lei et al. (2016) and our approach where the rationale is jointly modeled. Ribeiro et al. (2016) make any classifier interpretable by approximating it locally with a linear proxy model in an approach called LIME, and AlvarezMelis and Jaakkola (2017) propose a framework that returns inputoutput pairs that are causally related.
Learning from rationales.
Our work is different from approaches that aim to improve classification using rationales as an additional input (Zaidan et al., 2007; Zaidan and Eisner, 2008; Zhang et al., 2016). Instead, our rationales are latent and we are interested in uncovering them. We only use annotated rationales for evaluation.
Sparse layers.
Also arguing for enhanced interpretability, Niculae and Blondel (2017) propose a framework for learning sparsely activated attention layers based on smoothing the operator. They derive a number of relaxations to , including itself, but in particular, they target relaxations such as (Martins and Astudillo, 2016) which, unlike , are sparse (i.e. produce vectors of probability values with components that evaluate to exactly
). Their activation functions are themselves solutions to convex optimization problems, to which they provide efficient forward and backward passes. The technique can be seen as a deterministic sparsely activated layer which they use as a dropin replacement to standard attention mechanisms. In contrast, in this paper we focus on binary outcomes rather than
valued ones. Niculae et al. (2018) extend the framework to structured discrete spaces where they learn sparse parameterizations of discrete latent models. In this context, parameter estimation requires exact marginalization of discrete variables or gradient estimation via REINFORCE. They show that oftentimes distributions are sparse enough to enable exact marginal inference.Peng et al. (2018) propose SPIGOT, a proxy gradient to the nondifferentiable operator. This proxy requires an solver (e.g. Viterbi for structured prediction) and, like the straightthrough estimator (Bengio et al., 2013)
, is a biased estimator. Though, unlike ST it is efficient for structured variables. In contrast, in this work we chose to focus on unbiased estimators.
Rectified Distributions.
The idea of rectified distributions has been around for some time. The rectified Gaussian distribution
(Socci et al., 1998), in particular, has found applications to factor analysis (Harva and Kaban, 2005) and approximate inference in graphical models (Winn and Bishop, 2005). Louizos et al. (2017) propose to stretch and rectify samples from the BinaryConcrete (or GumbelSoftmax) distribution (Maddison et al., 2017; Jang et al., 2017). They use rectified variables to induce sparsity in parameter space via a relaxation to . We adapt their technique to promote sparse activations instead. Rolfe (2017)learns a relaxation of a discrete random variable based on a tractable mixture of a point mass at zero and a continuous reparameterizable density, thus enabling reparameterized sampling from the halfclosed interval
. In contrast, with HardKuma we focused on giving support to both 0s and 1s.8 Conclusions
We presented a differentiable approach to extractive rationales, including an objective that allows for specifying how much text is to be extracted. To allow for reparameterized gradient estimates and support for binary outcomes we introduced the HardKuma distribution. Apart from extracting rationales, we showed that HardKuma has further potential uses, which we demonstrated on premisehypothesis attention in SNLI. We leave further explorations for future work.
Acknowledgments
We thank Luca Falorsi for pointing us to Louizos et al. (2017), which inspired the HardKumaraswamy distribution. This work has received funding from the European Research Council (ERC StG BroadSem 678254), the European Union’s Horizon 2020 research and innovation programme (grant agreement No 825299, GoURMET), and the Dutch National Science Foundation (NWO VIDI 639.022.518, NWO VICI 27789002).
References

AlvarezMelis and Jaakkola (2017)
David AlvarezMelis and Tommi Jaakkola. 2017.
A causal framework for
explaining the predictions of blackbox sequencetosequence models.
In
Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing
, pages 412–421. Association for Computational Linguistics.  Bengio et al. (2013) Yoshua Bengio, Nicholas Léonard, and Aaron Courville. 2013. Estimating or propagating gradients through stochastic neurons for conditional computation. arXiv preprint arXiv:1308.3432.
 Bowman et al. (2015) Samuel R. Bowman, Gabor Angeli, Christopher Potts, and Christopher D. Manning. 2015. A large annotated corpus for learning natural language inference. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pages 632–642. Association for Computational Linguistics.
 Bowman et al. (2016) Samuel R. Bowman, Jon Gauthier, Abhinav Rastogi, Raghav Gupta, Christopher D. Manning, and Christopher Potts. 2016. A fast unified model for parsing and sentence understanding. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1466–1477. Association for Computational Linguistics.
 Chen et al. (2017) Qian Chen, Xiaodan Zhu, ZhenHua Ling, Si Wei, Hui Jiang, and Diana Inkpen. 2017. Enhanced lstm for natural language inference. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1657–1668. Association for Computational Linguistics.
 Craven and Shavlik (1996) Mark Craven and Jude W Shavlik. 1996. Extracting treestructured representations of trained networks. In Advances in neural information processing systems, pages 24–30.
 Devlin et al. (2019) Jacob Devlin, MingWei Chang, Kenton Lee, and Kristina Toutanova. 2019. Bert: Pretraining of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers). Association for Computational Linguistics.
 Faruqui et al. (2015) Manaal Faruqui, Yulia Tsvetkov, Dani Yogatama, Chris Dyer, and Noah A. Smith. 2015. Sparse overcomplete word vector representations. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 1491–1500. Association for Computational Linguistics.

Gilpin et al. (2018)
Leilani H Gilpin, David Bau, Ben Z Yuan, Ayesha Bajwa, Michael Specter, and
Lalana Kagal. 2018.
Explaining explanations: An overview of interpretability of machine
learning.
In
2018 IEEE 5th International Conference on Data Science and Advanced Analytics (DSAA)
, pages 80–89. IEEE.  Harva and Kaban (2005) Markus Harva and Ata Kaban. 2005. A variational bayesian method for rectified factor analysis. In Proceedings. 2005 IEEE International Joint Conference on Neural Networks, 2005., volume 1, pages 185–190. IEEE.
 Herbelot and Vecchi (2015) Aurélie Herbelot and Eva Maria Vecchi. 2015. Building a shared world: mapping distributional to modeltheoretic semantic spaces. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pages 22–32. Association for Computational Linguistics.
 Hochreiter and Schmidhuber (1997) Sepp Hochreiter and Jürgen Schmidhuber. 1997. Long ShortTerm Memory. Neural Computation, 9(8):1735–1780.
 Howard and Ruder (2018) Jeremy Howard and Sebastian Ruder. 2018. Universal language model finetuning for text classification. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 328–339. Association for Computational Linguistics.
 Jang et al. (2017) Eric Jang, Shixiang Gu, and Ben Poole. 2017. Categorical reparameterization with gumbelsoftmax. International Conference on Learning Representations.
 Jordan et al. (1999) MichaelI. Jordan, Zoubin Ghahramani, TommiS. Jaakkola, and LawrenceK. Saul. 1999. An introduction to variational methods for graphical models. Machine Learning, 37(2):183–233.
 Kingma and Welling (2014) Diederik P. Kingma and Max Welling. 2014. Autoencoding variational bayes. In International Conference on Learning Representations.
 Kumaraswamy (1980) Ponnambalam Kumaraswamy. 1980. A generalized probability density function for doublebounded random processes. Journal of Hydrology, 46(12):79–88.
 Lei et al. (2016) Tao Lei, Regina Barzilay, and Tommi Jaakkola. 2016. Rationalizing neural predictions. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pages 107–117. Association for Computational Linguistics.
 Lipton (2016) Zachary Chase Lipton. 2016. The mythos of model interpretability. ICML Workshop on Human Interpretability in Machine Learning (WHI 2016).
 Louizos et al. (2017) Christos Louizos, Max Welling, and Diederik P Kingma. 2017. Learning sparse neural networks through regularization. arXiv preprint arXiv:1712.01312.
 Maddison et al. (2017) Chris J. Maddison, Andriy Mnih, and Yee Whye Teh. 2017. The concrete distribution: A continous relaxation of discrete random variables. International Conference on Learning Representations.
 Martins and Astudillo (2016) Andre Martins and Ramon Astudillo. 2016. From softmax to sparsemax: A sparse model of attention and multilabel classification. In International Conference on Machine Learning, pages 1614–1623.
 McAuley et al. (2012) Julian McAuley, Jure Leskovec, and Dan Jurafsky. 2012. Learning attitudes and attributes from multiaspect reviews. In Data Mining (ICDM), 2012 IEEE 12th International Conference on, pages 1020–1025. IEEE.
 Nalisnick and Smyth (2016) Eric Nalisnick and Padhraic Smyth. 2016. Stickbreaking variational autoencoders. arXiv preprint arXiv:1605.06197.
 Niculae and Blondel (2017) Vlad Niculae and Mathieu Blondel. 2017. A regularized framework for sparse and structured neural attention. In Advances in Neural Information Processing Systems, pages 3338–3348.
 Niculae et al. (2018) Vlad Niculae, André F. T. Martins, and Claire Cardie. 2018. Towards dynamic computation graphs via sparse latent structure. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 905–911. Association for Computational Linguistics.
 Parikh et al. (2016) Ankur Parikh, Oscar Täckström, Dipanjan Das, and Jakob Uszkoreit. 2016. A decomposable attention model for natural language inference. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pages 2249–2255. Association for Computational Linguistics.
 Peng et al. (2018) Hao Peng, Sam Thomson, and Noah A. Smith. 2018. Backpropagating through structured argmax using a spigot. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1863–1873. Association for Computational Linguistics.
 Peters et al. (2018) Matthew Peters, Mark Neumann, Mohit Iyyer, Matt Gardner, Christopher Clark, Kenton Lee, and Luke Zettlemoyer. 2018. Deep contextualized word representations. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 2227–2237. Association for Computational Linguistics.
 Rezende et al. (2014) Danilo Jimenez Rezende, Shakir Mohamed, and Daan Wierstra. 2014. Stochastic backpropagation and approximate inference in deep generative models. In Proceedings of the 31st International Conference on Machine Learning, volume 32 of Proceedings of Machine Learning Research, pages 1278–1286, Bejing, China. PMLR.
 Ribeiro et al. (2016) Marco Ribeiro, Sameer Singh, and Carlos Guestrin. 2016. “why should i trust you?”: Explaining the predictions of any classifier. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Demonstrations, pages 97–101. Association for Computational Linguistics.

Rolfe (2017)
Jason Tyler Rolfe. 2017.
Discrete variational autoencoders.
In ICLR.  Socci et al. (1998) Nicholas D. Socci, Daniel D. Lee, and H. Sebastian Seung. 1998. The rectified gaussian distribution. In M. I. Jordan, M. J. Kearns, and S. A. Solla, editors, Advances in Neural Information Processing Systems 10, pages 350–356. MIT Press.
 Socher et al. (2013) Richard Socher, Alex Perelygin, Jean Wu, Jason Chuang, Christopher D. Manning, Andrew Ng, and Christopher Potts. 2013. Recursive deep models for semantic compositionality over a sentiment treebank. In Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, pages 1631–1642. Association for Computational Linguistics.

Thrun (1995)
Sebastian Thrun. 1995.
Extracting rules from artificial neural networks with distributed representations.
In Advances in neural information processing systems, pages 505–512.  Titov and McDonald (2008) Ivan Titov and Ryan McDonald. 2008. A joint model of text and aspect ratings for sentiment summarization. In Proceedings of ACL.

Williams (1992)
Ronald J Williams. 1992.
Simple statistical gradientfollowing algorithms for connectionist reinforcement learning.
Machine learning, 8(34):229–256.  Winn and Bishop (2005) John Winn and Christopher M Bishop. 2005. Variational message passing. Journal of Machine Learning Research, 6(Apr):661–694.
 Zaidan and Eisner (2008) Omar Zaidan and Jason Eisner. 2008. Modeling annotators: A generative approach to learning from annotator rationales. In Proceedings of the 2008 Conference on Empirical Methods in Natural Language Processing, pages 31–40, Honolulu, Hawaii. Association for Computational Linguistics.
 Zaidan et al. (2007) Omar Zaidan, Jason Eisner, and Christine Piatko. 2007. Using “annotator rationales” to improve machine learning for text categorization. In Human Language Technologies 2007: The Conference of the North American Chapter of the Association for Computational Linguistics; Proceedings of the Main Conference, pages 260–267. Association for Computational Linguistics.
 Zhang et al. (2016) Ye Zhang, Iain Marshall, and Byron C. Wallace. 2016. Rationaleaugmented convolutional neural networks for text classification. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pages 795–804, Austin, Texas. Association for Computational Linguistics.
Appendix A Kumaraswamy distribution
A Kumaraswamydistributed variable takes on values in the open interval and has density
(17) 
where and are shape parameters. Its cumulative distribution takes a simple closedform expression
(18a)  
(18b) 
with inverse
(19) 
a.1 Generalisedsupport Kumaraswamy
We can generalise the support of a Kumaraswamy variable by specifying two constants and transforming a random variable to obtain as shown in (20, left).
(20) 
The density of the resulting variable is
(21a)  
(21b)  
(21c) 
where by definition. This affine transformation leaves the cdf unchanged, i.e.
(22)  
Thus we can obtain samples from this generalisedsupport Kumaraswamy by sampling from a uniform distribution
, applying the inverse transform (19), then shifting and scaling the sample according to (20, left).a.2 Rectified Kumaraswamy
First, we stretch a Kumaraswamy distribution to include and in its support, that is, with and , we define . Then we apply a hardsigmoid transformation to this variable, that is, , which results in a rectified distribution which gives support to the closed interval . We denote this rectified variable by
(23) 
whose distribution function is
(24)  
where
(25)  
is the probability of sampling exactly , where
(26)  
is the probability of sampling exactly 1, and
(27) 
is the probability of drawing a continuous value in . Note that we used the result in (22) to express these probabilities in terms of the tractable cdf of the original Kumaraswamy variable.
a.3 Reparameterized gradients
Let us consider the case where we need derivatives of a function of the underlying uniform variable , as when we compute reparameterized gradients in variational inference. Note that
(28) 
by chain rule. The term
depends on a differentiable observation model and poses no challenge; the termis the derivative of the hardsigmoid function, which is
for or , for , and undefined for ; the term follows directly from (20, left); the term depends on the Kumaraswamy inverse cdf (19) and also poses no challenge. Thus the only two discontinuities happen for , which is a measure set under the stretched Kumaraswamy: we say this reparameterisation is differentiable almost everywhere, a useful property which essentially circumvents the discontinuity points of the rectifier.a.4 HardKumaraswamy PDF and CDF
Appendix B Implementation Details
b.1 Multiaspect Sentiment Analysis
Our hyperparameters are taken from Lei et al. (2016) and listed in Table 4. The pretrained word embeddings and data sets are available online at http://people.csail.mit.edu/taolei/beer/
. We train for 100 epochs and select the best models based on validation loss. For the MSE tradeoff experiments on all aspects combined, we train for a maximum of 50 epochs.
Optimizer  Adam 

Learning rate  0.0004 
Word embeddings  200D (Wiki, fixed) 
Hidden size  200 
Batch size  256 
Dropout  0.1, 0.2 
Weight decay  
Cell  RCNN 
For the Bernoulli baselines we vary weight among , just as in the original paper. We set the fused lasso (coherence) weight to .
For the HardKuma models we set a target selection rate to the values targeted in Table 2, and optimize to this end using the Lagrange multiplier. We chose the fused lasso weight from .
b.1.1 Recurrent Unit
b.1.2 Expected values for dependent latent variables
The expected is a chain of nested expectations, and we solve each term
(29)  
as a function of a sampled prefix, and the shape parameters are predicted in sequence.
b.2 Sentiment Classification (SST)
For sentiment classification we make use of the PyTorch bidirectional LSTM module for encoding sentences, for both the rationale extractor and the classifier. The BiLSTM final states are concatenated, after which a linear layer followed by a softmax produces the prediction. Hyperparameters are listed in Table 5. We apply dropout to the embeddings and to the input of the output layer.
Optimizer  Adam 

Learning rate  0.0002 
Word embeddings  300D Glove (fixed) 
Hidden size  150 
Batch size  25 
Dropout  0.5 
Weight decay  
Cell  LSTM 
b.3 Natural Language Inference (SNLI)
Our hyperparameters are taken from Parikh et al. (2016) and listed in Table 6. Different from Parikh et al. is that we use Adam as the optimizer and a batch size of 64. Word embeddings are projected to 200 dimensions with a trained linear layer. Unknown words are mapped to 100 unknown word classes based on the MD5 hash function, just as in Parikh et al. (2016), and unknown word vectors are randomly initialized. We train for 100 epochs, evaluate every 1000 updates, and select the best model based on validation loss. Figure 10 shows a correct and incorrect example with HardKuma attention for each relation type (entailment, contradiction, neutral).
Optimizer  Adam 

Learning rate  0.0001 
Word embeddings  300D (Glove, fixed) 
Hidden size  200 
Batch size  64 
Dropout  0.2 
Comments
There are no comments yet.