1 Introduction
Compression is a ubiquitous task for humans and machines alike [Cover & Thomas (2006), MacKay (2002)]. For example, machines must turn the large pixel grids of color that form pictures into small files capable of being shared quickly on the web [Wallace (1991)], humans must compress the vast stream of ongoing sensory information they receive into small changes in the brain that form memories [Kandel et al (2000)], and data scientists must turn large amounts of highdimensional and messy data into more manageable and interpretable clusters [MacKay (2002)].
Lossy compression involves an implicit decision about what is relevant and what is not [Cover & Thomas (2006), MacKay (2002)]. In the example of image compression, the algorithms we use deem some features essential to representing the subject matter well, and others are thrown away. In the example of human memory, our brains deem some details important enough to warrant attention, and others are forgotten. And in the example of data clustering, information about some features is preserved in the mapping from data point to cluster ID, while information about others is discarded.
In many cases, the criterion for “relevance” can be described as information about some other variable(s) of interest. Let’s call the signal we are compressing, the compressed version, the other variable of interest, and the “information” that has about (we will formally define this later). For human memory, is past sensory input,
the brain’s internal representation (e.g. the activity of a neural population, or the strengths of a set of synapses), and
the features of the future environment that the brain is interested in predicting, such as extrapolating the position of a moving object. Thus, represents the predictive power of the memories formed [Palmer et al (2015)]. For data clustering, is the original data, is the cluster ID, and is the target for prediction, for example purchasing or adclicking behavior in a user segmentation problem. In summary, a good compression algorithm can be described as a tradeoff between the compression of a signal and the selective maintenance of the “relevant” bits that help predict another signal.This problem was formalized as the “information bottleneck” (IB) by Tishby, Pereira, and Bialek [Tishby (1999)]. Their formulation involved an informationtheoretic optimazation problem, and resulted in an iterative soft clustering algorithm guaranteed to converge to a local (though not necessarily global) optimum. In their cost functional, compression was measured by the mutual information . This compression metric has its origins in ratedistortion theory and channel coding, where represents the maximal information transfer rate, or capacity, of the communication channel between and [Cover & Thomas (2006)]. While this approach has its applications, often one is more interested in directly restricting the amount of resources required to represent , represented by the entropy . This latter notion comes from the source coding literature and implies a restriction on the representational cost of [Cover & Thomas (2006)]. In the case of human memory, for example,
would roughly correspond to the number of neurons or synapses required to represent or store a sensory signal
. In the case of data clustering, is related to the number of clusters.In the following paper, we introduce an alternative formulation of the IB, called the deterministic information bottleneck (DIB), replacing the compression measure with , thus emphasizing contraints on representation, rather than communication. Using a clever generalization of both cost functionals, we derive an iterative solution to the DIB, which turns out to provide a hard clustering, or deterministic mapping from to , as opposed to the soft clustering, or probabilitic mapping, that IB provides. Finally, we compare the IB and DIB solutions on synthetic data to help illustrate their differences.
2 The original information bottleneck (IB)
Given the joint distribution
, the encoding distribution is obtained through the following “information bottleneck” (IB) optimization problem:(1) 
subject to the Markov constraint . Here denotes the mutual information between and , that is ,^{1}^{1}1Implicit in the summation here, we have assumed that , , and are discrete. We will be keeping this assumption throughout for convenience of notation, but note that the IB generalizes naturally to , , and continuous by simply replacing the sums with integrals (see, for example, [Chechik et al (2005)]). where
denotes the KullbackLeibler divergence.
^{2}^{2}2For those unfamiliar with it, mutual information is a very general measure of how related two variables are. Classic correlation measures typically assume a certain form of the relationship between two variables, say linear, whereas mutual information is agnostic as to the details of the relationship. One intuitive picture comes from the entropy decomposition: . Since entropy measures uncertainty, mutual information measures the reduction in uncertainty in one variable when observing the other. Moreover, it is symmetric (), so the information is mutual. Another intuitive picture comes from the form: . Sincemeasures the distance between two probability distributions, the mutual information quantifies how far the relationship between
and is from a probabilistically independent one, that is how far the joint is from the factorized . A very nice summary of mutual information as a dependence measure is included in [Kinney & Atwal (2014)]. The first term in the cost function is meant to encourage compression, while the second relevance. is a nonnegative free parameter representing the relative importance of compression and relevance, and our solution will be a function of it. The Markov constraint simply enforces the probabilistic graphical structure of the task; the compressed representation is a (possibly stochastic) function of and can only get information about through . Note that we are using to denote distributions that are given and fixed, and to denote distributions that we are free to change and that are being updated throughout the optimization process.Through a standard application of variational calculus (see Section 8 for a detailed derivation of the solution to a more general problem introduced below), one finds the formal solution:^{3}^{3}3For the reader familiar with ratedistortion theory, eqn 2 can be viewed as the solution to a ratedistortion problem with distortion measure given by the KLdivergence term in the exponent.
(2)  
(3) 
where is a normalization factor, and is a Lagrange multiplier that enters enforcing normalization of .^{4}^{4}4More explicitly, our cost function also implicitly includes a term and this is where comes in to the equation. See Section 8 for details. To get an intuition for this solution, it is useful to take a clustering perspective  since we are compressing into , many will be mapped to the same and so we can think of the IB as “clustering” s into their cluster labels . The solution is then likely to map to when is small, or in other words, when the distributions and are similar. These distributions are similar to the extent that and provide similar information about . In summary, inputs get mapped to clusters that maintain information about , as was desired.
This solution is “formal” because the first equation depends on the second and vice versa. However, [Tishby (1999)] showed that an iterative approach can be built on the the above equations which provably converges to a local optimum of the IB cost function (eqn. 1).
Starting with some initial distributions , , and , the update is given by:^{5}^{5}5Note that, if at step no s are assigned to a particular (i.e. ), then . That is, no s will ever again be assigned to (due to the factor in ). In other words, the number of s “in use” can only decrease during the iterative algorithm (or remain constant). Thus, it seems plausible that the solution will not depend on the cardinality of , provided it is chosen to be large enough.
(4)  
(5)  
(6) 
Note that the first equation is the only “meaty” one; the other two are just there to enforce consistency with the laws of probability (e.g. that marginals are related to joints as they should be). In principle, with no proof of convergence to a global optimum, it might be possible for the solution obtained to vary with the initialization, but in practice, the cost function is “smooth enough” that this does not seem to happen. This algorithm is summarized in algorithm 1. Note that while the general solution is iterative, there is at least one known case in which an analytic solution is possible, namely when and are jointly Gaussian [Chechik et al (2005)].
In summary, given the joint distribution , the IB method extracts a compressive encoder that selectively maintains the bits from that are informative about . As the encoder is a function of the free parameter , we can visualize the entire family of solutions on a curve (figure 1), showing the tradeoff between compression (on the axis) and relevance (on the axis), with as an implicitly varying parameter. For small , compression is more important than prediction and we find ourselves at the bottom left of the curve in the high compression, low prediction regime. As increases, prediction becomes more important relative to compression, and we see that both and increase. At some point, saturates, because there is no more information about that can be extracted from (either because has reached or because has too small cardinality). In this regime, the encoder will approach the trivially deterministic solution of mapping each to its own cluster. At any point on the curve, the slope is equal to , which we can read off directly from the cost functional. Note that the region below the curve is shaded because this area is feasible; for suboptimal , solutions will lie in this region. Optimal solutions will of course lie on the curve, and no solutions will lie above the curve.
Additional work on the IB has highlighted its relationship with maximum likelihood on a multinomial mixture model [Slonim & Weiss (2002)] and canonical correlation analysis [Creutzig et al (2009)] (and therefore linear Gaussian models [Bach & Jordan (2005)] and slow feature analysis [Turner & Sahani (2007)]). Applications have included speech recognition [Hecht & Tishby (2005), Hecht & Tishby (2008), Hecht et al (2009)], topic modeling [Slonim & Tishby (2000), Slonim & Tishby (2001), Bekkerman et al (2001), Bekkerman et al (2003)], and neural coding [Schneidman et al (2002), Palmer et al (2015)]
. Most recently, the IB has even been proposed as a method for benchmarking the performance of deep neural networks
[Tishby & Zaslavsky (2015)].3 The deterministic information bottleneck
Our motivation for introducing an alternative formulation of the information bottleneck is rooted in the “compression term” of the IB cost function; there, the minimization of the mutual information represents compression. As discussed above, this measure of compression comes from the channel coding literature and implies a restriction on the communication cost between and Here, we are interested in the source coding notion of compression, which implies a restriction on the representational cost of . For example, in neuroscience, there is a long history of work on “redundancy reduction” in the brain in the form of minimizing [Barlow (1981), Barlow (2001), Barlow (2001)].
Let us call the original IB cost function , that is . We now introduce the deterministic information bottleneck (DIB) cost function:
(7) 
which is to be minimized over and subject to the same Markov constraint as the original formulation (eqn 1
). The motivation for the “deterministic” in its name will become clear in a moment.
To see the distinction between the two cost functions, note that:
(8)  
(9) 
where we have used the decomposition of the mutual information . is sometimes called the “noise entropy” and measures the stochasticity in the mapping from to . Since we are minimizing these cost functions, this means that the IB cost function encourages stochasticity in the encoding distribution relative to the DIB cost function. In fact, we will see that by removing this encouragement of stochasticity, the DIB problem actually produces a deterministic encoding distribution, i.e. an encoding function, hence the “deterministic” in its name.
Naively taking the same variational calculus approach as for the IB problem, one cannot solve the DIB problem.^{6}^{6}6When you take the variational derivative of with respect to and set it to zero, you get no explicit term, and it is therefore not obvious how to solve these equations. We cannot rule that that approach is possible, of course; we have just here taken a different route. To make this problem tractable, we are going to define a family of cost functions of which the IB and DIB cost functions are limiting cases. That family, indexed by , is defined as:^{7}^{7}7Note that for , we cannot allow to be continuous since can become infinitely negative, and the optimal solution in that case will trivially be a delta function over a single value of for all , across all values of . This is in constrast to the IB, which can handle continuous . In any case, we continue to assume discrete , , and for convenience.
(10) 
Clearly, . However, instead of looking at as the case, we’ll define the DIB solution as the limit of the solution to the generalized problem :^{8}^{8}8Note a subtlety here that we cannot claim that the is the solution to , for although and , the solution of the limit need not be equal to the limit of the solution. It would, however, be surprising if that were not the case.
(11) 
Taking the variational calculus approach to minimizing (under the Markov constraint), we get the following solution for the encoding distribution (see Section 8 for the derivation and explicit form of the normalization factor ):
(12)  
(13) 
Note that the last equation is just eqn 3, since this just follows from the Markov constraint. With , we can see that the first equation just becomes the IB solution from eqn 2, as should be the case.
Before we take the limit, note that we can now write a generalized IB iterative algorithm (indexed by ) which includes the original as a special case ():
(14)  
(15)  
(16) 
This generalized algorithm can be used in its own right, however we will not discuss that option further here.
For now, we take the limit and see that something interesting happens with  the argument of the exponential begins to blow up. For a fixed , this means that will collapse into a delta function at the value of which maximizes . That is:
(17) 
where:
(18) 
So, as anticipated, the solution to the DIB problem is a deterministic encoding distribution. The above encourages that we use as few values of as possible, via a “richgetricher” scheme that assigns each preferentially to a already with many s assigned to it. The KL divergence term, as in the original IB problem, just makes sure we pick s which retain as much information from about as possible. The parameter , as in the original problem, controls the tradeoff between how much we value compression and prediction.
Also like in the original problem, the solution above is only a formal solution, since eqn 12 depends on eqn 13 and vice versa. So we will again need to take an iterative approach; in analogy to the IB case, we repeat the following updates to convergence (from some initialization):^{9}^{9}9As with the IB, the DIB has the property that once a cluster goes unused, it will not be brought back into use in future steps. That is, if , then and . So once again, one should conservatively choose the cardinality of to be “large enough”; for both the IB and DIB, we chose to set it equal to the cardinality of .
(19)  
(20)  
(21)  
(22)  
(23)  
(24) 
This process is summarized in Algorithm 2.
Note that the DIB algorithm also corresponds to “clamping” IB at every step by assigning each to its highest probability cluster . We can see this by taking the argmax of the logarithm of in eqn 2, noting that the argmax of a positive function is equivalent the argmax of its logarithm, discarding the term since it doesn’t depend on , and seeing that the result corresponds to eqn 18. We emphasize, however, that this is not the same as simply running the IB algorithm to convergence and then clamping the resulting encoder; that would, in most cases, produce a suboptimal, “unconverged” deterministic solution.
Like with the IB, the DIB solutions can be plotted as a function of . However, in this case, it is more natural to plot as a function of , rather than . That said, in order to compare the IB and DIB, they need to be plotted in the same plane. When plotting in the plane, the DIB curve will of course lie below the IB curve, since in this plane, the IB curve is optimal; the opposite will be true when plotting in the plane. Comparisons with experimental data can be performed in either plane.
4 Comparison of IB and DIB
To get an idea of how the IB and DIB solutions differ in practice, we generated a series of random joint distributions , solved for the IB and DIB solutions for each, and compared them in both the IB and DIB plane. To generate the , we first sampled from a symmetric Dirichlet distribution with concentration parameter (so ), and then sampled each row of from another symmetric Dirichlet distribution with concentration parameter (so ). In the experiments shown here, we set to 1000, so that each was approximately equally likely, and we set to be equally spaced logarithmically between and , in order to provide a range of informativeness in the conditionals. We set the cardinalities of and to and , with for two reasons. First, this encourages overlap between the conditionals , which leads to a more interesting clustering problem. Second, in typical applications, this will be the case, such as in document clustering where there are often many more documents that vocabulary words. Since the number of clusters in use for both IB and DIB can only decrease from iteration to iteration (see footnote 9), we always initialized .^{10}^{10}10An even more efficient setting might be to set the cardinality of based on the entropy of , say , but we didn’t experiment with this.
For the DIB, we initialized the cluster assignments to be as even across the cluster as possible, i.e. each data points belonged to its own cluster. For IB, we initialized using a soft version of the same procedure, with 75% of each conditional’s probability mass assigned to a unique cluster, and the remaining 25% a normalized uniform random vector over the remaining
clusters.An illustrative pair of solutions is shown in figure 2. The key feature to note is that, while performance of the IB and DIB solutions are very similar in the IB plane, their behavior differs drastically in the DIB plane.
Perhaps most unintuitive is the behavior of the IB solution in the DIB plane, where from an entropy perspective, the IB actually “decompresses” the data (i.e. ). To understand this behavior, recall that the IB’s compression term is the mutual information . This term is minimized by any that maps s independently of s. Consider two extremes of such mappings. One is to map all values of to a single value of ; this leads to . The other is to map each value of uniformly across all values of ; this leads to and . In our initial studies, the IB consistently took the latter approach.^{11}^{11}11Intuitively, this approach is “more random” and is therefore easier to stumble upon during optimization. Since the DIB cost function favors the former approach (and indeed the DIB solution follows this approach), the IB consistently performs poorly by the DIB’s standards. This difference is especially apparent at small , where the compression term matters most, and as increases, the DIB and IB solutions converge in the DIB plane.
To encourage the IB to perform closer to DIB optimality at small , we next altered our initialization scheme of to one biased towards singlecluster solutions; instead of each having most of its probability mass on a unique cluster , we placed most of the probability mass for each on the same cluster . The intended effect was to start the IB closer to solutions in which all data points were mapped to a single cluster. Results are shown in figure 3. Here, is the amount of probability mass placed on the cluster , that is ; the probability mass for the remaining clusters was again initialized to a normalized uniform random vector. “random” refers to an initialization which skips placing the mass and just initializes each to a normalized uniform random vector.
We note several features. First, although we can see a gradual shift of the IB towards DIBlike behavior in the DIB plane as , the IB solutions never quite reach the performance of DIB. Moreover, as
, the singlecluster initializations exhibit a phase transition in which, regardless of
, they “skip” over a sizable fraction of lower solutions that are picked up by DIB. Third, even for higher solutions, the singlecluster initializations seem to perform suboptimally, not quite extracting all of the information , as DIB and the multicluster initialization from the previous section do; this can be seen in both the IB and DIB plane.To summarize, the IB and DIB perform similarly by the IB standards, but the DIB tends to outperform the IB dramatically by the DIB’s standards. Careful initialization of the IB can make up some of the difference, but not all.
It is also worth noting that, across all the datasets we tested, the DIB also tended to converge faster, as illustrated in figure 4. The DIB speedup over IB varied depended on the convergence conditions. In our experiments, we defined convergence as when the relative steptostep change in the cost functional was smaller than some threshold ctol, that is when at step . In the results above, we used . In figure 4, we vary ctol, with the IB initialization scheme fixed to the original “multicluster” version, to show the effect on the relative speedup of DIB over IB. While DIB remained 25x faster than IB in all cases tested, that speedup tended to be more pronounced with lower ctol. Since the ideal convergence conditions would probably vary by dataset size and complexity, it is difficult to make any general conclusions, though our experiments do at least suggest that DIB offers a computational advantage over IB.
5 Related work
The DIB is not the first hard clustering version of IB.^{12}^{12}12In fact, even the IB itself produces a hard clustering in the large limit. However, it trivially assigns all data points to their own clusters. Indeed, the agglomerative information bottleneck (AIB) [Slonim & Tishby (1999)] also produces hard clustering and was introduced soon after the IB. Thus, it is important to distinguish between the two approaches. AIB is a bottomup, greedy method which starts with all data points belonging to their own clusters and iteratively merges clusters in a way which maximizes the gain in relevant information. It was explicitly designed to produce a hard clustering. DIB is a topdown method derived from a cost function that was not designed to produce a hard clustering. Our starting point was to alter the IB cost function to match the source coding notion of compression. The emergence of hard clustering in DIB is itself a result. Thus, while AIB does provide a hard clustering version of IB, DIB contributes the following in addition: 1) Our study emphasizes why a stochastic encoder is optimal for IB, namely due to the noise entropy term. 2) Our study provides a principled, topdown derivation of a hard clustering version of IB, based upon an intuitive change to the cost function. 3) Our nontrivial derivation also provides a cost function and solution which interpolates between DIB and IB, by adding back the noise entropy continuously, i.e. with . This interpolation may be viewed as adding a regularization term to DIB, one that may perhaps be useful in dealing with finitely sampled data. Another interpretation of the cost function with intermediate is as a penalty on both the mutual information between and and the entropy of the compression, .
The original IB also provides a deterministic encoding upon taking the limit that corresponds to the causalstate partition of histories [Still et al (2010)]. However, this is the limit of no compression, whereas our approach allows for an entire family of deterministic encoders with varying degrees of compression.
6 Discussion
Here we have introduced the deterministic information bottleneck (DIB) as an alternative to the information bottleneck (IB) for compression and clustering. We have argued that the DIB cost function better embodies the goal of lossy compression of relevant information, and shown that it leads to a nontrivial deterministic version of the IB. We have compared the DIB and IB solutions on synthetic data and found that, in our experiments, the DIB performs nearly identically to the IB in terms of the IB cost function, but far superior in terms of its own cost function. We also noted that the DIB achieved this performance at a computational efficiency 25x better than the IB.
Of course, in addition to the studies with synthetic data here, it is important to compare the DIB and IB on real world datasets as well to see whether the DIB’s apparent advantages hold, for example with datasets that have more explicit hierarchical structure for both algorthms to exploit, such as in topic modelling [Blei et al (2004), Slonim & Weiss (2002)].
One particular application of interest is maximally informative clustering, where it would be interesting to know how IB and DIB relate to classic clustering algorithms such as means [Strouse & Schwab (2017)]. Previous work has, for example, offered a principled way of choosing the number of clusters based on the finiteness of the data [Still & Bialek (2004)], and similarly interesting results may exist for the DIB. More generally, there are learning theory results showing generalization bounds on IB for which an analog on DIB would be interesting as well [Shamir et al (2010)].
Another potential area of application is modeling the extraction of predictive information in the brain (which is one particular example in a long line of work on the exploitation of environmental statistics by the brain [Barlow (1981), Barlow (2001), Barlow (2001), Atick & Redlich (1992), Olshausen & Field (1996), Olshausen & Field (1997), Simoncelli & Olshausen (2001), Olshausen & Field (2004)]). There, would be the stimulus at time , the stimulus a short time in the future , and the activity of a population of sensory neurons. One could even consider neurons deeper in the brain by allowing and to correspond not to an external stimulus, but to the activity of upstream neurons. An analysis of this nature using retinal data was recently performed with the IB [Palmer et al (2015)]. It would be interesting to see if the same data corresponds better to the behavior of the DIB, particularly in the DIB plane where the IB and DIB differ dramatically.
We close by noting that DIB is an imperfect name for the algorithm introduced here for a couple of reasons. First, there do exist other deterministic limits and approximations to the IB (see, for example, the discussion of the AIB in section 5), and so we hesitate to use the phrase “the” deterministic IB. Second, our motivation here was not to create a deterministic version of IB, but instead to alter the cost function in a way that better encapsulates the goals of certain problems in data analysis. Thus, the deterministic nature of the solution was a result, not a goal. For this reason, “entropic bottleneck” might also be an appropriate name.
7 Acknowledgements
For insightful discussions, we would like to thank Richard Turner, Máté Lengyel, Bill Bialek, Stephanie Palmer, and Gordon Berman. We would also like to acknowledge financial support from NIH K25 GM098875 (Schwab), the Hertz Foundation (Strouse), and the Department of Energy Computational Sciences Graduate Fellowship (Strouse).
8 Appendix: derivation of generalized IB solution
Given and subject to the Markov constraint , the generalized IB problem is:
(25) 
where we have now included the Lagrange multiplier term (which enforces normalization of ) explicitly. The Markov constraint implies the following factorizations:
(26)  
(27) 
which give us the following useful derivatives:
(28)  
(29) 
Now taking the derivative of the cost function with respect to the encoding distribution, we get:
(30)  
(31)  
(32)  
(33)  
(34)  
(35)  
(36)  
(37)  
(38)  
(39)  
(40)  
(41) 
Setting this to zero implies that:
(42) 
We want to rewrite the term as a KL divergence. First, we will need that . Second, we will add and subtract . This gives us:
(43)  
(44) 
The second term is now just . Dividing both sides by , this leaves us with the equation:
(45) 
where we have absorbed all of the terms that don’t depend on into a single factor:
(46) 
Solving for , we get:
(47)  
(48) 
where:
(49) 
is just a normalization factor. Now that we’re done with the general derivation, let’s add a subscript to the solution to distinguish it from the special cases of the IB and DIB.
(50) 
The IB solution is then:
(51) 
while the DIB solution is:
(52) 
with:
(53) 
References
 Atick & Redlich (1992) Atick, J.J. & Redlich, A.N. (1992). What Does the Retina Know about Natural Scenes? Neural Computation, 4, 196210.
 Bach & Jordan (2005) Bach, F.R., & Jordan, M.I. (2005). A probabilistic interpretation of canonical correlation analysis. Technical Report.
 Barlow (1981) Barlow, H.B. (1981). Critical Limiting Factors in the Design of the Eye and Visual Cortex. Proceedings of the Royal Society B: Biological Sciences, 212(1186), 1–34.
 Barlow (2001) Barlow, H. (2001). Redundancy reduction revisited. Network: Computation in Neural Systems, 12(3), 241–253.
 Barlow (2001) Barlow, H. (2001). The exploitation of regularities in the environment by the brain. Behav. Brain Sci. 24, 602–607.
 Bekkerman et al (2001) Bekkerman, R., ElYaniv, R., Tishby, N., & Winter, Y. (2001). On feature distributional clustering for text categorization. Proceedings of SIGIR2001.

Bekkerman et al (2003)
Bekkerman, R., ElYaniv,
R., & Tishby, N. (2003). Distributional word clusters vs. words
for text categorization.
Journal of Machine Learning Research, 3, 11831208.
 Blei et al (2004) Blei, D.M., Griffiths, T.L., Jordan, M.I. & Tenenbaum, J.B. (2004). Hierarchical Topic Models and the Nested Chinese Restaurant Process. Proc. of Advances in Neural Information Processing System (NIPS).
 Chechik et al (2005) Chechik, G., Globerson, A., Tishby, N., & Weiss, Y. (2005). Information Bottleneck for Gaussian Variables. Proc. of Advances in Neural Information Processing System (NIPS).
 Cover & Thomas (2006) Cover, T.M. & Thomas, J.A. (2006). Elements of Information Theory. John Wiley & Sons, Inc.
 Creutzig et al (2009) Creutzig, F., Globerson, A., & Tishby, N. (2009). Pastfuture information bottleneck in dynamical systems. Physical Review E, 79(4).
 Hecht & Tishby (2005) Hecht, R.M. & Tishby, N. (2005). Extraction of relevant speech features using the information bottleneck method. InterSpeech.
 Hecht & Tishby (2008) Hecht, R.M. & Tishby, N. (2007). Extraction of relevant Information using the Information Bottleneck Method for Speaker Recognition. InterSpeech.
 Hecht et al (2009) Hecht, R.M., Noor, E., & Tishby, N. (2009). Speaker recognition by Gaussian information bottleneck. InterSpeech.
 Kandel et al (2000) Kandel, E.R., Schwartz, J.H., Jessell, T.M., Siegelbaum, S.A., & Hudspeth, A.J. (2013). Principles of neural science (5th ed.). New York: McGrawHill, Health Professions Division.
 Kinney & Atwal (2014) Kinney, J.B. & Atwal, G.S. (2014). Equitability, mutual information, and the maximal information coefficient. PNAS.
 MacKay (2002) Mackay, D. (2002). Information Theory, Inference, & Learning Algorithms. Cambridge University Press.
 Olshausen & Field (1996) Olshausen, B.A., & Field, D.J. (1996). Emergence of simplecell receptive field properties by learning a sparse code for natural images. Nature, 381(6583), 607–609.
 Olshausen & Field (1997) Olshausen, B.A., & Field, D.J. (1997). Sparse coding with an overcomplete basis set: A strategy employed by V1? Vision Research.
 Olshausen & Field (2004) Olshausen, B.A., & Field, D.J. (2004). Sparse coding of sensory inputs. Current Opinion in Neurobiology, 14(4), 481–487.
 Palmer et al (2015) Palmer, S.E., Marre, O., Berry, M.J., & Bialek, W. (2015). Predictive information in a sensory population. PNAS, 112(22), 6908–6913.
 Schneidman et al (2002) Schneidman, E., Slonim, N., Tishby, N., deRuyter van Steveninck, R., & Bialek, W. (2002). Analyzing neural codes using the information bottleneck method. Proc. of Advances in Neural Information Processing System (NIPS).
 Shamir et al (2010) Shamir, O., Sabato, S., & Tishby, N. (2010). Learning and Generalization with the Information Bottleneck. Theoretical Computer Science, Volume 411, Issues 2930, Pages 26962711.
 Simoncelli & Olshausen (2001) Simoncelli, E.P., & Olshausen, B.A. (2001). Natural image statistics and neural representation. Annual Review of Neuroscience.
 Slonim & Tishby (1999) Slonim, N., & Tishby, N. (1999). Agglomerative information bottleneck. Proc. of Advances in Neural Information Processing System (NIPS).
 Slonim & Tishby (2000) Slonim, N., & Tishby, N. (2000). Document clustering using word clusters via the information bottleneck method. Proceedings of the 23rd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, 208–215.
 Slonim & Tishby (2001) Slonim, N., & Tishby, N. (2001). The Power of Word Clusters for Text Classification. Proceedings of the European Colloquium on IR Research, ECIR 2001, 1–12.
 Slonim & Weiss (2002) Slonim, N., & Weiss, Y. (2002). Maximum likelihood and the information bottleneck. Proc. of Advances in Neural Information Processing System (NIPS), 15, 335–342.
 Still & Bialek (2004) Still, S., & Bialek, W. (2004). How many clusters? An informationtheoretic perspective. Neural Computation, 16(12), 2483–2506.

Still et al (2010)
Still, S., Crutchfield, J.P.,
& Ellison, C.J. (2010).
Optimal causal inference: Estimating stored information and approximating causal architecture
. Chaos: an Interdisciplinary Journal of Nonlinear Science, 20(3), 037111.  Strouse & Schwab (2017) Strouse, DJ & Schwab, D. (2017). On the relationship between distributional and geometric clustering. In progress.
 Tishby (1999) Tishby, N., Pereira, F. & Bialek, W. (1999). The Information Bottleneck Method. Proceedings of The 37th Allerton Conference on Communication, Control, & Computing, Univ. of Illinois.
 Tishby & Zaslavsky (2015) Tishby, N., & Zaslavsky, N. (2015). Deep Learning and the Information Bottleneck Principle. arXiv.org.
 Turner & Sahani (2007) Turner, R.E., & Sahani, M. (2007). A maximumlikelihood interpretation for slow feature analysis. Neural Computation, 19(4), 1022–1038.
 Wallace (1991) Wallace, G.K. (1991). The JPEG Still Picture Compression Standard. Commun. ACM, vol. 34, pp. 3044.
Comments
There are no comments yet.