I Introduction
In the last decade, deep learning systems show a fascinating performance in solving different complex tasks. We have designed different system structures for different problems and revealed some general rules for the designing of deep learning systems. There are also theoretical attempts to understand deep learning systems from both mathematical and physical perspectives[1]. But still we are lacking of a theoretical framework to answer the question, why and how deep learning systems works. Also it’s highly desired that a theoretical framework of deep learning systems can be used to guide the design the structures of deep learning systems from an algorithmic level.
In this paper we try to fill this gap by drawing a geometric picture of deep learning systems. We build our geometric framework to understand deep learning systems by comparing the deep learning system with other two existing matural geometric structures, the geometry of quantum computations and the geometry of diffeomorphic template matching in the field of computational anatomy. We show that deep learning systems can be formulated in a geometric language, by which we can draw geometric pictures of different deep learning systems including convolutional neural networks, residual networks, recursive neural networks, fractal neural networks and recurrent neural networks. What’s more, these geometric pictures can be used to analysis the performance of different deep learning structures and provide guidance to the design of deep learning systems from an algorithmic level.
The rest of this paper is arranged as follows. We will first give a brief overview of the geometry of quantum computations and the geometry of diffeomorphic template matching. Then we will explain the geometric framework of deep learning systems and apply our framework to draw correspondent geometric pictures of different deep learning networks. Finally we will give a general optimization based framework of deep learning systems, which can be used to address the equilibrium propagation algorithm.
Ii What can geometry teach us
It’s well known that geometry is not only a core concept of mathematics, but also it plays a key role in modern physics. The great success of geometrization of physics tells us that the soul of physical systems lies in their geometric structures. It’s natural to ask if geometry can help to reveal the secret of our human intelligence and our stateoftheart artificial intelligence systems, deep learning systems. The answer is YES.
We will first introduce two interesting geometric structures in the fields of quantum computation and computational anatomy. We will see, their geometric structures share some similarities with deep learning systems and the geometric framework of deep learning systems can be built based on an understanding of these structures.
Iia Geometry of quantum computation
Geometry concepts have been widely discussed in formulating quantum mechanics and quantum information processing systems, including the geometry of quantum states and their evolution[2][3][4], the geometry of entanglement[5][6] and also the relationship between the spacetime structure and the geometry of quantum states[7][8][9].
a geometric framework was proposed for the complexity of quantum computations. Its basic idea is to introduce a Riemannian metric to the space of nqubit unitary operators so that the quantum computation complexity becomes a geometric concept as given by the slogan
quantum computation as free falling. The key components of this framework can be summarized as follows.
An algorithm of a nqubit quantum computation system is a unitary operation , which can evolve the nqubit initial state to the final state . is the space of the unitary operations of nqubits, which is both a manifold and a Lie group.

Any physical realization of the algorithm is a curve with , where is the identity operator. A smooth curve can be achieved by a Hamiltonian so that , where is in the tangent space of at and also an element of the Lie algebra .

A Riemannian metric can be introduced to the manifold so that a Riemannian structure can be built on . The metric is defined as
(1) where
are tangent vectors in the tangent space at
, and is the projection of to a simple local Hamiltonian space , and . With this metric any finite length curve connecting the identity and will only contain local Hamiltonian when . Usually the geodesic connecting and has the minimal length, which is defined as the complexity of the algorithm . We also know the geodesic is given by the EPDiff equation, which is obtained by solving the optimization problem . 
The quantum circuit model of quantum computations tells us that any can be approximated with any accuracy with a small universal gate set, which contains only simple local operators, for example only a finite number of 1 and 2 qubit operators. The quantum circuit model can then be understood as approximating with a piecewise smooth curve, where each segment corresponds to a universal gate. It’s proven in [10][11] that the number of universal gates to approximate within some constant error is given by , where is the geodesic distance between and . So an algorithm can be efficiently realized by the quantum circuit model only when its geodesic distance to the identity operator is polynomial with respect to .

The set of operators that can be reached by a geodesic with a length that is polynomial with the qubit number n is only a small subset of . Equivalently, starting from a simple initial state , the quantum state that can be obtained efficiently by local quantum operations is a small subset of the quantum state space. This defines the complexity of a quantum state.

Finding the geodesic can be achieved by either the lifted Jacobi field or standard geodesic shooting algorithm. The performance of the optimization procedure to find highly depends on the curvature of the Riemannian manifold. For the metric given in (1), the sectional curvature is almost negative everywhere when is large enough. This means the geodesic is unstable and it’s general difficult to find a geodesic.
The key lesson of the geometry of quantum computations is that: Under a local Hamiltonian preferred metric, the complexity of a general quantum algorithm is exponential with the number of qubit n. If the complexity is polynomial, it can be achieved efficiently with local operators, either by a continuous local Hamiltonian or a set of universal local operators. The performance of the optimization procedure to find a good realization of depends on the curvature of the Riemannian manifold.
We will see that this observation is closely related with the question ofwhy and how cheap deep learning works.
IiB Geometry of diffeomorphic template matching
The diffeomorphic framework of computational anatomy aims to achieve a diffeomorphic matching between images on a Riemannian structure or a Lie group. The geometry of diffeomorphic template matching not only provides a mathematical framework for image registration but also validates the possibility of statistical analysis of anatomical structures on manifolds[12][13][14][15][16][17][18].
We can summarize the geometry of diffeomorphic template matching as follows:

The Riemannian geometry of diffeomorphic template matching starts with the claim that a diffeomorphic matching between two images and defined on a space ( for 2 dimensional images or for 3 dimensional volume data) is achieved by a diffeomorphic transform . the diffeomorphism transformation group of the space .

A matching between and is achieved by a smooth curve on connecting the identity transform and . The optimal is obtained by an optimization problem
(2) where , and is a metric on the tangent space of the manifold with a linear operator.
This is the LDDMM framework[12] and the optimization results in the EulerLagrange equation given by
(3) where , and is the transformation generated by during the time internal to . is the operator defined by , i.e., . In computational anatomy is usually taken as a Gaussian kernel, whose effect is to generate a smooth velocity field that leads to a smooth transformation .
The geometric picture of LDDMM is to find a geodesic that transforms to a near neighbour of so that . The optimal velocity field is obtained with a gradient descent algorithm.

The sectional curvature of the volume preserving diffeomorphisms of the flat torus with a weak right invariant metric is negative in may directions so the geodesic on this manifold can be unstable. We indicate this point to show that here we might also face a similar negative curvature case as in the geometry of the quantum computation.
The key concept of the geometry of template matching is: The template matching can be achieved by either a Riemannian geodesic or a Lie group exponential mapping in that transforms the source image to approximate the destination image. Finding the optimal curve can be formulated as an optimization problem. The Riemannian metric or the Lie group structure of play a central role in the optimization procedure.
Obviously the two geometric structures share lots of common characteristics. Both systems consider the geometry of a mapping , where is a vector space (the state space of nqubits or the space of images on respectively) and is the space of transformations on ( or ). By introducing a Riemannian metric on and an inner product on , both of them address the problem of finding a curve on a Riemannian manifold ( or ) to connect the identity operator with a target operator( or ). Equivalently this curve can achieve a transformation from an initial state ( or ) to a near neighbourhood (defined by the inner product on ) of a final state ( or ) in . The problem can be formalized as an optimization task and the optimal curve is a Riemannian geodesic (an Lie group exponential for SVF). The stability of geodesics and the convergence property of the optimization procedure highly depend on the Riemannian metric of the Riemannian manifold. In both geometric structure, we have to fact the negative sectional curvature problem.
Besides their similarities, there exist a key difference between these two geometric structures. The geometry of the quantum computation addresses the problem to reach a point efficiently by local operations. So it focuses on the computability of a physical system. For more discussion on the computability and physical systems, please refer to [10]. On the contrary, in the template matching problem, we do not emphasis to use local operations. The reason falls in that the deformation curve is in fact generated purely by local operations since it’s an integration curve of a timevarying velocity field, which is essentially local. Also in the template matching problem, it seems that we do not care if a transformation can be efficiently computed.
The reason for the difference between them is related to the concept of the diameter of a Riemannian manifold, which is defined by the maximal geodesic distance between arbitrary points on the manifold. So for the geometry of the quantum computation, the manifold has an exponentially large diameter with respect to the qubit number . So only a subset of the operations can be efficiently carried out. For the template matching problem, the diameter is usually finite. In fact the real situation is quite complex. For example, if we work on the volumepreserving diffeomorphism group of a ndimensional cube in , the diameter is finite for but infinite for if a rightinvariant metric is used. For , a metric results in a finite diameter, and a metric can even lead to a vanishing geodesic distance, which means any two points can be connected with a geodesic with an arbitrary small distance. So generally we assume all the transformations in the template matching problem can be efficiently computed.
We will see that the diameter of the Riemannian manifold of the deep learning systems is closed related with the question why deep learning works.
Iii Geometry of Deep Learning
Based on the above mentioned geometric structures, we propose a geometric framework to understand deep learning systems. We will first focus on the most popular deep learning structure, the deep convolution neural network(CNN). Then the same framework can be used to draw geometric pictures of other systems including the Residual Network(ResNet), the recursive neural network, the fractal neural network and the recurrent network.
Iiia Geometry of convolutional neural networks
The convolutional neural network is the most well studied deep learning structure, which can achieve reliable object recognition in computer vision systems by stacking multiple CONVReLUPooling operations followed by a fully connected network. From a geometric point of view, CNN achieves the object classification by performing a nonlinear transformation
from the input image space, where the object classification is difficult, to the output vector space, in which the classification is much easier. We will show that there exists an exact correspondence between CNN and the geometry of the quantum circuit model of quantum computation, where corresponds to the unitary operator of the quantum computation and the CONVReLUPooling operation plays the role of local universal gates.The geometric structure of CNN is constructed as follows.

The manifold
We first construct the mapping . In a CNN with layers, each layer accomplishes a transformation of the input data. We can easily embed the input and output data of each step into a Euclidean space with a high enough dimension . We call it and its dimension is determined by the structure of the CNN including the input data size and the kernel number of each level. Accordingly can be taken as the automorphism of . The goal of CNN is then to find and realize a proper transformation by constructing a curve on , which consists of L line segments, to reach . 
The Riemannian metric
It’s easy to see that the local CONVReLUPooling operation of each layer corresponds to the local unitary operations in the quantum computation system. The details of each CONVReLUPooling operation define the allowed local operations, i.e., the set in the quantum computation system.Now we have two different ways to understand the metric on . We can define a similar metric as (1), which means that at each layer of the CNN, this metric only assigns a finite length to the allowed local operations defined by the CONVReLUPooling operation of this layer. So in this picture the metric changes during the forward propagation along the CNN since the configuration of the CONVReLUPooling operator differs at each layer. So is a curve on a manifold with a timevarying metric. To remedy our Riemannian structure from this complicated situation, we can define another computational complexity oriented metric by scaling the metric at each layer by the complexity of the CONVReLUPooling operation. For example, if at a certain layer we have kernels with a kernel size and the input data size is , then we scale the metric by a factor . With this metric, is a curve on with this complexity oriented metric and the length of roughly corresponds to the complexity of the CNN, if the nonlinear operations are omitted. Of course, the nonlinear ReLU and pooling operation in fact also change the metric. Existing research works showed that they have a strong influence on the metric as will be explained below.

The curvature
If the CNN is formulated as a curve on a Riemannian manifold, then obviously the convergence of the training procedure to find the curve highly depends on the curvature of the manifold. It’s difficult to compute explicitly the curvature of . But we may have a reasonable guess since we know that the Riemannian manifold of the quantum computation, which has a very similar metric, has a curvature which is almost negative everywhere. Of course the metric of the CNN is more complex since here we also have the nonlinear component. Ifalso has a almost negative curvature under our first definition of the metric shown above, then our second scaled metric will result in a highly curved manifold with a large negative curvature depending on the size of kernels and the selection of the activation function. Roughly speaking, a larger kernel size and a stronger nonlinearity of the activation function lead to a higher negative curvature.

The target transformation
Different with the case of the quantum computation, where the algorithm is known, here we do not have . is found in a similar way as in the case of the template matching, where the target transformation is obtained by an optimization to minimize the error between the transformed source image and the destination image . In CNNs, we find in a similar way. The only difference is that now the optimization is carried out on an ensemble of source and destination points, i.e., the training data with their corresponding labeling. 
The curve
is a piecewise smooth curve consisting of L line segments, where the length of each line segment corresponds to the computational complexity of each layer. So geometrically a CNN is to find a curve with a given geometrical structure to connect and on .
The above given Riemannian is built on the transformation space . In fact just like the cases in both the quantum computation and the template matching, equivalently we can have another Riemannian manifold on the data space . We can regard the CNN achieves a change of the metric from the input data space to the final output data space defined by , where and are the metric on the input and output manifolds respectively and is the differential of the transformation
. So in this construction, the metric of the data space changes along the data flow in the network. As we indicated before, the input data can not be easily classified in the input data space by defining a simple distance between points, but it’s much easier in the output data space even with an Euclidean distance. This means by applying the transformation
, the CNN transforms a highly curved input data space to a flatter output data space by a curvature flow like mechanism. This gives an intuitive picture on how a complex problem can be solved with the CNN by flattening a highly curved manifold as in [23]. But it’s difficult to estimate how the curvature of the input data space except for some special cases as in
[10]. In the rest of this paper we will focus on the Riemannian structure on since it’s relatively easier to analysis the curvature of .We now show how the design of the CNN structure influence the geometry and how the geometric structure can affect the performance of CNNs.
First we need to clarify how the components of CNNs will affect the geometric structure.

Fully connected network
Usually there is a fully connected network at the final stage of CNNs. From existing experimental works we know that its main function is to collect multiple modes of a certain category of object. For example, if the task is to identify cats, the FC network is in charge of collecting the information of cats with different locations and gestures. Also we know with the increase of the kernel number and depth, the FC part can be omitted. In another word, the convolutional network before the FC network in fact has already achieved an over classification of the objects, the FC network only achieve a clustering of the information. So the FC network does not play an important role in the geometry of CNN. 
Depth of CNN
The depth of a CNN is related to the total length of the curve , i.e., the complexity of the CNN. Obviously a deeper CNN can potentially achieve more complex transformation in . 
CONV operation
The size and the number of kernel of the CONV operaion leads to different and as described above. Also as a component of the local CONVReLUPooling operation, CONV influences the complexity of local operators and the metric as well. Generally a bigger kernel size and kernel number increase the complexity of allowed local operators. This can be understood if we compare it with the Riemannian metric defined in the geometry of quantum computations. Also the CONV of each layer changes the structure of the curve . 
ReLU operation
The role of ReLU operation (or a general activation function) falls in two folds: Firstly it provides the necessary nonlinear component to achieve the target nonlinear transformation . Secondly different selections of the activation function show different nonlinear property, so it also influences the complexity of local operations and the Riemannian metric. The existing experimental works on the influence of the activation function on the convergence property of CNN indicate that it may have a very strong effect on the Riemannian metric, which leads to a change of the curvature of the Riemannian manifold . This observation helps to analysis the influence of the activation function on the convergence of the training procedure. 
Pooling operation
The function of the pooling operation is also two folds: On one hand, it can reduce the complexity by changing and . On the other hand it also plays a role in constructing the metric, correspondent to the kernel in the LDDMM framework. By using pooling operations, we can improve the smoothness and robustness of the curve since is estimated by a statistical optimization using samples from input image space, i.e., the training data. The pooling operation can be understood as an extrapolation of the samples so that the information of each sample is propagated to its near neighbourhood.
Given the above analysis, we can answer the following questions from a geometric point of view:

Why can CNNs work well in computer vision systems?
There are some theoretical considerations to answer this question. In [1] it’s claimed that the special structure of the image classification task enables the effectiveness of CNN, where the problem is formulated as the efficient approximation of a local Hamiltonian system with local operations of CNNs. By checking the analogy between the CNN and the quantum computation, we see that the argument of [1] is equivalent to say that for a task that can be solved by a deep CNN, its falls in the subset that can be achieved by local CONVReLUPooling operators, just as the subset of that can be approximated by simple local universal gates in quantum computations. So the answer to this question is trivial, CNN works since the problem falls in the set of solvable problems of CNNs. A direct question is, can we design a task that does not belong to this solvable problem set? A possible design of such a task is to classify images that are segmented into small rectangular blocks and then the blocks are randomly permuted. Obviously such images are generated by global operations which can not be effectively approximated by local operations, and therefore the classification of such images can not be achieved efficiently by structures like CNN. So we can say, the diameter of the Riemannian manifold is exponentially large with respect to the size of the input data and only those problems which can be solved by a transformation that has a polynomially large geodesic distance to the identity transformation can be solved by CNNs. 
Can we justify if a problem can be efficiently solved by CNNs? Generally this is difficult since this is the same question as to ask if a unitary operator can be decomposed into a tensor product of simple local unitary operators, or to ask if a nqubit quantum state has a polynomial state complexity with respect to
. We already know both of them are difficult questions. 
Why deep CNN?
Geometrically, training a CNN with a given deepth is equivalent to find a curve of length roughly proportional to in to reach the unknown transformation . In deep learning systems, this is achieved by a gradient descent based optimization procedure, i.e., the training of CNNs. It should be noted that the CNN is not to find a geodesic, instead it aims to find a curve with a fixed geometric structure, i.e., a curve consisted of multiple line segments with given lengthes determined by the CNN structure. The reason to use the deep CNN is just that the length of should be longer than the length of the geodesic connecting the identity transformation and on the Riemannian manifold . Otherwise it will not work since a short will never reach . On the other side, it’s not difficult to guess that a too deep CNN may also fail to converge if the curve is too long due to both the depth and an improper design of CONVReLUPooling operations. This can be easily understood since to reach a point with a longer curve is just to find an optima in a larger parameter space. Intuitively we prefer a design of CNNs that leads to a curve with a length just above the geodesic distance between and . To achieve this, we need a balance between the network depth and the complexity of CONVReLUPooling operations. This observation will play a key role in the analysis of ResNet below. 
Why does ReLU work better than the sigmoid in deep CNNs?
A typical answer to this problem is that ReLU can improve the vanishing gradient problem so that a better performance can be achieved. Another idea is that ReLU can change the Fisher information metric to improve the convergence but the sigmoid function can not. Here we explain this from a complexity point of view. As mentioned before, the selection of the activation function will influence the complexity of the local operation. Obviously ReLU is a weaker nonlinear function than the sigmoid. This means for a given CNN structure, using ReLU will have a lower representation capability than using the sigmoid so that the CNN using ReLU has a smaller complexity. This is to say that the a CNN using a sigmoid need to search for an optima in a larger state space. It’s obvious that as far as the complexity of the CNN is above the length of the geodesic, it’s preferable to use a simpler CNN so that a better convergence can be achieved. It seems there exists a balance between the complexity of the CONV operation and the activation function so that the complexity of the complete system can be distributed equally along the network. With our first definition of the metric, the intuitive geometric picture is that the sigmoid function will introduce a metric so that the curvature of the Riemannian manifold will be more
negative, if our guess of the curvature of is true. A higher negative curvature of course will make the curve more unstable so that a small change of the parameters of the CNN will result in a large deviation of the final point of . This corresponds to the observed vanishing gradient problem during the backpropagation procedure. So for a general deep CNN, the sigmoid is too complex and ReLU is a proper activation function that can produce a Riemannian manifold with a proper curvature. 
How difficult is it to optimize the structure of CNNs? By the optimization of the structure of CNNs, we mean to find the optimal CNN structure (for example the depth, the number and sizes of kernels of a CNN), which can accomplish the classification task with the minimal computational complexity. Geometrically this is to construct and define a metric on the Riemannian fold , under with the length of a curve is proportional to the computational complexity of a CNN structure; and (b)realize the geodesic that connects the identity operator with . Generally this is difficult, especially when the curvature property of the Riemannian manifold is unknown. This is the same as to find the optimal realization of an algorithm with a given universal logic gate set. For the geometry of quantum computation, the curvature is almost negative so that it’s difficult to achieve a general operator[10]. If the manifold of the fundamental law, the quantum computation, has a negative curvature, it will not be surprising if we assume the geometry of deep learning has a similar property. This point seems to be partially confirmed by the recent work[23], where a manifold with an exponentially growing curvature with the depth of a CNN can be constructed. If a CNN is applied to such an input data manifold with an exponentially large negative curvature, then the optimization will be difficult to converge.
IiiB Geometry of residual networks
ResNet is a superdeep network structure that shows superior performance than normal CNNs[24][25]. Given the geometry of CNNs, it’s straight forward to draw the geometric picture of ResNets.
Similar with CNNs, ResNets also aim to find a curve to reach . The difference is that consists of a large number of short line segments corresponding to simple operations . It’s easy to tell that when is big enough, is near the identity operation so that the firstorder approximation of gives , where is a weak local nonlinear transformation. This gives the structure of the ResNet, i.e., each layer of the ResNet is the addition of the original and the residual information.
The reason that ResNet works better falls in the following facts:

The ResNet achieves a roughly uniform distribution of the complexity of the problem along the curve
. Such a structure may lead to a smooth or stable information flow along the network and help to improve the convergence of the optimization procedure. 
The fact that reduces the searching space of . For example, if the space is , then is in the Lie algebra , which has a smaller dimension than . This may improve the performance of the optimization procedure.
We can also use this framework to analysis the properties of ResNets observed in [25]. In [25] the ResNet is formulated in a general form
(4)  
(5) 
where and are input and output of the thlevel and is a residual function. In the original ResNet, and .
The main results of [25] are:

If both the shortcut connection and the afteraddition activation are identity mappings, the signal could be directly propagated in both the forward and backward directions. Training procedure in general becomes ealier when the architecture is closer to these conditions.

When is selected far from the identity mapping, for example as multiplicative manipulations(scaling, gating, convolutions and dropout), the information propagations is hampered and this leads to optimization problems.

The preactivation of ReLU by putting the ReLU as part of and setting as the identity operation improves the performance. But the impact of is less severe when the ResNet has fewer layers.

The success of using ReLU in ResNet also confirms our conclusion that ReLU is a proper weak nonlinear function for deep networks and the sigmoid is too strong for deep networks.
From the geometric point of view, the above mentioned observation are really natural.
For a deep ResNet, , this is exactly the case that both and are identical mappings. If is far from the identity mapping, then the difference between the identity mapping and need to be compensated by , this will make more complex and leads to optimization problems.
The preactivation also aims to set as the identity mapping and the weak nonlinear ReLU is absorbed in . If the ResNet has fewer layers, then is a little far from the identity mapping so that is valid. This can explain the observation that the preactivation is less crucial in a shallower ResNet.
The geometric picture easily confirms that the best ResNet structure is to replicate . Another related observation is that the highway network[26] usually performs worse than the ResNet since it introduces unnecessary complexity to the system by adding extra nonlinear transformations to the system so the structure of the highway network does not match the natural structure given by the firstorder approximation.
IiiC Geometry of recursive neural networks
Recursive neural network is commonly used in the natural scene images or natural language processing tasks.
Taking the recursive neural network for the structure prediction described in [27][28] as an example, the goal is to learn a function , where is the set of all possible binary parse trees. An input consists of (a) a set of activation vectors which represent input elements such as image segments or words of a sentence, and (b) a symmetric adjacency matrix to define which elements can be merged.
The recursive neural network accomplishes this task by finding the two sub mappings: (a) A mapping from the words of natural languages or segments of images to the semantic representation space. Combining this map with the adjacent matrix, the original natural input data is represented in a space ; (b)A fixed rule to recursively parse the input data to generate a parsing tree.
Geometrically we are interested in the following aspects:

What’s the correspondent geometric picture of the fixed recursive rule?

Why both the recursive rule and the representation mapping can be learned during training?
It’s easy to see that given the representation mapping and the adjacent matrix, the recursive operation is nothing but a transformation on . The fixed operation rule reminds us its similarity with the Lie group exponential mapping. Though we can not say is a Lie group, but conceptually the recursive neural network can be understood as a discretized Lie group exponential mappings. Given any input in , the transformation between and is achieved by this Lie group exponential mapping.
In CNN and ResNet, the goal is to find a free curve with a fixed length, which is neither a geodesic nor a Lie group exponential mapping. So the complexity or the dimension of freedom of the curve is much higher. In recursive neural networks, the exponential mapping like curve is much simpler, this is why we have the capability to optimize both the representation mapping and the exponential mapping simultaneously during training.
We can also observe that the recursive neural network does not emphasis on local operations as in CNNs or ResNets. This means that the recursive neural network is dealing a class of problems that is similar with the template matching problem. And its analogy with the SVF framework of the template matching confirms this point.
IiiD Geometry of recurrent neural networks
Another category of neural networks is the recurrent neural networks[29]. The key feature of the RNN is that it allows us to process sequential data and exploit the dependency among data. RNNs are widely used in language related tasks such as the language modeling, text generating, machine translation, speech recognition and image caption generation. Commonly used RNN structures include bidirectional RNNs, LSTM networks and deep RNNs.
Given a sequence of input data , a standard RNN compute the hidden vector sequence and the output sequence for every time step as:
(6)  
(7) 
where and
are the weight matrices and the bias vectors.
is the activation function of the hidden layer.Here we are not going to give a complete overview to the different variations of RNNs. Instead we will try to understand the feature of RNNs from a geometrical point of view.
Let’s first go back to the geometry of CNNs, where a CNN is a curve on to reach a point . For each given data point on , it’s transformed by to another point on . Geometrically a CNN accomplishes a pointtopoint mapping with a simple curve as its trajectory on .
RNN is more complex than CNN. We can check the geometric picture of RNNs in two different ways.
The first picture is to regard the sequential input data as a series of points on , then the geometric picture of a RNN is a mapping of string, a series of interconnected points , to a destination string with an intermediate trajectory . This is not a collection of independent parallel trajectories as since they are coupled by the recursive structure of the RNN through the hidden states in (6).
Another picture of RNNs is to regard the sequential data, the string in the first picture, as a single data point, then the trajectory of the input data is a curve in a higher dimensional space . Accordingly is the space of transformations on . This is similar with the geometric picture of CNNs.
The difference of these two pictures is that in the first picture, the transformation working on the trajectory of each point of the string can be any transformation in and the trajectories of different points are coupled with each other. In the second picture, the transformation working on the data point, the complete sequence of data , is now local operations in , which only work on each individual data and pairs of successive data also through as in LABEL:. So the second picture is somehow trivial since now the RNN is only a special case of the CNN. So we from now on we will focus on the first picture to see how this picture can help us to understand the RNN related topics.
For a standard RNN, only earlier trajectories will influence later trajectories so that a later input data feels the force of earlier data. For bidirectional RNNs, the force is bidirectional. For deep bidirectional RNNs, the trajectory is a complex sheet consisting of multiple interconnected trajectories.
This stringsheet picture reminds us of their counterparts in physics. A deep CNN corresponds to the worldline of a particle, which can achieve human vision functions. A deep RNN corresponds to the worldsheet of a string, which can be used to solve linguistic problems. It’s natural to ask, what’s the correspondent structure of the trajectory of a membrane and what kind of brain function can be achieved by such a structure? Just as that RNNs can interconnect points into a string, we can also interconnect multiple strings to build a membrane. We can even further to interconnect multiple membranes to generate higher dimensional membranes. To build such structures and investigate their potential functionalities may lead to interesting results.
We show that in fact the LSTM, the stacked LSTM, the attention mechanism of RNN and the grid LSTM are all special cases of this membrane structure.

LSTM
The structure of the standard LSTM is shown in Fig. 8. It’s straight forward to see that it’s the evolution of two parallel but coupled strings, the hidden states and the memory . Compared with normal RNNs where there is only a single information passage, it’s possible that the existence of the parallel information passages in LSTM makes the LSTM system capable of modeling long range dependencies. 
Stacked LSTM
The stacked LSTM is in fact a deep LSTM whose geometric picture is a cascaded double string as in Fig. 8. 
Grid LSTM
The grid LSTM[30] is an extension of the LSTM to higher dimensions. Accordingly the strings in the standard LSTM are replaced by higher dimensional membranes as given in [30], where a 2d deep grid LSTM can be understood as stacked multiple 2d membranes. So if the deep LSTM is a cascaded coupled double strings, then the deep 2d grid LSTM is a cascaded membranes (Fig. 10). Similar geometric pictures can be constructed for higher dimensional and deep grid LSTM systems. 
Attention mechanism
The attention mechanism is getting popular in the decoder of the machine translation and the image caption generation. The key feature is that during the text generation, an alignment or an attention model (a feedforward network in
[31]or a multiple layer perceptron network in
[32]) is used to generate a timevarying attention vector to indicate which part of the input data should be paid more attention to. This structure can be understood as a coupling between the normal decoder network (a RNN or a LSTM) with the attention network as shown in Fig. 9.
It need to be noted that there is an essential difference between the LSTM systems and the attention mechanism. We can see that the LSTM systems are the coupling of homogeneous subsystems but the attention mechanism is in fact a coupling of two heterogeneous subsystems, i.e. one decoder RNN/LSTM network and one attention model with a different structure. Other similar systems, such as the recurrent models of visual attention[33]
and the neural Turing machines
[34][35], also hold a similar picture of coupling heterogeneous subnetworks.It can be expected that more complex and powerful systems can be constructed by such a mechanism. It will be interesting to see how such a mechanism may help us to understand key human intelligence components such as our memory, imagination and creativity.
Similar with the case of CNNs, we can also check the metric and the curvature of the manifold of the coupled networks. Roughly we are now working with the direct product of the manifold of each subsystem with the tangent spaces of the subsystems as orthogonal subspaces of the tangent space of the complete manifold. This is very similar with the case of the semidirect group product in the metamorphsis framework in the template matching system[18]. It can be expected that the structure of the manifold is much more complex than the relatively simple CNN case. But if the curvature of each subsystem’s manifold is almost negative as in the quantum computation, then the convergence of the training of the coupled system naturally will be more difficult, just as indicated in the case of neural Turing machines.
IiiE Geometry of GANs
Generative adversarial networks (GANs) is a framework for unsupervised learning. The key idea of GANs is to train a generative model that can approximate the real data distribution. This is achieved by the competition between a generative model
and a discriminative model . In the standard configuration of GANS, both and are CNNlike networks.We now give a geometrical description of GANs and try to answer the following questions:

Why is it difficult to train GANs?

Why can the refined GANs, including LAPGANs, GRANs and infoGANs, perform better than standard GANs?
Since both and are realized by CNNs, the geometric picture of GANs can be built on the geometry of CNNs. From the geometric model of CNNs, it’s easy to give the geometric picture of the standard DCGANs as a twopiece curve in as shown in Fig 11.
The two curve pieces represent the generator and the discriminator respectively. The goal of the training procedure is to find the two transformations that can minmax the cost function of GANs. GANs are different from CNNs in the following aspects:

The goal of CNNs is to find a curve that can reach a transformation from . For GANs, we need to find two transformations and by constructing two curves to reach them from simultaneously.

The target transformation of is relatively simple but the destination transformation of is much flexible than . This can be easily understood since we set more constraints on by labeling the training data in CNN based object classification systems. An easy example to illustrate this point is shown in Fig. 12. Both CNNs and the inverse of G aim to transform the training data into an Euclidean space so that the training data are properly clustered. For CNNs, the labeling of training data results in that the output clustering pattern of CNNs is relatively unique. But for the unsupervised learning of the generator of GANs, their are a set of equivalent clustering patterns, which may compete with each other during the training procedure.

The information pathways are also different in GANs and CNNs as shown in Fig. 11. Firstly we can see that in the design of GANs, the basic idea is to learn a representation of the training data with so that the generated examples from can have the same distribution of the training data, i.e. the generated data can fool the discriminator . So the information flow in this design is to start from the training data, go through the reverse generator , return back through and pass the discriminator . Only at this point, the output of can send error information back to guide the update structure of . During the network training, the information flow for is similar with a normal CNN. But the optimal is defined by the cost function of GANs, which need an information passage passing both and to set constraints on , which is longer than the CNN case.

The geometric picture of GANs is to find a twosegment curve connecting the input space of to the output space of , which is asked to pass the neighbourhood of the training data in a proper way. But we can only use the information from the output of to justify if the curve fulfills our requirement. So any loss of information at the output of , due to either the mismatch between the training of and or an improper cost function that fails to measure the distance between the curve and the training data, will deteriorate the convergency of GANs.
Given the above observations, the reason that GANs are difficult to be trained is straight forward.

Compared with CNNs, GANs have the same Riemannian manifold as , which has a high possibility of owning a negative sectional curvature. GANs have to optimize two curves and a more flexible target transformation on this manifold.

The information pathways are longer in GANs than in CNNs. Also the information pathways in the training procedure do not coincide with the information pathways in the designing idea of GANs. This mismatch may lead to a deterioration of the system performance since the information flow in the direction is lost in the training procedure.
To eliminate the difficulties of GANs, there exist the following three possible strategies:

To decrease the flexibility of the generator transformation by adding more constraints to .

To improve the structures of the curves of and . From the analysis of CNNs, we know this is to change the metric of the manifold or equivalently to rearrange the lengths of the line segments in the curves of G and D.

To eliminate the information loss at the output of so that it’s more reliable to justify how the twosegment curve of GAN passes the neighbourhood of the training data. Basically this is exactly what the WGAN does.

To change the information pathways during the training procedure. This may improve the efficiency of the information flow during the training so that possibly a better convergence may happen.
Now we explain how LAPGANs, GRANs and infoGANs can improve the performance of GANs utilizing the above given strategies.

InfoGAN
InfoGANs aim to decrease the flexibility of the transformation or equivalently by introducing extra constraints on . The original GAN cost function is based on the sample based expectation of the network outputs. On the contrary, infoGANs enhance the constraints on by setting constraints on expectations of clusters of sample data. This means that the clustering pattern in the representation space of the training data, indicated by the latent variables in the input data space of the generator , should be kept by the generator, so that in the hidden states of the discriminator (as a kind of decoder of the generator to reveal the clustering pattern) the same clustering pattern can be observed. This is exactly why the infoGAN add a mutual information item, which is used to check if the clustering patten, to its cost function. This cluster based constraints in fact constructs a hierarchical structure in both the training data space and the representation space. Asking the generator G to keep this hierarchical structure effectively reduce the flexibility of and improves the convergence stability. 
LAPGAN
The key feature of LAPGANs is to decompose the image space into orthogonal subspaces with different frequency bands and run different GANs sequentially on subspaces at higher frequency bands conditioned on the results on lower frequency bands. From our geometrical point of view, the decomposition of input data space results in a correspondent decomposition of the transformation space into multiple subspaces where with the number of bands of the paramid. So the pyramid based LAPGAN is essentially to construct the curve and in by a set of curves in each subspace using simpler local operations. This is similar to achieving a displacement in 3 dimensional space by three displacements in x, y and z directions.The reason such a structure can improve the convergence of LAPGANs can be understood from different points of view.
Firstly LAPGANs replace a complex curve by a set of simpler curves in each orthogonal subspace and then optimize those curves sequentially. So the local operations constructing the curve are simpler and the curve is also shorter than . From our above analysis of CNNs and ResNets, simpler local operations will introduce a flatter transformation manifold or equivalently a curve consisted of shorter line segments. This will improve the convergence property of each curve . This is to say a curve with complex local operations by a series of shorter curves with simpler local operations in subspaces .
Secondly in LAPGANs, a set of generators and discriminators are trained with each subband of training data. Obviously each pair of and determines a representation pattern of one data subband. Also these representations are consistent since during the training procedure, the generator at a higher frequency band is conditioned on the lower band image. This means LAPGANs replace the original cost function on the expectation of the complete data set by a set of cost functions on the expectation of each frequency band of the input data. This is in fact to construct a hierarchical structure on the training data, i.e. the paramid, and set a set of consistent constraints on this hierarchical structure. So essentially we add more constraints to the generator transformation .

GRAN
GRANs improve the performance of GANs by a RNN based generator so that the final generated data is constructed from the sequential output of the RNN network. We already know that the RNN essentially decompose a single input data point of a normal CNN network into a data sequence and therefore the trajectory of a data point of CNNs becomes a sheet swept by a string in RNNs. From this point of view, the operations of a RNN are essentially all simpler local operations of the complete input data. What’s more wimilar with LAPGANs, GRANs also explore a hierarchical structure of the system which is encoded in the RNN structure. But this hierarchical structure is not so explicit as in LAPGANs. In another word, LAPGANs explore the structure in frequency domain and GRANs explore a kind of selfsimilarity like structure of the input data.
As a summary, LAPGANs, GRANs and infoGANs all improve the performance by constructing a hierarchical structure in the system and adding more constraints on . Their difference lies in their ways to construct the hierarchy. What’s more, LAPGANs and GRANs can potentially further improve the performance by changing the structure of the curve by only using simpler local operations compared with the standard GANs.
Another possible way to improve the performance is to change the information passway. For example in the training of and , the training data information is fed to D directly but to G indirectly. Feeding the training data directly to G in the inverse direction (as the information passway given by the black dash arrow in Fig. 11) can potentially further explore the structure of the representation space .
IiiF Geometry of equilibrium propagation
Equilibrium propagation[36]
is a new framework for machine learning, where the prediction and the objective function are defined implicitly through an energy function of the data and the parameters of the model, rather than explicitly as in a feedforward net. The energy function
is defined to model all interactions within the system and the actions with the external world of the system, where is the parameter to be learned, is a parameter to control the level of the influence of the external world, is the state of the external world(input and expected output) and is the state of the system.The cost function is define by
(8) 
where is a directional vector in the space of so the cost function is the directional derivative of the function at the point in the direction .
For fixed , and , a local mininum of which corresponds to the prediction from the model, is given by
(9) 
Then the equilibrium propagation is a twophase procedure as follows:

Run a 0phase until the system settle to a 0fixed poin and collect the statistics .

Run a phase for some small to a fixed poin and collect the statistics .

Update the parameter by
(10)
An equivalent constrained optimization formulation of the problem is to find
(11) 
subject to the constrain
(12) 
This leads to the Lagrangian
(13) 
By this formulation we can see the update of in equilibrium propagation is just one step of gradient descent on with respect to as
(14) 
where is determined by
(15) 
and
(16) 
It’s claimed that the phase can be understood as performing the backpropagation of errors. Because when a small on is introduced, the energy function induces a new ’external force’ acting on the output units of the system to drive the output to output target. This perturbation at the output layer will propagates backward across the hidden layers of the network so that it can be thought of a backpropagation.
Here we will show that the geodesic shooting algorithm[20][19] in the diffeomorphic template matching is in fact an example of the equilibrium propagation. We will use it to give a geometrically intuitive picture of the equilibrium propagation to show how the backpropagation is achieved in the phase.
The geodesic shooting addresses that diffeomorphic template matching as find the optimal matching between two images and . Under the smooth transformation of , we define the transformed source image as . The energy function is given as
(17) 
In the language of the equilibrium propagation, corresponds and as the input and output data, the state of the system is given by ). The cost function is .The prediction of the state is obtained by , which leads to the EulerPointcare equation given by
(18)  
(19)  
(20)  
(21) 
where is the momentum map and is the adjoint variable of . It’s easy to know that the optimal state is completely determined by , which can be regarded as the parameter to be optimized.
Taking the EP equation as the constraint, the variation of 13 results in
(22)  
(23)  
(24)  
(25)  
(26)  
(27)  
(28)  
(29)  
(30) 
where is the momentum map of .
In the geodesic shooting framework, the above equations include two components:

Shooting The first three equations determine given the initial momentum and .

Adjoint advection
The second five equations determines by solving the equations in the reverse time direction given the boundary conditions of . 
Gradient
The last equation given the gradient of the cost function on the initial momentum , while updating equals to update the parameter .
It’s very easy to see that the three components correspond exactly to the three step in the equilibrium propagation. The shooting and advection phases correspond to the 0phase and phase. The gradient is equivalent to the gradient descent .
This example gives an intuitive picture of the equilibrium propagation as follows:

The 0phase is to shoot a geodesic staring from a given point with a given initial velocity vector.

The phase is to propagate the error information (the distance between the end point of the geodesic and the expected destination ) in a backward direction.

The update of parameter is based on the information obtained by the backpropagation procedure.
Of course, the equilibrium propagation is a general framework for deep learning. We can not give a concrete geometric description of it due to its flexibility. Here we just show that in some special cases, the equilibrium propagation does have a geometric picture, in which how the backpropagation of error information can be directly observed.
Another observation is that the equilibrium propagation is not only a framework for training the so called implicitly defined deep learning systems as described in [36], in fact it’s also potentially a framework to optimize the structure of the explicitly defined deep learning systems as well.
To see this point, taking the deep CNN as an example where the prediction is explicitly defined by the structure of the CNNs. We can define an energy as
(31) 
where the parameter to be learned includes also the parameters of the structure of a CNN, for example the depth, the number and size of each layer, besides the normal papameters of a CNN. is then the cost function for a CNN with a given structure defined by and defines the complexity of the CNN, i.e., the number of operations of the feedforward network structure defined by . does not depend on since the state of the CNN is determined by and .
If we run the equilibrium propagation on this system, the system prediction can be achieved by the current gradient descent based CNN training method, i.e., training a CNN with a fixed configuration. Then the optimization on is in fact a global optimization framework to find the optimal CNN structure defined by . Geometrically this is to find a Riemannian manifold with its metric defined by , so that the optimal CNN can realize a geodesic , which gives a minimal system complexity, i.e., the most efficient CNN structure that can achieve the task.
Of course, the discrete parameters of the structure of the CNN can not be updated by the gradient descent strategy. This is the same as in the reinforcement learning neural Turing machines
[34] if we regard the discrete structure parameters as the actions in the neural Turing machines. So theoretically we can train the systems also with the reinforcement learning to find the optimal structure parameters, while the backpropagation in [34] is replaced by the equilibrium propagation. Also this framework might not work due to the complexity in practice, just as the neural Turing machine can only achieve simple algorithmic tasks.Iv Conclusional remarks
This paper is based on the belief that geometry plays a central role in understanding physical systems. Based on a brief overview of the geometric structures of two fields, the quantum computation and the diffeomorphic computational anatomy, we show that deep learning systems hold a similar geometric picture and the properties of deep learning systems can be understood from this geometric perspective.
We summarize our observations on different deep learning systems as follows:
Convolutional neural networks

The structure of a CNN defines a manifold, the transformation space , and the Riemannian metric on it.

The goal of a CNN is to find a transformation on this Riemannian manifold, which is achieved by a curve defined by the local operations of the CNN structure.

A CNN is not to find the geodesic but to find a curve with a given structure to reach a point in the Riemannian manifold. A deep CNN works only when the curve it defines is longer than the geodesic between the identity operator and . A too shallow network will never work efficiently.

The convergence property of a CNN depends on both the curvature of the Riemannian manifold and the complexity distribution of the curve. The linear complexity defined by the configuration of kernels and the nonlinear complexity defined by the activation function should be balanced. A too deep network with improper complexity distribution may also fail.

Generally to find the optimal CNN structure is difficult.

How the structure of a CNN influence the curvature of the manifold and the performance of the network needs to be investigated.

An alternative geometric structure is built on the data space , where a CNN works by disentangling a highly curved input manifold into a flatter output manifold.
Residual networks

Residual networks is a special case of CNNs, where the curve to reach the target transformation consists of a large number of short line segments.

Each line segment corresponds to a nearidentity transformation on the manifold, which can be achieved by the oneorder approximation. The optimal ResNet structure is determined by this oneorder approximation.

The superior performance of ResNet comes from the balanced complexity distribution along the curve and a potentially smaller parameter space defined by the ResNet structure.
Recursive neural networks

Recursive neural networks have a structure that can be understood as a Lie group exponential mapping if the manifold of transforamtions is regarded as a Lie group.

The relative simple structure of the recursive neural network makes it possible to find the representation mapping and the recursive operation during the training procedure.
Recurrent neural networks

Recurrent neural networks can be understood as a sheet swept by a string on the data space .

A stacked or deep LSTM network accomplishes a cascaded structure of a pair of coupled strings.

The grid LSTM network extends the coupled stings into weaved membranes.

The attention mechanism and the neural Turing machine are coupled structure of heterogeneous subnetworks.
Generative adversarial networks

Generative adversarial networks are the simultaneous optimization of two curves leading to and , in , where each curve has a similar structure as CNNs. The unsupervised learning of the target transformation
is much flexible than than the supervised learning of
. These features explain the difficulty of training GANs. 
LAPGANs achieve a better performance by both exploring the frequency domain structure of the training data and constructing the curve by simpler local operations on subspaces of the training data.

GRANs also work in a similar way as LAPGAN. The difference is that GRANs construct a selfsimilarity like structure which is encoded in the RNN structure of . Also GRANs only apply simpler local operations in .

InfoGANs reduce the flexibility of by setting extra constraints on clusters of data samples.
Equilibrium propagation

The geodesic shooting algorithm of the template matching is an example of the equilibrium propagation. It provides an intuitive picture of how the equilibrium propagation can fill the gap between the implicit framework of optimization of a cost function and the backpropagation of information in the explicit system.

Theoretically the equilibrium propagation can be used as a framework for find the optimal deep learning system.
References
 [1] H. W. Lin and M. Tegmark, “Why does deep and cheap learning work so well?,” arXiv:1608.08225, 2016.
 [2] A. Ashtekar and T. A. Schilling, “Geometrical formulation of quantum mechanics,” arXiv:grqc/9706069, 1997.
 [3] H. Heydari, “Geometric formulation of quantum mechanics,” arXiv:1503.00238, 2015.
 [4] O. Andersson and H. Heydari, “Geometry of quantum evolution for mixed quantum states,” Physica Scripta, arXiv:1312.3360v1, 2014.
 [5] P. Levay, “The geometry of entanglement: metrics, connections and the geometric phase,” Journal of Physics A: Mathematical and General, arXiv:quantph/0306115v1, 2004.
 [6] P. Levay, “The twistor geometry of threequbit entanglement,” arXiv:quantph/0403060v1, 2003.
 [7] M. V. Raamsdonk, “Building up spacetime with quantum entanglement,” International Journal of Modern Physics D, arXiv:1005.3035v1, 2010.
 [8] D. Stanford and L. Susskind, “Complexity and shock wave geometries,” arXiv:1406.2678v2, 2014.
 [9] L. Susskind and Y. Zhao, “Switchbacks and the bridge to nowhere,” arXiv:1408.2823v1, 2014.
 [10] M. G. M. A. Nielsen, M. R. Dowling and A. C. Doherty, “Quantum computation as geometry,” Science 311,1133, 2006.
 [11] M. R. Dowling and M. A. Nielsen, “The geometry of quantum computation,” Quantum Information and Computation, vol. 8, no. 10, pp. 861–899, 2008.
 [12] M. F. Beg, M. I. Miller, A. Trouve, and L. Younes, “Computing large deformation metric mappings via geodesic flow of diffeomorphisms,” Int. Journal of Computer Vision, vol. 61, pp. 139–157, 2005.
 [13] M. Hernandez, M. Bossa, and S. Olmos, “Registration of anatomical images using paths of diffeomorphisms parameterized with stationary vector flows,” International Journal of Computer Vision, vol. 85, pp. 291–306, 2009.
 [14] M. Lorenzi and X. Pennec, “Geodesics, parallel transport and oneparameter subgroups for diffeomorphic image registration,” International Journal of Computer Vision, pp. 12:1–17, 2012.
 [15] S. Durrleman, X. Pennec, A. Trouvé, G. Gerig, and N. Ayache, “Spatiotemporal atlas estimation for developmental delay detection in longitudinal datasets,” in MICCAI, vol. 12 of 297304, 2009.
 [16] D. D. H. M. Bruveris, F. GayBalmaz and T. Ratiu, “The momentum map representation of images,” Journal of Nonlinear Science, vol. 21, no. 1, pp. 115–150, 2011.
 [17] M. Bruveris and D. Holm, “Geometry of image registration: The diffeomorphism group and momentum maps,” Lecture Notes, 2013.
 [18] D. D. Holm, A. Trouve, and L. Younes, “The eulerpoincare theory of metamorphosis,” Quarterly of Applied Mathematics, vol. 67, pp. 661–685, 2009.
 [19] F. X. Vialard, L. Risser, D. Rueckert, and D. D. Holm, “Diffeomorphic atlas estimation using geodesic shooting on volumetric images,” Annals of the BMVA, vol. 5, pp. 1–12, 2012.
 [20] F. X. Vialard, L. Risser, D. Rueckert, and C. J. Cotter, “Diffeomorphic 3d image registration via geodesic shooting using an efficient adjoint calculation,” Int. Journal of Computer Vision, vol. 97, pp. 229–241, 2012.
 [21] F. X. Vialard, L. Risser, D. D. Holm, and D. Rueckert, “Diffeomorphic atlas estimation using karcher mean and geodesic shooting on volumetric images,” in Medical Image Understanding and Analysis, 2011.
 [22] M. M. Postnikov, Geometry VI: Riemannian Geometry, Encyclopedia of mathematical science. Springer, 2001.
 [23] B. Poole, S. Lahiri, M. Raghu, J. Sohldickstein, and S. Ganguli, “Exponential expressivity in deep neural networks through transient chaos,” 2016.
 [24] S. R. K. He, X. Zhang and J. Sun, “Deep residual learning for image recognition,” arXiv:1512.03385, 2015.
 [25] S. R. K. He, X. Zhang and J. Sun, “Identity mappings in deep residual networks,” arXiv:1603.05027, 2016.
 [26] K. G. R. K. Srivastava and J. Schmidhuber, “Highway networks,” arXiv:1505.00387, 2015.
 [27] A. Y. N. R. Socher, C. Lin and C. D. Manning, “Parsing natural scenes and natural language with recursive neural networks,” in Proceedings of the 28 th International Conference on Machine Learning, 2011.
 [28] R. Socher, A. Perelygin, J. Wu, J. Chuang, C. D. Manning, A. Ng, and C. Potts, “Recursive deep models for semantic compositionality over a sentiment treebank,” 2013.
 [29] J. B. Z. C. Lipton and C. Elkan, “A critical review of recurrent neural networks for sequence learning,” arXiv:1506.00019, 2015.
 [30] I. D. N. Kalchbrenner and A. Graves, “Grid long shortterm memory,” arXiv:1507.01526, 2015.
 [31] R. K. K. C. A. C. R. S. R. Z. K. Xu, J. L. Ba and Y. Bengio, “Show, attend and tell: neural image caption generation with visual attention,” 2015.
 [32] K. C. D. Bahdanau and Y. Bengio, “Neural machine translation by jointly learning to align and translate,” arXiv:1409.0473, 2014.
 [33] A. G. V. Mnih, N. Heess and K. Kavukcuoglu, “Recurrent models of visual attention,” pp. 2204–2212, 2014.
 [34] G. W. A. Graves and I. Danihelka, “Neural turing machines,” arXiv:1410.5401, 2014.
 [35] W. Zaremba and I. Sutskever, “Reinforcement learning neural turing machines  revised,” arXiv:1505.00521, 2015.
 [36] B. Scellier and Y. Bengio, “Equilibrium propagation: Bridging the gap between energybased models and backpropagation,” arXiv:1602.05179, 2016.
Comments
There are no comments yet.