1. Introduction
1.1. Introduction.
Let be a simple, undirected, connected tree on vertices . The degree matrix is the diagonal matrix , the adjacency matrix encodes the connections between the vertices. The matrix
is known as the Laplacian matrix of . It is symmetric and has eigenvalues that we order by their size
We refer to [9, 11, 22] for an introduction to the spectral theory on graphs. It is not difficult to see that the unique eigenvector associated to the eigenvalue is the vector having constant entries and that
(1) 
This shows that if and only if is connected. The eigenvector associated to the second smallest eigenvalue is also known as the Fiedler vector [13, 14, 15, 16, 30]. The following crucial result is due to Fiedler [15].
Theorem (Fiedler).
The induced subgraph on is connected.
This, together with many other desirable properties, motivates the classical spectral cut whereby the sign of is used to decompose a graph. Overall, relatively little seems to be known about the actual behavior of the Fiedler vector:
However, apart from the original results from M. Fiedler, very few is known about the Fiedler vector and its connection to topological properties of the underlying graph […] ([16], 2018)
1.2. The Problem
Let be a tree. Equation (1) suggests that is the ‘smoothest’ vector that is orthogonal to the constants, so maximum and minimum of should be attained at the points of largest distance. This was explicitly conjectured in [10], a counterexample was then produced by Evans [12] and is shown in Figure 1. Maximum and minimum are assumed far away from one another but not at the two points of maximum distance from one another. A natural question that remains is (1) to understand the behavior of the Fiedler vector and, as discussed by Lefèvre [20] and Gernandt & Pade [16], (2) to understand under which conditions such a result might still be true.
We hope that our paper contributes to both problems and can be helpful in clarifying the situation; we also raise a number of questions.
1.3. The Continuous Case.
Let
be a domain and consider the second smallest eigenfunction of the Laplace operator
with Neumann boundary conditions, i.e. the equationRauch conjectured in 1974 that the maximum and the minimum are assumed at the boundary. This is known to fail at this level of generality [6, 7] but widely assumed to be true for convex domains . The second author [29] recently showed that if satisfy , then every maximum and minimum is assumed within distance of and , where is a universal constant (which is the optimal scaling up to the value of ). Therefore, up to an inradius, the maximum and minimum are essentially assumed at maximum distance.
2. Main Results
We present two main results: the first is a representation formula for that is very useful. It allows to quickly recover some of the existing results and gives a better understanding of the behavior of . In particular, it will explain that for generic trees there is little to no reason to assume that the extrema of are assumed at vertices that are at distance . We hope that the representation formula will also be useful in other settings. The second contribution is an explicit application of the representation formula to construct families of trees on which the desired statement indeed holds: the extrema of the Fiedler vector are assumed at vertices which are distance apart.
2.1. A Representation Formula
Let us fix to be a Graph on vertices. Let be two arbitrary vertices. We introduce a game that results in a representation formula for any eigenvector associated to the eigenvalue .

You start with zero payoff and in the vertex .

If you find yourself in a vertex ,

add to your payoff and

then jump to a randomly chosen neighbor of .


If you find yourself in the vertex , the game ends.
We note that the game ends in finite time almost surely.
Theorem 1.
The expected payoff of the game satisfies
Throughout the rest of the paper, we will only use this result for (since only then do we have Fiedler’s theorem at our disposal). Results of such a flavor are very easy to obtain for the random walk normalized Laplacian (which, by its very nature, is strongly tied to random walks). Our game adjusts for the canonical Laplacian . The game could also be interpreted as a discretized version of the FeynmanKac formula (see e.g. [31]). We believe that this representation theorem has a substantial amount of explanatory power. A simple example is the following (known) Corollary: a simple form of monotonicity of the second eigenvector along paths in a tree.
Corollary 1 (see also [16, 20]).
Let be a path in the tree such that only assumes positive values on the path. If are two vertices on the path and is at a greater distance than from the closest vertex where is negative, then
In particular, maxima and minima are assumed in vertices with degree 1.
Proof.
Let be the vertex where is negative but where is positive for one of the neighbors. By Fiedler’s theorem, that vertex is unique. Let us now assume and are vertices on a path and is from a greater distance from than .
We use Theorem 1 and start the game in the vertex . We then collect payoff in every run of the game and thus in expectation and therefore as desired. ∎
2.2. How to produce counterexamples.
We now return to the original question from [10]: whether the second eigenvector assumes maximum and minimum at the endpoints of the longest path in the tree. This was disproven by Evans [12] by means of an explicit example. The question has also been studied by [16, 20]
. The purpose of this subsection is to argue that our representation formula from Theorem 1 allows to heuristically explain why, generally, there is no reason the second eigenvector should assume extreme values at the endpoints of the longest path – it shows that Evans’ counterexample is actually representative of one of the main driving forces behind localization of large values of the Fiedler vector in substructures. We are not making any precise claims at this point, this section tries to provide a good working heuristic that (a) allows to contruct counterexamples quite easily and (b) will underlie all our formal arguments later in the paper.
One of the crucial ingredients in the representation formula is the number of steps a typical random walk will need to reach another vertex: if the tree has a complicated structure (say, many vertices with large degree), then it will take a very long time for a random walk to reach a specific vertex (this principle is already embodied in Evans’ counterexample [12] shown in Figure 1). Put differently, distance is not so crucial as complexity – this immediately implies a large family of counterexamples whose type is shown in Figure 4: we consider a path graph of length and attach a tree to vertex . We assume the tree has diameter much smaller than but has vertices of very large degree.
The game then suggests that, if the tree has vertices of sufficiently large degree, one extremum is at the ‘most remote’ part off the path in the tree – in particular, one of the two extrema would not be on the path and thus not at the endpoints of the longest path in the Graph. The distance between the extrema would be
A sketch of the argument to show that this type of construction works is as follows. There are two cases: either the sign change of the second eigenvector happens inside or it happens on the path. If the diameter of is sufficiently small compared to , known inequalities on the eigenvalue (which we use below) suggest that the first case cannot occur. This means that the sign change happens on the path. If the value of the eigenvector in the vertex that connects to is nonzero, we can play the game with vertices starting in and ending in . Corollary 1 shows that the values of the eigenfunction inside are (in absolute value) at least as big as the value in . Then the game leads to a nonzero contribution for each step of the random walk that is not arbitrarily small. This means that in order to ensure large (absolute) values inside the tree, the quantity to maximize is the expected number of steps in the game – this, in turn, can be achieved by having vertices of large degree. We emphasize that this heuristic is nonrigorous but quickly motivates the construction of many counterexamples. All the positive results in our paper can be understood as ensuring the absence of such a structure.
To build further intuition, we quickly sketch another type of counterexample. Take a path graph of length and add a path graph of length to the middle vertex. What we observe is that the eigenvector changes along the long path, that it assumes extrema at its end and that the eigenvector is small and changes slowly on the little path in the middle.
However, if we start adding paths of length 1 to the vertices of the short path in the middle (or short trees, even ones with bounded
diameter), then after a while the eigenvector flips and assumes an extremum in the tip of short path in the middle.
Perhaps the main contribution of our paper is a framework that clearly establishes why this happens. The
theorems we give are one way of capturing the phenomenon but presumably there are many other possible formulations
that could be proven by formalizing the same kind of mechanism that we use here.
A particular consequence of these ideas is that a generic tree should not have the desired property of assuming its extrema at the endpoints of a path of length . We refer to numerical work done by Lefèvre [20] showing that all trees with vertices do have the property but already of trees with vertices do not. Lefèvre specifically asks whether a typical tree on vertices does not have the property as becomes large and we also consider this to be an interesting problem.
2.3. An Admissible Class.
The purpose of this section is to construct a large family of treelike graphs for which the following statement is true: the second eigenvector of the Graph Laplacian does indeed assume maximum and minimum at the endpoints of the longest path. We assume that we are given a Graph that can be constructed by taking a path of length and then possibly adding to each vertex one or several attached graphs that are isolated from each other except for being connected to the path. Of course, trees have this property.
We will now assign to each such Graph a natural quantity: for any vertex , we can consider a random walk started in that jumps uniformly at random to an adjacent vertex until it hits the path. We can then, for each such vertex , compute the expected hitting time (that is: the expected number of steps in the random walk until one hits the path) and define
We argue that this quantity, in a certain sense, captures the essence of the underlying dynamics. We refer to [8] for a paper where the same quantity has been used in a similar way. To get a feeling for the scaling of things, we observe that if is itself a path Graph, then
(2) 
This classical scaling result follows easily from observing that the problem is structurally similar to a random walk on the lattice and that the standard random walk on after
random steps has variance
(and thus standard deviation
).Theorem 2.
Let be a graph whose longest path of length has the following property: the graphs attached to the vertices on the path are isolated (any path from any vertex in to the complement goes through ) and each graph attached to vertex satisfies

the attached Graph does not have too many vertices

and the hitting time is not too large
Then the second eigenvector of the Graph Laplacian assumes its extrema at the endpoints of the graph.
The structure of the proof in §3.3. exploits the heuristic developed in §2.2. By considering a path of length and then attaching another path of length to the th vertex on the long path, we see that both assumptions are optimal up to the values of the constants. We point out that it would be of interest to obtain inverse results: explicit conditions under which one of the extrema is not attained at the endpoints of the longest path. We give one sample application of Theorem 2 to Evans’ counterexample and consider what he called the Fiedler rose (see Fig. 6): let denote the Fiedler rose with vertices. If we start in an outermost vertex, then one step of the random walk leads to the center and the next step leads to the path with likelihood . This means that the expected number of steps required until one hits the path is
This means that if we have a path graph of length and attach a Fiedler rose with vertices to the middle point of the path graph, then the rose can have up to vertices without violating the result. Much more precise asymptotics for this special case were given by Lefèvre [20].
2.4. A Hitting Time Bound.
The purpose of this section is to establish bounds on hitting times under an assumption on the maximum degree. Let be a connected graph and assume that vertex is marked (in the setting above, is the vertex that lies on the long path). Evans’ counterexample shows that we necessarily need to make some assumptions on the maximum degree of and we introduce
Proposition.
Let be a connected graph with maximum degree and marked vertex . The maximum expected time of a random walk started in a vertex in until it hits can be bounded by
where is a constant that depends only on .
Revisiting Theorem 2, if we only have an assumption on the maximal degree, then we are allowed to attach graphs of diameter up to at most
. In light of Evans’ counterexample, this estimate is perhaps not surprising (one can attach Fiedler roses on top of Fiedler roses on top of Fiedler roses etc. to the desired effect). However, we also point out that if the graphs do not have a ‘labyrinth’ type structure where random walkers can easily get lost (in the sense of hitting time being large), then one could attach graphs of larger diameter without violating the conditions of Theorem 2.
2.5. Caterpillar graphs.
We conclude with a simple example: a caterpillar graph [3, 18] is path of length where to each vertex we may add trees of size 1 (alternatively: after removing all vertices of degree 1, a path graph remains). Gernandt & Pade [16] proved that the extrema of the second eigenvector are assumed at the endpoints of the longest graph and established various generalizations of this result. We give another one.
Corollary 2.
Let be a path graph of length with vertices order . Suppose we attach to the vertex an arbitrary number of paths of length at most , where
Then the global extrema are assumed at the endpoints of the longest path.
This Corollary follows almost immediately from Theorem 2 and the behavior of hitting times for path graphs. It is natural to conjecture that stronger results should be true, maybe even
2.6. A Hitting Time Problem.
An interesting question is the following: suppose is a connected Graph with a marked vertex and is a function such that the expected number of steps a random walk started in takes until it hits . What bounds (both from above and from below) can be proven on
A trivial bound is
Amusingly, this might be close to optimal. Fix a degree and consider the following type of Graph where each vertex has the maximal number of children () up to a certain level. Let us then connect all the vertices in the last level to the root of the tree. The induced random walk can be regarded as a biased random walk in terms of the level and will quickly lead to the root of the tree.
A simple question is the following: what sort of hitting time bounds are possible and how do they depend on the graph. For example, if is a tree, then we have . What other results are possible?
3. Proofs
3.1. Proof of Theorem 1
Proof.
Recall that or
Evaluating this equation in a single vertex means that
This can be interpreted as one step of the game: you add to your account and then jump to a random neighbor (‘random’ because we properly normalized the sum) and evaluate the function there. We now iteratively apply this identity to every term involving except for those involving which we keep. The arising terms can be bijectively mapped to the random walks in the game. ∎
3.2. Some Preliminary Considerations
Before embarking on the proof of Theorem 2, we recall several helpful statements. A result of McKay [21] states that
We also know, from the variational characterization
that the eigenvalue decreases if the Graph is enlarged. Since the Graph contains a path of length , we can use the second eigenvalue of the path graph as an upper bound. It is known (see e.g. [9]) that
and therefore
(3) 
McKay’s bound shows that this upper bound using only the diameter is optimal up to constants. These facts are well known (see e.g. [21, 22]). We also observe that changes sign. This means that for every vertex , we can estimate the size of by summing over a path from to the nearest vertex where is negative. For a normalized eigenvector , this shows that
where the second line uses the Cauchy–Schwarz inequality, the fourth line uses equation (1) and the fifth line uses equation (3). Since this holds for every vertex, we have
(4) 
The normalization in of implies by the Hölder inequality that
and therefore, by equation (4):
Moreover, has mean value 0 and therefore the positive part and the negative part cancel out and therefore
(5) 
3.3. Proof of Theorem 2
Proof.
The proof decuples into two parts: first, we show that the sign change of the second eigenfunction occurs somewhere on the long path and not within any attached Graph (see Fig. 5). The second part of the proof makes use of the Game Interpretation. We start by assuming that is the normalized eigenvector associated to the smallest nontrivial eigenvalue .
We now assume that the statement is false and that the sign change occurs somewhere inside the Graph . In particular, appealing to Fiedler’s theorem, the eigenvector has the same sign everywhere on the long path which we can assume without loss of generality to be negative.
Part 1: The argument is rather simple and exploits the bounds derived in §3.2. Let us assume that changes sign in . Then, by Fiedler’s theorem, all the positive values are attained inside . Then, however, by Equation (5), we have
and thus
which is a contradiction to the assumption (1) in the Theorem.
Part 2: It remains to show that the maxima occur at the endpoints of the path under the assumption that there is a sign change on the long path. Let us
assume that this is not the case and that, without loss of generality, the maximum is assumed in .
Since there is exactly one sign change along the long path, one of the endpoints of the long path also
has a positive value; since we have not specified anything about the value of , we can assume without loss of generality that
is positive (see Fig. 8). By Fiedler’s theorem, we have that is nonnegative and so are
and the values in all the attached graphs.
We now play the game twice: first to obtain an upper bound on the maximum of (under the assumption that this maximum is assumed in ) and then to obtain a lower bound on . The game in Section 2.1 implies that
(7)  
(8) 
where is the vertex on the path where is connected to the path.
We now start the game of Section 2.1 in the vertex 1 and obtain
It remains to understand the game. We jump around randomly and add
Suppose we are on the path and . This means that there are graphs attached and since there is a chance of going into these graphs which we will show increasees the expected payoff. We will obtain a lower bound on the expected number of times we encounter the vertex before going towards the next one on the path. We have encountered this type of computation before when computing hitting times for the Fielder rose. Introducing
we can bound
This means that in expectation, a visit to a vertex along the path would contribute at least before we continue to the next vertex on the path. For comparison, if , so that the vertex is not connected to any additional graph, the vertex contributes to the game and in the next step we move to one of the two vertices attached. Corollary 1 yields
Abbreviating the maximum hitting time on a path graph by , we have
and thus the lower bound
(11) 
Putting together Equations (10) and (11) and the assumption of part 2 of this proof, we have
(12) 
and since we observed that ,
(13) 
In other words, since the denominator is positive,
(14) 
and therefore
(15) 
Therefore,
(17) 
Thus,
(18) 
Using the hitting time bound (6) and the second assumption of Theorem 2,
(19) 
wWhich is a contradiction.
∎
3.4. Proof of the Proposition.
Proof.
We give a simple estimate that does not yield the sharp constant (for which we refer to Aldous & Fill [2]). Wherever we are, there is always at least one adjacent vertices that decreases the distance to the marked vertex (because the Graph is connected). If we start anywhere, then the likelihood of going number of times into a direction that decreases the distances to the marked vertex is at least
This is not very likely, the expected number of runs until this happens is
This, however, establishes the desired result. ∎
References
 [1] N. M. M. de Abreu, Old and new results on algebraic connectivity of graphs. Linear Algebra Appl. 423 (2007), no. 1, 53–73.

[2]
D. Aldous and J. Fill, Reversible Markov Chains and Random Walks on Graphs, preprint online.
 [3] S. ElBasil. Applications of caterpillar trees in chemistry and physics. J. Math. Chem., 1(2):153–174, Jul 1987.

[4]
A. Blum, J. Hopcroft and R. Kannan, Foundations of Data Science, Cambridge University Press, 2020.
 [5] T. Biyikoglu, J. Leydold and P. Stadler, Laplacian eigenvectors of graphs. PerronFrobenius and FaberKrahn type theorems. Lecture Notes in Mathematics, 1915. Springer, Berlin, 2007.
 [6] K. Burdzy, The hot spots problem in planar domains with one hole, Duke Math. J. 129 (2005), 481–502.
 [7] K. Burdzy and W. Werner, A counterexample to the ”hot spots” conjecture, Ann. Math. 149 (1999), p. 309–317
 [8] X. Cheng, M. Rachh and S. Steinerberger, On the Diffusion Geometry of Graph Laplacians and Applications, arxiv, Appl. Comp. Harm. Anal., 46, p. 674–688 (2019).
 [9] F. R. Chung, Spectral Graph Theory. CBMS Regional Conference Series in Mathematics, 92. Published for the Conference Board of the Mathematical Sciences, Washington, DC; by the American Mathematical Society, Providence, RI, 1997.

[10]
Moo Chung, Seongho Seo, Nagesh Adluru, and Houri Vorperian. Hot spots conjecture and its application to modeling tubular structures. In Kenji Suzuki, Fei Wang, Dinggang Shen, and Pingkun Yan, editors, Machine Learning in Medical Imaging, volume 7009 of Lecture Notes in Computer Science, pages 225–232.
 [11] E. Brian Davies, G. Gladwell, J. Leydold and P. Stadler, Discrete nodal domain theorems. Linear Algebra Appl. 336 (2001), 51–60.
 [12] L. C. Evans. The Fiedler Rose: On the extreme points of the Fiedler vector. arXiv:1112.6323, 2013.
 [13] M. Fiedler. Algebraic connectivity of graphs. Czechoslovak Math. J., 23(98):298–305, 1973
 [14] M. Fiedler. A property of eigenvectors of nonnegative symmetric matrices and its application to graph theory. Czechoslovak Math. J., 25(100)(4):619–633, 1975.
 [15] M. Fiedler. Laplacian of graphs and algebraic connectivity. In Combinatorics and graph theory (Warsaw, 1987), volume 25 of Banach Center Publ., pages 57–70. PWN, Warsaw, 1989.
 [16] H. Gernandt and J. Pade, Schur reduction of trees and extremal entries of the Fiedler vector, Linear Algebra and its Applications Volume 570, p. 93–122 (2019)
 [17] D. Hammond, P. Vandergheynst, R. Gribonval, Wavelets on graphs via spectral graph theory, Applied and Computational Harmonic Analysis 30, p. 129–150, (2011).
 [18] F. Harary and A. Schwenk, The number of caterpillars, Discrete Mathematics, 6 (4): 359–365 (1973).
 [19] J. Irion and N. Saito, Hierarchical graph Laplacian eigen transforms. JSIAM Lett. 6, 21–24, (2014).
 [20] J. Lefèvre, Fiedler Vectors and Elongation of Graphs: A Threshold Phenomenon on a Particular Class of Trees, arXiv:1302.1266
 [21] B. Mohar, Eigenvalues, diameter, and mean distance in graphs. Graphs Combin. 7 (1991), no. 1, 53–64.
 [22] B. Mohar, Laplace eigenvalues of graphs–a survey. Algebraic graph theory (Leibnitz, 1989). Discrete Math. 109 (1992), no. 13, 171–183.
 [23] A. Ortega, P. Frossard, J. Kovacevic, J. Moura, P. Vandergheynst, Graph Signal Processing: Overview, Challenges, and Applications, Proceedings of the IEEE 106, p. 808 – 828 (2018)
 [24] N. Perraudin and P. Vandergheynst, Stationary signal processing on graphs, IEEE Transactions on Signal Processing 65, p. 3462–3477, (2017).
 [25] N. Saito, How can we naturally order and organize graph Laplacian eigenvectors?, arXiv:1801.06782, 2018 IEEE Statistical Signal Processing Workshop, SSP 2018, Freiburg im Breisgau, Germany, June 1013, 2018.

[26]
D. Shuman, S. Narang, P. Frossard, A. Ortega and P. Vandergheynst, The emerging field of signal processing on graphs: Extending highdimensional data analysis to networks and other irregular domains, IEEE Signal Processing Magazine 30, p. 83–98, (2013).

[27]
D. Shuman, B. Ricaud and P. Vandergheynst, A windowed graph Fourier transform, Statistical Signal Processing Workshop (SSP), p. 133–136, (2012).
 [28] D. Shuman, M. Faraji and P. Vandergheynst, Multiscale Pyramid transform for graph signals, IEEE Transactions on Signal Processing 64, p. 2119–2134, (2016).
 [29] S. Steinerberger, Hot Spots in Convex Domains are in the Tips (up to an Inradius), arXiv:1907.13044
 [30] E. Stone and A. Griffing, On the Fiedler vectors of graphs that arise from trees by Schur complementation of the Laplacian. Linear Algebra Appl. 431 (2009), no. 10, 1869–1880.

[31]
M. Taylor, Partial differential equations. II. Qualitative studies of linear equations. Applied Mathematical Sciences, 116. SpringerVerlag, New York, 1996.