Scholarly article on topic 'Random walks on simplicial complexes and harmonics'

Random walks on simplicial complexes and harmonics Academic research paper on "Computer and information sciences"

0
0
Share paper
Academic journal
Random Structures & Algorithms
OECD Field of science
Keywords
{""}

Academic research paper on topic "Random walks on simplicial complexes and harmonics"

Random Walks on Simplicia! Complexes and Harmonics*

Sayan Mukherjee,1 John Steenbergen2

1 Departments of Statistical Science, Mathematics, and Computer Science, Duke

University; e-mail: sayan@stat.duke.edu

2 Department of Mathematics, University of Illinois at Chicago;

e-mail: jbergen@uic.edu

Received 29 October 2013; revised 25 February 2015; accepted 11 December 2015 Published online in Wiley Online Library (wileyonlinelibrary.com). DOI 10.1002/rsa.20645

ABSTRACT: In this paper, we introduce a class of random walks with absorbing states on simplicial complexes. Given a simplicial complex of dimension d, a random walk with an absorbing state is defined which relates to the spectrum of the ¿-dimensional Laplacian for 1 < k < d. We study an example of random walks on simplicial complexes in the context of a semi-supervised learning problem. Specifically, we consider a label propagation algorithm on oriented edges, which applies to a generalization of the partially labelled classification problem on graphs. © 2016 Wiley Periodicals, Inc. Random Struct. Alg., 00, 000-000, 2016

Keywords: spectral theory; random walks; simplicial complexes

1. INTRODUCTION

The relation between spectral graph theory and random walks on graphs is well studied and has both theoretical and practical implications [2,13,14,16]. Consider an unweighted and undirected graph G = (V, E). A random walk on the graph G is defined via a Markov chain on the set of vertices, with transition probability matrix P = I — D—1W where W is the graph adjacency matrix, and D is a diagonal matrix with Dii the degree of vertex i. The graph Laplacian is defined as A = D — W so P = D—1A. Connections between spectral properties of the graph Laplacian and properties of random walks are well understood [13], the mixing time of a random walk on a graph is one connection. A key observation is that the stationary distribution of a random walk on a graph is related to the harmonics of the

Correspondence to: Sayan Mukherjee

»Supported by NIH (System Biology) (to S.M.) (5P50-GM081883); AFOSR (to S.M.) (FA9550-10-1-0436); NSF (to S.M.) (CCF-1049290); NSF (to J.S.) (DMS-1045153; DMS-12-09155). © 2016 Wiley Periodicals, Inc.

graph Laplacian [13]. Expander graphs are another important example that highlights the relationship between random walks and spectral graph theory [2,7,13].

Graphs are common objects on which stochastic processes are defined, however they are limited in their ability to represent interactions between more than two objects. Simplicial complexes provide a language to describe such high-order interactions. In this paper we will define stochastic processes on simplicial complexes and examine what properties of random walks on graphs extend to random walks on simplicial complexes. The idea of extending results of random walks on graphs to random walks on simplicial complexes have only recently been explored [6,17,19,20]. There has also been recent activity studying the related problem of extending the notion of expander graphs to simplicial complexes and finding isoperimetric inequalities for simplicial complexes [8,15,17,18,21]. In contrast, the understanding of quasirandom properties of hypergraphs is relatively well developed [4,5,10,12].

The main objective of this paper is to define random walks on simplicial complexes by a Markov chain and relate the stationary distribution of the chain to harmonics of the Hodge Laplacian. The extension of the graph Laplacian to simplcial complexes goes back to the work of Eckmann [9] where higher-order combinatorial or Hodge Laplacians were defined. The spectrum of the k-th order Hodge Laplacian provides information on whether the k-th (co)homology class is trivial. The size of the spectral gap can be considered a "measure" of how far the complex is from having nontrivial (co)homology. In this paper we will show that the stationary distribution of the random walk on the simplicial complex captures nontrivial k-th (co)homology classes.

The main differences between random walks on graphs and random walks on simplicial complexes are as follows:

1. Detection of higher-order homology groups: The topological properties studied on graphs are usually limited to connected components and cycles-these are the zeroth and first homology groups, H0 and H respectively. In the case of simplicial complexes we can examine higher dimensional analogs of cycles, represented by the higher homology groups H2, H3 respectively. Understanding the relationship between random walks on simplicial complexes and their homology is one of our aims.

2. The role of orientation: In order to study simplicial homology each simplex in a complex is required to have an orientation, an ordering on its vertices that can be either positive or negative. Conceptually, the biggest difference form the graph case is that for random walks on simplicial complexes a Markov chain needs to be defined on each orientation and the two chains need to be coupled.

3. Defining neighbors: In a graph there is only one logical way to define neighboring vertices-if they are connected by an edge. Two simplexes in a simplicial complex can be connected in various ways, see Fig. 1. In this paper we focus on the case where two k-simplexes are considered to be neighbors if they share a common (k — 1)-dimensional face. In [17] two k-simplexes are considered neighbors if they are faces of a common (k + 1)-dimensional simplex (a coface). These different definitions of neighborhoods yield random walks with different properties.

1.1. Motivation and Related Work

A strong motivation for studying random walks on simplicial complexes is a desire to capture geometry as well as topology. Consider two objects: (a) a solid three dimensional disc and

Fig. 1. (A) The image on the left corresponds to the random walk between two edges that share a triangle. (B) The image on the right corresponds to the random walk between two triangles that share an edge.

(b) the same disc with a small hole punctured in the middle. Generate a graph Gs from a random sample of the solid disc and a graph Gp from a random sample of the punctured disc. If the puncture is small the spectra of the graph Laplacian of Gs and the graph Laplacian Gp will be very similar. In contrast if one generates a simplicial complex or triangulation of the solid and punctured disc, Ts and Tp respectively, the spectra of higher-order Hodge Laplacians will not be the same since the inner boundary of the disc with a hole will be detected by the Hodge Laplacian.

The two operators needed to define the Hodge Laplacian are the boundary map and the coboundary map. Here we provide an intuition for these maps and describe how they relate to the different walks shown in Fig. 1. A rigorous formulation will be provided in Section 2.3. The boundary operator dk maps k-dimensional complexes to (k — 1) dimensional simplices, for example 31 maps edges to vertices and d2 maps triangles to edges. The coboundary operator 8k goes in the opposite direction, it defines a map from lower-dimensional complexes to higher-dimensional complexes. The k-th Hodge Laplacian for dimension k > 0 is composed of two components

Ak = dk+18k + 5k—1 dk.

These two components will correspond to the two definitions of random walks displayed in Figure 1. One can define a random walk between triangles that share an edge, Fig. 1b, this corresponds to 81 d2 since each move in the walk can be thought of as a composition of moving from a triangle to an edge and back to a triangle. One can also define a random walk between edges that share a triangle, Fig. 1a, this corresponds to d281 since each move in the walk can be thought of as a composition of moving from an edge to a triangle back to an edge. One can also define a walk that combines both types of walks. Walks corresponding to 3k+18k were introduced by Parzanchevski and Rosenthal [17]. In this paper we will introduce walks corresponding to 8k—1 dk as well as walks that relate to the Hodge Laplacian.

The following example motivates the walk in Fig. 1B. Consider the 2-dimensional simplicial complex formed by a hollow tetrahedron (or any triangulation of the 2-sphere). We know that the complex has nontrivial 2-dimensional homology since there is a void. However, this homology cannot be detected by the random walk defined in [17], because there are no tetrahedrons that can be used by the walk to move between the triangles. In general,

the walk defined in [17] can detect homology from dimension 0 to co-dimension 1, but never co-dimension 0. Hence, a new walk which can travel from triangles to triangles through edges is needed to detect the void.

In addition to the two types of random walks explored in this paper and in [17], a recent paper by Rosenthal [19] considers a notion of branching processes on simplicial complexes. We will return to a discussion of the similarities and differences of these Markov processes in Section 7.

1.2. Overview of paper

In Section 2 we state the notation used in this paper and define chains, cochains, and the Hodge Laplacian. In Section 3 we define a random walk from a k-simplex to another k-simplex through a (k — 1)-dimensional face. We then relate the stationary distribution of the walk to the harmonics of the Sk—1dk component of the Hodge Laplacian and provide mixing times. In Section 4 we define a random walk from a k-simplex to another k-simplex through a (k + 1)-dimensional coface and relate the stationary distribution to the harmonics of the dk+iSk component of the Hodge Laplacian. The results in this section can be considered a restatement of the results in [17]. In Section 5 we define a walk that captures the harmonics of the entire Hodge Laplacian. In Section 6 we provide some examples of random walks to illustrate some of our ideas and present a novel machine learning algorithm based on random walks on simplicial complexes.

2. DEFINITIONS

In this section we define the notion of a simplicial complex X, the chain and cochain complexes associated to it, and the k-Laplacian.

2.1. Simplicial Complexes

By a simplicial complex we mean an abstract finite simplicial complex. Simplicial complexes generalize the notion of a graph to higher dimensions. Given a set of vertices V, any nonempty subset a c V of the form a = {v0, v1,..., vj} is called a j-dimensional simplex, or j-simplex. A simplicial complex X is a finite collection of simplexes of various dimensions such that X is closed under inclusion, i.e., t c a and a e X implies t e X. While we will not need it for this paper, one can include the empty set in X as well (thought of as a (-i)-simplex). Given a simplicial complex X, denote the set of j-simplexes of X as Xj. We say that X is d-dimensional or that X is a d-complex if Xd = 0 but Xd+1 = 0. Graphs are i-dimensional simplicial complexes. We will assume throughout that X is a d-complex for some fixed d > 1.

If a e Xj, t e Xj-i, and t c a, then we call t a face of a and a a coface of t . Every j-simplex a for j > 1 has exactly j + 1 faces but may have any number of cofaces. Given a e Xj we define deg(a) (called the degree of a) to be the number of cofaces of a. Two simplexes are upper adjacent if they share a coface and lower adjacent if they share a face. The number of simplexes upper adjacent to a j-simplex a is (j + 1) ■ deg(a) while the number of simplexes lower adjacent to a is Y,tca(deg(T) - 1) where the sum is over all faces t of a.

Orientation plays a major role in the geometry of a simplicial complex. For j > 0, an orientation of a j-simplex a is an equivalence class of orderings of its vertices, where two orderings are equivalent if they differ by an even permutation. Notationally, an orientation is denoted by placing one of its orderings in square brackets, as in [v0,..., Vj]. Every j-simplex a has two orientations which we think of as negatives of each other. We abbreviate these two orientations as a+ and a— = —a+ (which orientation a+ corresponds to is chosen arbitrarily). For any j > 1, we will use X+ = {a+ : a e Xj} to denote a choice of positive orientation a+ for each j-simplex a. The set of all oriented j-simplexes will be denoted by X±, so that X± = {a± : a+ e X+} and |X± | = 2|Xj | for any choice of orientation X+.

An oriented simplex a+ = [v0,..., Vj] induces an orientation on the faces of a as (—1)'[V0,... , Vi—1, Vi+1,..., Vj]. Conversely, an oriented face (—1)' [V0,..., v,—1, v,+1,..., Vj] of a induces an orientation a+ = [v0, ..., Vj] on a. Two oriented j-simplexes a+ and a'+ are said to be similarly oriented, and we write a+ ~ a+, if a and a' are distinct, lower adjacent j-simplexes and a+ and a'+ induce the opposite orientation on the common face (if a and a' are upper adjacent as well, this is the same as saying that a+ and a'+ induce the same orientation on the common coface). If they induce the same orientation on the common face, then we say they are dissimilarly oriented and write a— ~ a+. We say that a d-complex X is orientable if there is a choice of orientation X+ such that for every pair of lower adjacent simplexes a, a' e Xd, the oriented simplexes a+, a'+ e X+ are similarly oriented.

Orientation is typically not a property of vertices of a graph and the term oriented graph is taken to mean a directed graph. For j = 0 there are no distinct orderings, and one can think of each vertex v as being positively oriented by default (so, v+ = v) and having an oppositely-oriented counterpart v— := —v. However, introducing orientation to a graph is not standard. The collection of oriented 0-simplexes, X±± does not reduce to the standard graph setting, since graphs are not considered to be oriented. We will discuss orientation in more detail in Section 3. We will see in Section 3 that the random walks we define on simplicial complexes do not reduce to a simple random walk on a graph.

2.2. Chain and Cochain Complexes

Given a simplicial complex X, we can define the chain and cochain complexes of X over R. The space of j-chains Cj := Cj(X; R) is the vector space of linear combinations of oriented j-simplexes with coefficients in R, with the stipulation that the two orientations of a simplex are negatives of each other in Cj (as implied by our notation). Thus, any choice of orientation X+ provides a basis for Cj. The space of j-cochains Cj := Cj(X; R) is then defined to be the vector space dual to Cj. These spaces are isomorphic and we will make no distinction between them. Usually, we will work with cochains using the basis elements {1a+ : a+ e X+}, where 1a+ : Cj ^ R is defined on a basis element t+

1o+(r+) =

1 t+= a+ 0 else

The boundary map dj : Cj ^ Cj—1 is the linear map defined on a basis element [V0,..., Vj] as

dj[V0,..., Vj] = Y^(—1)i[V0,..., Vi—1, Vi+1,..., Vj].

The coboundary map 8 1 : Cj 1 ^ Cj is then defined to be the transpose of the boundary map. In particular, for f e Cj—1,

(8—1f )([vc,..., Vj]) = £( - 1)f ([vo, • • •, Vi_i, v/+i,..., Vj]). ¡=1

When there is no confusion, we will denote the boundary and coboundary maps by d and 8. It holds that dd = 88 = 0, so that (Cj, dj) and (Cj, 8j) form chain and cochain complexes respectively.

The homology and cohomology vector spaces of X over R are

„ ker d: „ ker &

Hj := Hj(X; R) =-and Hj := Hj(X; R) =-.

im dj+1 im &—1

It is known from the universal coefficient theorem that Hj is the vector space dual to Hj. Reduced (co)homology can also be used, and it is equivalent to including the nullset as a (—1)-dimensional simplex in X.

2.3. The Hodge Laplacian

The k-Laplacian of X is defined to be

A, := AUp + Ad

AUp = 9*+,8k and Ad°™ = 8k-1dk.

1k = ^k + "k

V = °k+1c

The Laplacian is a symmetric positive semi-definite matrix, as is each part Aup and Aj|o' From Hodge theory, it is known that

ker Ak = Hk = Hk

and the space of cochains decomposes as

Ck = im 9k+1 ® ker Ak ® im 8k—1

where the orthogonal direct sum © is with respect to the "usual" inner product

a+eX+l

For much of this paper we will focus on the A;do™ half of the Laplacian. Trivially, im dj+i c ker Ado™ and the smallest nontrivial eigenvalue of Aj|own is therefore given by

T • ii II2 Xk = min -—,

r-ck ifii2

f lim d

where If ||2 := y/ff) denotes the Euclidean norm on Ck. A cochain f that achieves the minimum is an eigenvector of Xk. It is easy to see that any such f is also an eigenvector of Ak with eigenvalue Xk and that, therefore, Xk relates to homology:

Xk = 0 ^ ker Ak = 0 ^ Hk = 0. Random Structures and Algorithms DOI 10.1002/rsa

Remark 2.1. Given a choice of orientation X+, A)Jown can be written as a matrix with rows and columns indexed by Xk+, the entries of which are given by

/ A down\ (Ak )a+ ,a+

k + 1 o'+ = a+ 1 o+

— 1 a+ ~ a+

0 otherwise.

Changing the choice of orientation X+ amounts to a change of basis for Aj|own. If the row and column indexed by a+ are instead indexed by a—, all the entries in them switch sign except the diagonal entry. Alternatively, Aj|own can be characterized by how it acts on cochains

A,downf (r+) = (k + 1) ■ f (t+) + J2 f (a+) — J2 f (a+).

a— ~r+ a+ ~r+

Note that since Adownf is a cochain, Ak°wnf (t—) = — Adownf (t+).

The remainder of this section states some relevant properties on the behavior of Aj|own.

Definition 2.2. A d-complex X is called k-connected (1 < k < d) if for every two k-simplexes a, a' there exists a chain a = a0, ai,..., an = a' of k-simplexes such that ai is lower adjacent to ai+1 for all i. For a general d-complex X, such chains define equivalence classes of k-simplexes, and the subcomplexes induced by these are called the k-connected components of X.

Definition 2.3. A d-complex X is called disorientable if there is a choice of orientation X+d of its d-simplexes such that all lower adjacent d-simplexes are dissimilarly oriented. In this case, the d-cochain f = Ya+sX+ 1°+ is called a disorientation.

Remark 2.4. Disorientability was defined in [17] and shown to be a higher-dimensional analog of bipartiteness for graphs. Note that one can also define X to be k-disorientable if the k-skeleton of X (the k-complex given by the union (Ji<k X') is disorientable, but this can only happen when k = d. For example if k < d then there exists a (k + 1)-simplex a+ = [v0,..., vk+1]. Given any two dissimilarly oriented faces of a+, say, [v1, v2,..., vk+1] and [v0, v2,..., vk], we find that the simplex {v0, v1, v3,..., vk} cannot be dissimilarly oriented to both of them simultaneously.

The following Lemma is similar to Proposition 2.7 in [17].

Lemma 2.5. LetX be a d-complex, 1 < k < d and M = maxffeXk—1 deg(a).

1. Spec(Ad0Hn) is the disjoint union of Spec(Ad0H,n|X;) where Xt are the k-connected components ofX.

2. The spectrum of Adkown is contained in [0, (k + 1)M].

3. The kernel of Adkown is exactly ker dk = im dk+1 © ker Ak.

4. The upper bound (k + 1)M is attained if and only ifk = d and X has a d-connected component that is both disorientable and of constant (d — 1)-degree.

Proo/.Statement (1) follows from the fact that A)Jown can be written as a block diagonal matrix with each block corresponding to a component Xi. Statement (3) is easy to verify.

For statement (2), let/ be an eigenvector of Aj|own with eigenvalue X, let X+ be a choice of orientation such that / (a+) > 0 for all a+ e X+ and suppose / (t+) = maxa+eX+/ (a+). Then by Remark 2.1,

X/ (t+) = A,down/ (t+)

= (k + 1) ■ /(t+) + J2 /(a+) — J2 /(a+)

a— a+

< (k + 1) ■ /(T+) + £ /(T+) + £ /(T+)

< (k + 1) ■ /(T+) + (k + 1)(M — 1) ■ /(a+)

< (k + 1)M ■ /(t+)

where the third inequality results from the fact that any k-simplex is lower adjacent to at most (k + 1)(M — 1) other k-simplexes. Therefore, X < (k + 1)M.

It now remains to prove statement (4). Looking back at the inequalities, it holds that X = (k + 1)M only if a— ~ t+ and/(a+) = /(t+) whenever a and t are lower adjacent, and the faces of a all have degree M. But since/(a+) = /(t+), the same reasoning can be applied to/(a+) for all a lower adjacent to t and eventually to all k-simplexes in the same k-connected component Xi. Ultimately, this implies that Xi has constant (k — 1)-degree and is k-disorientable (and hence k = d).

To see that this bound is indeed attainable, consider a disorientable d-complex with constant (d — 1)-degree M (this includes, for instance, the simplicial complex induced by a single d-simplex). Let X+ be a choice of orientation such that all lower adjacent d-simplexes are dissimilarly oriented. Then a disorientation/ on Xd will satisfy

A,down/(t+) = (k + 1) ■ /(t+) + J2 /(a+) — J2 /(a+)

a— a+

= (k + 1) ■ /(t+) + J2 /(a+)

= (k + 1) ■ 1 + 1 = (k + 1)M = (k + 1)M ■ / (t+) for every t+. ■

3. RANDOM WALKS AND AD°wn

In this section we will define a random walk on a d-complex. The stationary distribution of the random walk is related to the harmonics of Akdown for 1 < k < d. The mixing time of the walk is a function of the spectral gap of Akdown. A notion of maximal degree M will appear in our results.

The following random walk is a Markov chain that is related to Aj|own.

Definition 3.1. The state space is the set of oriented complexes as well as a death state {©}, 5 = Xk± U{©}.

1. Neighbors: two oriented k-cells a and a' are called neighbors, which we denote as a I a' if they share a (k — 1) face and have opposite orientation on the shared face.

2. Transition matrix P: the transition matrix is the time-homogenous Markov chain on the state space S = Xk± U {©} with transition probabilities

Pa, = Prob(a ^ a') =

a0 when a = © and a' = a

ax when a = © and a I a' a2 when a = © and a I a'

P when a = © and a' = ©

1 a = a' = ©,

where a' denotes the opposite orientation of a', a I a' is the notation for moving to an oriented neighboring cell, and a I a' is the notation for moving to a disoriented neighboring cell. The probability of transitioning into a death state is P = 1 — Y,a/¿Lka) Prob (a ^ a') — a0 where Lk(a) are the set of simplexes that are oriented or disoriented neighbors of a. The transition probabilities are constrained to ensure P is a stochastic matrix.

The intuition behind the steps in the transition matrix follows. We first consider the case where there is no death state. In this case with probability a0 one stays in the same state. The probability of moving to an oriented neighboring cell is ax. The probability of moving in one step to disoriented neighboring cell is a2. The death state is needed when the total probability, 1 — P, of either moving to a new cell or staying at the same cell is less than one. In this case the remaining probability P corresponds to moving to an absorption or death state. There is also an interpretation of the Markov chain as two coupled Markov chains. Again for simplicity we consider the case where there is no death state. Consider one Markov chain on the state space X+ and another chain on X—. The two chains are coupled as follows: at each time step one can either jump to another state in the chain or jump to the corresponding state of opposite orientation in the other chain and proceed in the new chain.

We now specify the parameter values of the transition matrix P.

Definition 3.2. We define transition matrix P based on Definition 3.1 with parameter values

a0 = p, a, = a2 =-,

0 1 2 (m — 1)(k + 1)'

where M is the maximal degree maxaeXk—1 deg(a), 1 < k < d corresponds to the order of Adown, and 0 < p < 1.

We selected this parameter specification for ease of analysis and to impose certain intuitive properties of the walk. The value 0 < p < 1 determines the laziness of the walk, the probability of not changing states. We set the probability of moving to a neighbor as equal to the probability of moving to a disoriented neighbor, a1 = a2. One reason for the existence of the death state is that if ax = a2 the total transition probability at each state may not sum to one so we use the death state as an absorption state. A natural choice for a1 is a uniform probability of transitioning to each of the neighbors of a complex. We set ax = {M—:—Pk+V), this would be a uniform transition probability if all complexes had maximal degree M, this is the analog of a M-regular graph. We suspect that the qualitative phenomena and insights

we derive for the specified walk will transfer to walks defined by other parameters, such as the more natural uniform walk.

In the remainder of this section we will relate the stationary distribution of the stochastic matrix P to the geometry of the simplicial complex on which the random walk is defined. Specifically we are interested in relating the stationary distribution to the harmonics of Af™. From Hodge theory we know that the relevant geometry is characterized by the space of k-cochains, f e Ck .A k-cochain is antisymmetric under exchange of any pairs of indices which implies the following constraint

{f : X± ^ R | f (a) = —f (a), Va e X±}.

Given the transition matrix P we can compute the following expected value from a k-cochain at time t as

Et [f] :=J2 P' (a )f (a) = J2 P (a) — Pt (W))f (a),

a eX± a eXk

where pt(a) is the probability vector over states X+ at time t and pt(a) is the probability vector over states X_ at time t. We will be interested in the convergence of the following "expectation process"

Et = Pt (a) — pt (a).

If we consider the random walk as two coupled chains, on X+ and X- respectively, the above can be thought of as the difference in the marginal probability vectors of each chain.

In the case of a graph with a stochastic matrix PG one studies the convergence of pt the probability vector over states at time t to a stationary distribution n. The rate of convergence of the probability vector pt to n as well as the stationary distribution n are related to harmonics of the graph Laplacian. The graph case has some fundamental differences from the case of simplicial complexes in that graphs are not defined with orientation and the antisymmetric constraint that is required of k-cochains is not relevant.

Note that the orientation of the complexes does not affect the Hodge Laplacian. One can define the following propagation matrix on X+k that corresponds to the random walk P in Definition 3.2.

Definition 3.3. The propagation matrix B of the lower k-walk is defined to be a square matrix indexed by Xk with

(B)a+^a+

a0 = p a+ = a+

_ 1—p / |

a1 = — (M—1)(k+1) a+ + a+

__1—p / I -

a2 = (M—1)(k+1) a+ ^ a +

0 otherwise.

There is a natural relation between the propagation matrix and the Markov matrix. We define a matrix Q with rows indexed by X+ and columns indexed by S. The values of the matrix are specified as follows: Qj = 1 if state i and state j are neighbors, Qj = — 1 if state i and state j with reverse orientation are neighbors, and Qj = 0 if state j is a death state. This matrix has the following property QP = BQ. We now state some properties of the propagation matrix and its relation to the transition probability matrix P.

Proposition 3.4. The propagation matrix B is given by

E_p(M - 2) + 17_ M - 1

(M - 1)(k + 1)

In addition, B satisfies QP = BQ and

QP' v = B'Qv.

Proof. The first claim is a straightforward computation based on Definition 3.3. The second claim is equivalent to the equality QP = BQ, which we now prove. If a' e S and Pa/ is the column of P indexed by a', then the column of QP indexed by a' is QPa/. Using the definition of Q, the following holds

(QP)a+,a' = QPa' (a+)

= Pa>(a+) - Pa>(a-)

(P^a+ ,a ' (P^)a-,a ' '±p

a = a±

(M-1)(k+1) s = ® and a'i a±

0 otherwise.

Similarly, note that (BQ)a+,a/ = B(Q1ar)(a+) where 1a/ is the vector assigning 1 to a' e S and 0 to all other elements in S. If a' = ©, Q1a/ is the zero vector. Otherwise, if a' = t± then Q1a/ = ±1T+ and

(BQ)a+,a' = ±B lr+ (a+)

= ±(B)a+,T+ '±p

(M-1)(k+1) 1-p (M-V)(k+V)

(M-1)(k+1)

T+ i a+ t- i a+ else

a ' = a± a ' i a± otherwise.

For what follows, we define £rt+ := B' 1T+ to be the marginal difference of the random walk on X starting at t+ at time t. Also, let X+ be a choice of orientation and denote M = maxaeXk-i deg(a).

Corollary 3.5.

1. The spectrum ofB is contained in [2p - 1, p(MM^_>^1'\, with the upper bound acheived by cochains in ker dk and the lower bound acheived if and only ifk <= d and there is a disorientable d-connected component of constant (d - 1)-degree.

2. If t has a coface, then

3. Ifp = 0,1 then

fT+1 >

p(M - 2) + 1\ ' 1

M - 1 ) Vk+2

p(M - 2) + 1

U2 < - II' ^ m - 1

Proof. Statement (1) is easy to verify using Lemma 2.5 and Proposition 3.4. Statement (3) follows from the inequality \\Af ||2 < ||A|| ||f1|2 where A is a matrix, f is a vector, and || Ay is the spectral norm of A.

It remains now to prove statement (2). If t has a coface a, let f = dk+x1a+ (with a+ being any orientation of a) which implies f e ker dk. Let f,f1,..., f be an orthogonal basis for Ck such thatf1,..., f are eigenvectors of B. Then,

lE'T+ I2 = lB' 1t+ I2

= UaB'f + ai B'fi + ... + aB'f |2

^ \\aB'f I2

(p(M - 2) + 1V

M- 1 ^^ "2

p(M - 2) + 1

M - 1 p(M - 2) + 1\ ' |f (t+

M - 1 p(M - 2) + 1 \ ' 1

M - 1 ) Vk+2

By Corollary 3.5 we know that Et = 0 when p = 0,1. We rescale B to define the

normalized propagation matrix

M - 1 B :=-B.

p(M - 2) + 1

for which the trivial convergence due to (PM-r1) < 1 is eliminated. We also define the normalized marginal difference £]+ := B' 1t+ . The next two theorems show that the homology of X can be determined from the limiting behavior of the normalized marginal difference.

Theorem 3.6. The limi' := lim^^ Ext + of'he normalized marginal difference exists for all t+ if and only if B has no eigenvalue X < -1. Furthermore, ¿S" = projker d 1t+ whenever exists, where projker ^ is 'he projection map on'o ker dk.

Proof. Note that by Corollary 3.5, the spectrum of B is upper bounded by 1 and the eigenspace of the eigenvalue 1 is exactly ker dk. Letf1,..., f be an orthogonal basis for Ck

such thatf1,...,fi are eigenvectors of B with eigenvalues y1,..., yt. Then any 1T+ can be written as a linear combination 1T+ = af + ... + ai, f so that

£t0+= B' 1t+= aiYi/i + ..., ayf

Since the f form a basis, £00+ converges if and only if ajyj converges for each j. In other words, ££0+ converges if and only if for every j, aj = 0 or yj > -1. Furthermore, the limit (when it exists) is always

J2 j = Projker % 1T+

jYj=i>

Finally, suppose B has an eigenvalue X <—1. Then there is an eigenvector / such that B'f = X'f does not converge. Since the set of cochains {1T+ : t+ e X±} spans Ck(R),/ can be written as a linear combination of the cochains and therefore B' 1T+ must not converge for some t+. ■

Theorem 3.7.

1. If < p < 1 'hen 'he limit £00+ exists for all t+ and

dim(span{projkersk £0 : e X±}) = dim(^k(X))

where projker sk denotes 'he projection map onto ker Sk.

2. The same holds when p = M—4 ork = dor 'here are no disorien'a'able d-connec'ed components of constant (d — 1)-degree.

3. We can say more ifp > 2 .In 'his case,

II fT+ _ fT+ || — o

\\°n °œ || 2 _ O

1----xk

(p(M - 2) + 1)(k + 1) k

Proof. The proof follows mostly from Theorem 3.6. According to Theorem 3.6, £00 exists for all t+ if and only if the spectrum of B is contained in (—1,1]. Using Corollary 3.5 and the definition B := B, we know that the spectrum of B is contained in

(2p — 1) p(M——)+i, 1 . Now each of the following statements imply each other

M — 1 p(M — 2) + 1

(2p - 1)- > -1, —--- > 1 - 2p

F p(M -2) + 1 M - 1 F

M - 2 \ 1 M - 2

P 77-7 + 2 > 1 ^77, P >

M — 1 M 3M — 4

which proves that the spectrum of B is indeed contained in (—1,1] when p > M—4. Since the 1T+ span all of Ck, the £0+ = projker d 1T+ span all of ker dk, and hence the projker Sk £0+ span all of ker Sk.

In the case that p = M—i, the spectrum of B is contained in [—1,1]. However, as long as — 1 is not actually an eigenvalue of B, the result still holds. According to Corollary 3.5, — 1 is an eigenvalue if and only if k = d and there is a disorientable d-connected component of constant (d — 1)-degree. The case p = 1 is trivial (B = I) and is not considered.

Finally, if the spectrum of B lies in (-1,1] and k is the eigenvalue of B contained in (-1,1) with largest absolute value, then

B'f - lim B'f <|k|' |f H2

i^t» 2

for all f. To see this, letf1,..., f be an orthonormal basis for Ck such thatf1,..., f are eigenvectors of B with eigenvalues y\,..., Y. Then any f can be written as a linear combination

f = af + ..., +af so that |

= V Ei I«; I and

B'f - lim B'f = aiyf + ... + «i y'ift - J2 «fi

IjYj=1}

E a Yj'f

jYj =1}

< W -If ||2.

In particular, if p > 2 then the spectrum of B is contained in [0,1] and therefore k = 1 - ,„., „ kk. .

(p(M-2)+1)(k+1)

Note the dependence of the theorem on both the probability p of remaining in a state and on M. We can think of M as the maximum amount of "branching", where M = 2 means there is no branching, as in a pseudomanifold of dimension d = k, and large values of M imply a high amount of branching. In particular, the walk must become more and more lazy for larger values of M in order to prevent the marginal difference from diverging. However, since M-2i < 1 for all M a lazy probability of at least 1 will always ensure convergence. While there is no explicit dependence on k or the dimension d, it is easy to see that M must always be at least d - k + 1 (for instance, it is not possible for a triangle complex to have maximum vertex degree 1). A natural question is how varying the transition matrix P in Definition 3.2 will change the scaling behavior of the walk with respect to M and p.

We would also like to know whether the normalized marginal difference always converges to 0. Note that if t+ has a coface, then we already know that || || 2 stays bounded away from 0 according to Corollary 3.5. However, if t has no coface, then 1T+ may be perpendicular to ker dk, allowing |f(T+ L to die in the limit as we see in the following corollary.

Corollary 3.8. If t has no coface, Hk = 0, and if -,

< p < 1 'hen

\K II2 =

The same is 'rue when p = M-l componen's of constan'(d - \)-degree,

and ork = d and 'here are no disorien'able d-connec'ed

Proof. Under all conditions stated, converges. If t has no coface, then 1T+ is in the orthogonal complement of im d+, because all elements of im d+ are supported on oriented faces of (k + 1)-simplexes. If Hk = 0 then ker dk = im d+ and

\\Eï II 2 = projker % 1T+ = -

4. RANDOM WALKS AND AUP

The random walk described by Parzanchevski and Rosenthal in [17] is the "dual" of the random walk defined in the previous section in that one traverses between simplexes via cofaces rather than faces. We include Rosenthal and Parzanchevski's proposed walk in this section so the paper is self contained and we can compare the walk on Aup with the one in the previous section as well as a walk corresponding to the full Hodge Laplacian. We do not present any new results in this section.

Let X be a d-complex, 0 < k < d — 1, and 0 < p < 1.

Definition 4.1. The State space is the set of oriented complexes as well as a death state, S = X± U {©}.

1. Co-neighbors: two oriented k-cells are called co-neighbors, which we denote as a f a' if they share a (k + 1) coface and have opposite orientation on the shared face.

2. Transition matrix P: the transition matrix is the time-homogenous Markov chain on the state space S = X± U {©} with transition probabilities

Pa,ff, = Prob (a ^ a ') =

ao = p

a1 = k-ijg(a)

a2 = a1

P = 1 - p 1

when a = © and a' = a when a = © and a' f a when a = © and a f a' when deg(a) = 0 and a' = © a = a' = ©,

for all a, a ' e Xk.

The definition of the walk in [17] is slightly different. There is no death state in [17] because the case of k = d — 1 was examined and it was assumed that every k-simplex had at least one coface. As in the previous section we can relate the above Markov matrix to Aup and QP' v is the marginal difference after' steps for the random walk via co-neighbors. Here Q is the same matrix as in the previous section. There is again a propagation matrix A such that QP' v = A'Qv and A relates to Aup. As before the marginal difference converges to 0 for all initial distributions. We scale A by a constant to provide a normalized propagation matrix £ and a normalized marginal distribution A'Qv that describes the limiting behavior. The limiting behavior of the normalized marginal difference reveals homology similar to Theorem 3.7

There are a few differences between the walk across faces and the walk across cofaces. The norm of the normalized marginal difference for the walk across cofaces starting at a single oriented simplex stays bounded away from 0 (see Proposition 2.8 of [17]), whereas this need not hold for the walk across faces (as in Corollary 3.8). This is because in the walk across cofaces every starting point 1T+ has some nonzero inner product with an element of im Sk—1 c ker Sk. The second difference is the threshold values for p in Theorem 3.7 and Theorem 2.9 of [17]. For the walk across faces, homology can be detected for p > M—i (where M = maxaeXk—1 deg(a)) whereas for the walk across cofaces the threshold is p > 3kk"2. Hence, the walk across cofaces is sensitive to the dimension while the walk across faces is sensitive to the maximum degree. In both cases, p > | is always sufficient to detect homology and p > | allows us to put a bound on the rate of convergence.

5. RANDOM WALK RELATED TO THE FULL HODGE LAPLACIAN

The existence of random walks across faces and cofaces that respectively correspond to the two parts of the Hodge Laplacian suggests that there may be a random walk for the entire Laplacian, Ak = Aup + Aj|own . In this section, we state results for a random walk that captures both parts of the Hodge Laplacian. One can define a weighted notion of the Hodge Laplacian to weighted Laplacians

Lk,w := Wr1/23k+Wk+1 SkW-1/2 + W1/25k-1Wkl119kW!/2

where Wj denotes a diagonal matrix with diagonal entries equal to positive weights, one for each j-simplex. In general, one can also define an operator Lk that is a generalized notion of the k-th Hodge Laplacian.

Definition 5.1. Let X+ be a choice of orientation. A generalized Laplacian matrix is a square matrix L such that

1. the rows and columns of L are indexed by X+,

2. L has nonnegative diagonal entries,

3. whenever L has a zero on the diagonal, all other entries in the same row or column are also zero,

i. L is a non-negative operator.

Definition 5.2. Let X+ be a choice of orientation, L a generalized Laplacian matrix, and p e [0,1]. Given L we define a normalization matrix D- as a diagonal matrix with (D-1)a+,a+ = (La+,a+ )-1 if La+,a+ > 0 and (D-1)a+,a+ = 0 otherwise. We define the p-lazy propagation matrix related to L to be

Alp := pK-J2±1, - L-l . LD-K K L

. The case K = 0 is degenerate

where p e 1], K := maxff+6x+E a+=a+ (LD-1)

and is not considered. In addition, we define the normalized p-lazy propagation matrix

1 - P i ( K

alp := I----ld-1[ =-alp

L,p p(K - 1) + 1 L V P(K - 1) + i L,p

Note that whenever K = 1, AL,P = AL,P. In particular, this is true in the graph case when L = A°.

We will now define the Markov transition matrix in terms of the propagation matrix.

Definition 5.3. Let X+ be a choice of orientation, L a generalized Laplacian matrix, p e [0,1], and let AL,P be defined as above. The state space is S := X± U {©} which is a (2n + 1) x (2n + 1) matrix where the number of unoriented complexes is n. The Markov transition matrix is defined as

PL,p =

A A' 0 A' A °

where the first n rows and columns correspond to X++, the second n rows and columns correspond to X-, and the last row and column correspond to the death state. The elements of matrix A correspond to transition probabilities between simplexes of the same orientation and Aj = (ALpP)ij v 0 for i, j = 1,..., n. The elements of matrix A' correspond to transition probabilities between simplexes of the opposite orientation and A'ij = (-AL,p)ij v 0 for i, j = 1,..., n. The elements in the column vector v are the probabilities of entering the death set from one of the complexes and

Vi = 1 - £ (Pup)o\o for all a.

a '^S\{%)

The row vector v' is the transpose of v. The 1 in the lower diagonal states implies that one does not leave the death state.

The following lemma shows that PL,p is a left stochastic matrix.

Lemma 5.4. LetXk+ be a choice of orientation, L a generalized Laplacian, andp e [0,1]. The matrix PLp defined above is a left stochastic matrix for an absorbing Markov chain on the state space S (i.e., (PL)a— Prob(a — a')) such that © is an absorbing state and Prob(a —> a) — p for all a — ©.

Proof. It is clear by the definition of PL,p that © is an absorbing state. To see that Prob (a — a ) — p for all s — ©, note that

) p(K - 1) +1 1 -p 1

(^l,p )a+,a+ k k

p(K - 1) + 1 - 1 + p —-— p

and hence by the definition of PL,p,

(PL,p )a-,a- — (PL,p)a+ ,a+ — p

for all a. It is also clear by the definition of PL,p that the entries (PL,p)a>_,a+ — (PL,p)a+ ,a-are nonnegative for any a, a'. Hence, in order to show that PL,p is left stochastic we need only to prove that Y,a>eS\[©}(PL,p)aV < 1 for all a e S \ {©}. By the symmetries inherent in PL,p, the value of the sum is the same for a — a+ as it is for a — a_. For any a — a+,

V (PL,p)a',a — V (AL,p)a

a'eS\{&} a^eX+

_^ L,p/a ,a ^_^ V*L,p/a+,a+

(AL,p)a+ ,a+

=p + E

a'+eX+\{a+}

1 - p K

a+eX+\{a+}

(LDl-1)

L >a+ ,a+

< p + (1 -p) = 1. .

The following theorem relates the Markov transition matrix PLp to the generalized Hodge Laplacian L.

Theorem 5.5. Let X+ be a choice of orientation, L a generalized Laplacian matrix, p e [0,1], and let AL,p and PL,p be defined as above. In addition, let Q be defined as in Section 3. Then

AL,pQ = QPL,p.

In other words, the evolution of the marginal differences QP'L,pv after n steps with initial distribution v is governed by the propagation matrix: QP'ipV — AL pQv.

Proof. Using the definition of Q

(QPl,p)o+ ,s = (PL,p)a+,s - (PL,p)a- ,s = i±(AL,p)a+,a+ s =

|0 s = ©

Similarly, note that (AL,pQ)a+,s = ALp(Q1s)(a+) where 1s is the vector assigning 1 to s e S and 0 to all other elements in S. If s = ©, Q1s is the zero vector. Otherwise, if s = t± then Q1S = ±1T+ .Thus,

(AL,pQ)a+,s =

|±Al,p1t+ (<Z+ ) s = T± |0 s = ©

\±(AL,p)a+,t+ s = T± [0 s = © .

Finally, we conclude with a few results motivating the normalized propagation matrix and showing how the limiting behavior of the marginal difference relates to the kernel and spectrum of L.

Theorem 5.6. Let Xk+be a choice of orientation, L a generalized Laplacian matrix with

Spec(L) c [0, A] (A > 0). Then for

< p < 1 the following statements hold:

0 for every initial distribution v,

1 llAL,pQ, ,2

2. A'LpQv ^ projkerLQv for every initial distribution v, where projkerL denotes the projection map onto the kernel of L,

3. If k is the spectral gap (smallest nonzero eigenvalue) ofL then

ila'l,pqv - projkerlqv\2 = o

p(K - 1) + 1 J

Proof.The proof is the same as in the proofs of Corollary 3.5 and Theorem 3.7 and mostly boil down to statements about the spectra of AL,p and AL,p. Note that since k^xty < p < 1, Spec(AL,p) c [0,1] where the eigenspace of the eigenvalue 1 is equal to the kernel of L, and the largest eigenvalue of AL,p less than 1 is 1 - p(K1_T!)+! k. ■

As an example of the applicability of this framework, AL,p is used with L = Ak to perform label propagation on edges in the next section.

6. WALKS ON TRIANGLE COMPLEXES AND RANDOM WALKS FOR SEMI-SUPERVISED LEARNING

In this section we provide some examples of random walks on simplicial complexes to provide some intuition. In addition we extend the label propagation algorithm used in machine learning for semi-supervised learning from graphs to simplicial complexes.

6.1. Triangle Complexes

We begin by reviewing local random walks on graphs as defined by Fan Chung in [3]. Given a graph G = (V, E) and a designated "boundary" subset S c V, a ¿-lazy random walk on S = V \ S can be defined to satisfy a Dirichlet boundary condition on S (meaning a walker is killed whenever it reaches S). The walker starts on a vertex v0 e S and at each step remains in place with probability 1 or else jumps to one of the adjacent vertices with equal probability. The boundary condition is enforced by declaring that whenever the walker would jump to a vertex in S, the walk ends. Thus, the left stochastic matrix P for this walk can be written as

(P)v,ves = Prob (v ^ V) =

2 if v = v

d if v ~ V

0 else

where v ~ V denotes that vertices v and V are adjacent and dv is the number of edges connected to v. Note that P is indexed only by S, and that its columns sums may be less than 1. The probability of dying is implicitly encoded in P as the difference between the column sum and 1. As was shown in [3], P is related to a local Laplace operator also indexed by S.If D is the degree matrix and A the adjacency matrix, the graph Laplacian of G is A = D — A. We denote the local Laplacian as AS, where S in the subscript means rows and columns indexed by S have been deleted. The relation between P and AS is

1 , p = i — ^ asd—'.

Hence, the existence and rate of convergence to a stationary distributions can be studied in terms of the spectrum of the local Laplace operator.

Now suppose we are given an orientable 2-dimensional non-branching simplicial complex X = (V, E, T) where T is the set of triangles (subsets of V of size 3). Non-branching means that every edge is contained in at most 2 triangles. We can define a random walk on triangles fundamentally identical to a local walk on a graph which reveals the 2-dimensional homology of X. The ¿-lazy 2-walk on T starts at a triangle t0 and at each step remains in place with probability ¿ or else jumps to the other side of one of the three edges. If no triangle lies on the other side of the edge, the walk ends. The transition matrix B for this walk is given by

(B)t,,t = Prob(t ^ t') =

2 if t = f 6 if t ^ t' 0 else

where t ~ t' denotes t and t' share an edge. This is the same transition matrix as P, in the case that dv — 3 for all v e S. In this case, the analog of the set S is the set of triangles that

Fig. 2. Making the Dirichlet boundary condition explicit, and translating into a graph.

do not share an edge, this is the boundary of X. To draw an explicit connection, imagine adding a triangle to each boundary edge, obtaining a larger complex X = (V, E, T). See Fig. 2 t

Then take the "dual graph" G = (V, E) of X by thinking of triangles as vertices (so, V = T) and connecting vertices in G with an edge if the corresponding triangles in X share an edge. We do not add a vertex for each outer face. Choose the vertices corresponding to the added triangles T \ T to be the boundary set S. Now the matrix P associated to the local random walk on G is indistinguishable from the matrix B associated to the random walk on X. In addition, it can be seen that AS on G is the same as A2, the 2-dimensional Laplacian on X defined with respect to a given orientation (recall that we have assumed orientability). The following states the relation between the transition matrices and Laplacians:

1 1 B = P = I -- As = I -- A2.

See Section 2 for the definition of A2, and the appendix of [21] for more on the connection between AS and A2.

It is a basic fact that the kernel of A2 corresponds to the 2-dimensional homology group of X over R. Therefore, there exists a stationary distribution for the random walk if and only if X has nontrivial homology in dimension 2. Additionally, the rate of convergence to the stationary distribution (if it exists) is governed by the spectral gap of A2. In particular, the following statements hold:

1. Given a starting triangle to, the marginal distribution of the random walk after n steps is E0 := Bn1t0 where 1t0 is the vector assigning a 1 to t0 and 0 to all other triangles. For any t0, the marginal distrubition converges, i.e., Ex :— limn^cx) En exists.

2. The limit E^ is equal to 0 for all starting triangles t0 if and only if X has trivial homology in dimension 2 over R.

3. The rate of convergence is given by

? - E'0II2 = o

where k2 is the smallest nonzero eigenvalue of A2.

The example given here is constrained by certain assumptions (orientability and the non-branching property), which allows for the most direct interpretation with respect to previous work done on graphs.

6.2. Label Propagation on Edges

In machine learning random walks on graphs have been used for semi-supervised learning. In this section we will generalize a class of algorithms on graphs called "label propagation" algorithms to simplicial complexes, specifically we extend the algorithm described in [23] (for more examples, see [1,11,22]). The goal of semi-supervised classification learning is to classify a set of unlabelled objects {vi,..., vu}, given a small set of labelled objects {vu+i,..., vu+i} and a set E of pairs of objects {vt, vj} that one believes a priori to share the same class. Let G — (V, E) be the graph with vertex set V — {v1,..., vu+t} and let P be the probability matrix for the usual random walk, i.e.,

(P)ij — Prob(vj ^ vi) — -

where dj is the degree of vertex j. We denote the classes an object belongs to as c — 1,..., C and an initial distribution /0c : V ^[0,1] is the a priori confidence that each vertex is in class c, a recursive label propagation process proceeds as follows.

1. For t — 1,...,t and c — 1,..,C:

(a) Set/c ^ P/-1

(b) Resetfc(vi) — 1 for all vi labelled as c.

2. Consider/H as an estimate of the relative confidence that each object is in class c.

3. For each unlabelled point v*, i < u, assign the label

arg max/ (vi)}.

c—1,..C

The number of steps t is set to be large enough such that /tc is close to its limit /^ :— lim^TC/c. If G is connected, it can be shown that/^ is independent of the choice of /0c. Even if G is disconnected, the algorithm can be performed on each connected component separately and again the limit /c for each component will be independent of the choice

of /c.

We will now adapt the label propagation algorithm to higher dimensional walks, namely, walks on oriented edges. Given any random walk on the set of oriented edges (and an absorbing death state ©), its probability transition matrix P could be used to propagate labels in the same manner as the above algorithm. However, this will treat and label the two orientations of a single edge separately as though they are unrelated. We will use walks that are related to the 1-forms and harmonics of the Hodge Laplacian. A major difference with propagation on a goal is that the labels will be oriented. For example, given an oriented edge e+ and a class c, the propagation algorithm may assign a positive confidence that e+ belongs to class c or a negative confidence that e+ belongs to class c, which we view as a positive confidence that e+ belongs to class —c or, equivalently, that e_ belongs to class c. This construction applies to systems in which every class has two built-in orientations or signs, so the class information has a directed sense of "flow".

For example, imagine water flowing along a triangle complex in two dimensions. Given an oriented edge, the water may flow in the positive or negative direction along the edge. A "negative" flow of water in the direction of e+ can be interpreted as a positive flow in the direction of e_. Perhaps the flow along a few edges is observed and one wishes to infer the direction of the flow along all the other edges. Unlike in the graph case, a single class of flow already presents a classification challenge. Alternately, consider multiple streams

of water colored according to the C classes, in which case we may want to know which stream dominates the flow along each edge and in which direction. In order to make these inferences, it is necessary to make some assumption about how labels should propagate from one edge to the next. When considering water flow, it is intuitive to make the following two assumptions.

1. Local Consistency of Motion. If water is flowing along an oriented edge [v,, Vj] in the positive direction, then for every triangle [v,, Vj, vk] the water should also tend to flow along [v,, vk] and [vk, Vj] in the positive directions.

2. Preservation of Mass. The total amount of flow into and out of each vertex (along edges connected to the vertex) should be the same.

In fact, either one of these assumptions is sufficient to infer oriented class labels given the observed flow on a few edges. Depending on which assumptions one chooses, different normalized propagation matrices ALpP (see Section 5) may be applied. For example, setting the Laplacian matrix L to A7 = d251 will enforce local consistency of motion without regard to preservation of mass, while L = Afown = S0d1 will do the opposite. A reasonable way of preserving both assumptions is by using the full Hodge 1-Laplacian L = A1, see Fig. 3 for a visual example of the contrast between walks.

We now state a simple algorithm, analogous to the one for graphs, that propagates labels on edges to infer a partially-observed flow. Let X be a simplicial complex of dimension d > 1 and let X+ = {e1,..., en} be a choice of orientation for the set of edges. Without loss of generality, assume that oriented edges eu+1,..., en=u+t have been classified with class c (not -c). Similar to the graph case, we apply a recursive label propagation process to an initial distribution vector/0c : X+ ^ R measuring the a priori confidence that each oriented edge is in class c. See Algorithm 1 for the procedure. The result of the algorithm is a set of estimates of the relative confidence that each edge is in class c with some orientation.

Algorithm 1: Edge propagtion algorithm.

Data: Simplicial complex X, set of oriented edges

X+ = {e1,..., eu, eu+1,..., eu+t}

with eu+1,..., eu+t labelled with oriented classes ±1,.., ±C, initial distribution vector /0c : X+ ^ R, number of iterations t Result: Confidence of class membership and direction for unlabelled edges

{/: (e1),..,/c (eu)}C=1 for c = 1 to C do for t = 1 to t do

ft ^ TL,pftC-1;

/tc(ei) 1 for ei labelled with class c;

ftc(ei) <--1 for ei labelled with class -c

{fc(e1),...,fc(eu)}C=1 ^ (fc(e1),...,fc(ea)}f= 1;

Fig. 3. (A) An edge labelled on a 2-complex. (B) Label propagation with Alp, note the gradient like flows. (C) Label propagation with Alown, note the short cycles or curl structure. (D) Label propagation with Ai, note the cycle or Harmonic around the boundary.

After running the algorithm, an unlabelled edge e^ is assigned the oriented class sgn (fc (et ))c where c = argmaxc=1,c {f (a )|}.

We now prove that given enough iterations i the algorithm converges and the resulting assigned labels are meaningful. The proof uses the same methods as the one found in [23] for the graph case.

Proposition 6.1. Using the notation of Section 5, assume that L is a Laplacian matrix with Spec(LD-1) с [0, А]. Let ALp be the normalizedp-lazy propagation matrix as defined in Definition 5.2. If 2k+a-2 < p < 1 and if no vector in ker L is supported on the set of unclassified edges, then Algorithm 1 converges. That is,

Hm fT =:fc = ^i - A^)-1 A3 f c) ,

where A4 andA3 are submatrices ofAL,p and tyc is the class function on edges labelled with ±c (for which tyc (ei) = ±1). In addition, f^ depends neither on the initial distributionf0c nor on the lazy probability p.

Proof. First, note that we are only interested in the convergence off£ (ei) for ei not labelled ±c. Partition fc and AL,p according to whether ei is labelled ±c or not as

* = © and AL,p = (A3 J) .

The recursive definition offC in Algorithm 1 can now be rewritten as f£ = Af— + A3fc Solving for f£ in terms off0c yields

f[ = A)kfc + J2 Am*c

In order to prove convergence of//, it suffices to prove that A4 has only eigenvalues strictly less than 1 in absolute value. This ensures that (A4)kf£ converges to zero (eliminating dependence on the initial distribution) and that J] J—1 (A4)lA3fc converges to (I -A4)-1A3fc as k ^ x. We will prove that Spec(A4) c (-1,1) by relating Spec(A4) to Spec(LD-:) c [0, A] as follows.

First, partition L and DL in a similar way to AL,p as

so that

a4 = I -

p(K - 1) + 1

Hence Spec (A4) is determined bySpec(L4D-1),or to be more specific, X e Spec(L4D4:) ^ 1 - Py-P)+iX e Spec(A4). Furthermore, note that L4D-: and D-1/2L4D-1/2 are similar

matrices that share the same spectrum. It turns out that the spectrum of D^1/2L4D41/2 is bounded within the spectrum of D-1/2LD-1/2, which in turn is equal to Spec(LD-1) c [0, A] by similarity. Let g be an eigenvector of D-1/2L4D-1/2 with eigenvalue X and let g1,..., gl be an orthonormal basis of eigenvectors of D-1/2LD-1/2 (such a basis exists since it is a symmetric matrix) with eigenvalues y1,..., yl e [0, A]. We can write

= aigi + ... + atgt

for some a1,..., at, where 0c is the vector of zeros with length equal to the number of edges classified as ±c. Then

a^Yi + ... + at y =

D- 1/2L2D,

D-1/2L1D-1/2

D4-1/2L3D11'2 D-i/2LAD.

D-1/2L2D-1/2

D11/2L2D4-1/2g Xg

= X (a2 + ... + a2)

Because we assumed that Yj e [0, A] for all j, it would be a contradiction if X > A or X < 0. The case X = 0 is possible if and only if aj Yj = 0 for all j and therefore (0gc) e ker L. Since we assumed that no vector in ker L is supported on the unlabelled edges, we conclude that Spec(L4D-1) c (0, A]. Finally, since we assumed that 2kA-a-2 < P < 1, we conclude that

Spec(A4) c - P-K-P+ A, 1 c (-1,1).

To see that the solution fC = (I — A4) 1A3f-c does not depend onp, note that I — A4 is a submatrix of (K1-P)+iLD—1 so that p(K———>+1 (I — a4) does not depend on p. Then write fC as

fc = =

p(K — 1) + 1 . 1 — P

(I — A4)

x --A3fc

1 — P

and note that p(K__—^p+1A3 is an off-diagonal submatrix of p(K-—_^)p+'1I — LD_1 and therefore does not depend on p either. ■

Note that while the limit/c exists, the matrix I — A4 could be ill-conditioned. In practice, it may be better to approximate / c with /tc for large enough t. Also, the algorithm will converge faster for smaller values of p and if/0c — 0.

6.2.1. Simulation Study We use two simulations to illustrate how Algorithm 1 works. In the first example, we describe how a single oriented edge of one class is propagated via the different walks. In the second example, we describe the case where to oriented edges with different class labels are propagated.

Propagating a single edge with one label Figure 3A shows a simplicial complex in which a single oriented edge e1 has been labelled red and all other edges are unlabelled. Figure 3B displays the result of propagating this single label 1000 steps using Algorithm 1 with L — A1p, with the parameter p — 0.9, and/0 equal to the indicator function on e1 (edge e1 is labelled red). The edges in Fig. 3B are oriented according to the sign of/1000. Figures 3C and 3D display the results of propagating the same red edge by L — Afown and L — A1 respectively. Propagation by A1p results in gradient like flows while propagation by Afown results in a more curl like structure. Propagation by the walk corresponding to the full Laplacian A1 results in longer cycles or harmonics as can be seen by the cycle around the boundary of the simplcial complex.

Propagating edges of different labels Figure 4A shows a simplicial complex in which two edges have been labelled with class c — 1 (indicated by the red color) and two other edges have been labelled with class c — 2 (indicated by the blue color). Figure 4B displays the result of 1000 iterations of Algorithm 1 with L — A1, p — 0.9, and/0c equal to the indicator function on the oriented edges labelled with classes c — 1, 2. The orientation of the edges are given by the sign of//—1, if \/c—1\ > tfT—:21, or2, if /c—1\ < tfT—221. Notice that only a small number of labels are needed to induce large-scale circular motion. Near the middle, a few blue labels mix in with the red due to the asymmetry of the initial labels.

7. DISCUSSION

In this paper, we introduced a random walk on simplicial complexes with a stationary distribution that is related to part of the k-th Hodge Laplacian Aj|own = Sk—1dk. We compared our result to the result in [17] which focused on the walk corresponding to part of the k-th Hodge Laplacian Aup = dk+1Sk. We also state a walk that corresponds to the full Hodge Laplacian Ak.

There remain many open questions about random walks on simplicial complexes and the spectral theory of higher order Laplacians. Possible future directions of research include:

Fig. 4. (A) A 2-complex with two different labels on four edges. (B) Edge propagation with two classes with Ai.

1. What is the continuum limit for these random walks for a model where the random complex is generated by sampling a manifold.

2. Is it possible to use conditioning techniques from stochastic processes such as Doob's h-transform to analyze these walks?

3. What applications do these walks have to problems in machine learning and statistics?

ACKNOWLEDGEMENTS

SM would like to thank Anil Hirani, Misha Belkin, and Jonathan Mattingly for useful comments. JS would like to thank Kevin McGoff for proofreading and useful comments.

REFERENCES

[1] J. Callut, K. Frangoisse, M. Saerens, and P. Dupont, Semi-supervised classification from discriminative random walks, Machine learning and knowledge discovery in databases, Springer-Verlag, Berlin-Heidelberg-New York, 2008, pp. 162-177.

[2] F. R. K. Chung, Spectral graph theory, American Mathematical Society, Providence, RI, 1997.

[3] F. R. K. Chung, Random walks and local cuts in graphs, Linear Algebra Appli 423 (2007), 22-32.

[4] F. R. K. Chung, Quasi-random hypergraphs revisited, Random Struct Algorithms 40 (2012), 39-48.

[5] F. R. K. Chung and R. L. Graham, Quasi-random hypergraphs, Random Struct Algorithms 1 (1990), 105-124.

[6] E. Cohen, D. Mubyi, P. Ralli, and P. Tetali, Inverse expander mixing for hypergraphs, Arxiv preprint arXiv:1407.2285, 2014.

[7] J. Dodziuk, Difference equations, isoperimetric inequality and transience of certain random walks, Trans Am Math Soc 284 (1984), 787-794.

[8] D. Dotterrer and M. Kahle, Coboundary expanders, J Topol Anal 4 (2012), 499-514.

[9] B. Eckmann, Harmonische Funktionen und Randwertaufgaben in einem Komplex, Comment Math Helv 17 (1944), 240-255.

[10] J. Friedman and A. Wigderson, On the second eigenvalue of hypergraphs, Combinatorica 15 (1995), 43-65.

[11] T. Jaakkola and M. Szummer, Partially labeled classification with Markov random walks, Advances in neural information processing systems (NIPS), MIT Press, Cambridge, MA, Vol. 14, 2002, pp. 945-952.

[12] M. Kahle, Sharp vanishing thresholds for cohomology of random flag complexes, Ann Math 179 (2014), 1085-1107.

[13] D. A. Levin, Y. Peres, and E. L. Wilmer, Markov chains and mixing times, American Mathematical Society, Providence, RI, 2008.

[14] L. Lovasz, Random walks on graphs: A survey, In D. Miklos, V. T. Sos, and T. Szonyi, editors, Combinatorics, Paul Erdos is eighty, Vol. 2, Janos Bolyai Mathematical Society, Budapest-Hungary, 1996, pp. 353-398.

[15] A. Lubotzky, Ramanujan complexes and high dimensional expanders, Japanese Journal of Mathematics 9 (2014), 137-169.

[16] M. Meila and J. Shi, A random walks view of spectral segmentation, In AI and Statistics (AISTATS) 2001, Key West, Florida, 2001.

[17] O. Parzanchevski and R. Rosenthal, Simplicial complexes: Spectrum, homology and random walks, arXiv:1211.6775v2, 2012.

[18] O. Parzanchevski, R. Rosenthal, and R. J. Tessler, Isoperimetric inequalities in simplicial complexes, Combinatorica, 1-33.

[19] R. Rosenthal, Simplicial branching random walks and their applications, Arxiv preprint arXiv:1412.5406, 2014.

[20] J. Steenbergen, Towards a spectral theory for simplicial complexes, PhD thesis, Duke University, Durham, NC, 2013, AAI3605065.

[21] J. Steenbergen, C. Klivans, and S. Mukherjee, A Cheeger-type inequality on simplicial complexes, Adv Appl Math 56 (2014), 56-77.

[22] D. Zhou and B. Scholkopf, Learning from labeled and unlabeled data using random walks, Pattern recognition, Springer, Berlin-Heidelberg, 2004, pp. 237-244.

[23] X. Zhu, Semi-supervised learning with graphs, PhD thesis, Carnegie Mellon University, Language Technologies Institute, School of Computer Science, Pittsburgh, PA, 2005.