Finding Stationary Probability Vector of a Transition Probability Tensor Arising from a Higher-order Markov Chain

Similar documents
NOTES ON LINEAR TRANSFORMATIONS

Continuity of the Perron Root

Similarity and Diagonalization. Similar Matrices

Community Mining from Multi-relational Networks

UNCOUPLING THE PERRON EIGENVECTOR PROBLEM

Mathematics Course 111: Algebra I Part IV: Vector Spaces

THE $25,000,000,000 EIGENVECTOR THE LINEAR ALGEBRA BEHIND GOOGLE

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh

Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013

Solutions to Math 51 First Exam January 29, 2015

Notes on Symmetric Matrices

1 if 1 x 0 1 if 0 x 1

Inner Product Spaces

4.5 Linear Dependence and Linear Independence

DATA ANALYSIS II. Matrix Algorithms

1 Solving LPs: The Simplex Algorithm of George Dantzig

MATH APPLIED MATRIX THEORY

Continued Fractions and the Euclidean Algorithm

MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix.

Duality of linear conic problems

160 CHAPTER 4. VECTOR SPACES

Notes on Determinant

1 Sets and Set Notation.

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

Inner Product Spaces and Orthogonality

Chapter 6. Orthogonality

by the matrix A results in a vector which is a reflection of the given

Lecture 3: Finding integer solutions to systems of linear equations

Mathematical finance and linear programming (optimization)

Lectures notes on orthogonal matrices (with exercises) Linear Algebra II - Spring 2004 by D. Klain

University of Lille I PC first year list of exercises n 7. Review

A characterization of trace zero symmetric nonnegative 5x5 matrices

Matrix Representations of Linear Transformations and Changes of Coordinates

Vector and Matrix Norms

What is Linear Programming?

LINEAR ALGEBRA W W L CHEN

CONSTANT-SIGN SOLUTIONS FOR A NONLINEAR NEUMANN PROBLEM INVOLVING THE DISCRETE p-laplacian. Pasquale Candito and Giuseppina D Aguí

IRREDUCIBLE OPERATOR SEMIGROUPS SUCH THAT AB AND BA ARE PROPORTIONAL. 1. Introduction

Notes V General Equilibrium: Positive Theory. 1 Walrasian Equilibrium and Excess Demand

Section Inner Products and Norms

1. (First passage/hitting times/gambler s ruin problem:) Suppose that X has a discrete state space and let i be a fixed state. Let

Au = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.

ON SEQUENTIAL CONTINUITY OF COMPOSITION MAPPING. 0. Introduction

MA106 Linear Algebra lecture notes

No: Bilkent University. Monotonic Extension. Farhad Husseinov. Discussion Papers. Department of Economics

Introduction to Markov Chain Monte Carlo

a 11 x 1 + a 12 x a 1n x n = b 1 a 21 x 1 + a 22 x a 2n x n = b 2.

1 VECTOR SPACES AND SUBSPACES

Orthogonal Diagonalization of Symmetric Matrices

A Spectral Clustering Approach to Validating Sensors via Their Peers in Distributed Sensor Networks

On the representability of the bi-uniform matroid

Practical Guide to the Simplex Method of Linear Programming

FINITE DIMENSIONAL ORDERED VECTOR SPACES WITH RIESZ INTERPOLATION AND EFFROS-SHEN S UNIMODULARITY CONJECTURE AARON TIKUISIS

PÓLYA URN MODELS UNDER GENERAL REPLACEMENT SCHEMES

LS.6 Solution Matrices

Math 4310 Handout - Quotient Vector Spaces

Similar matrices and Jordan form

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

CONTROLLABILITY. Chapter Reachable Set and Controllability. Suppose we have a linear system described by the state equation

Factorization Theorems

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

Cross product and determinants (Sect. 12.4) Two main ways to introduce the cross product

Metric Spaces. Chapter Metrics

Methods for Finding Bases

SHARP BOUNDS FOR THE SUM OF THE SQUARES OF THE DEGREES OF A GRAPH

Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components

How To Prove The Dirichlet Unit Theorem

THE FUNDAMENTAL THEOREM OF ALGEBRA VIA PROPER MAPS

Perron vector Optimization applied to search engines

A PRIORI ESTIMATES FOR SEMISTABLE SOLUTIONS OF SEMILINEAR ELLIPTIC EQUATIONS. In memory of Rou-Huai Wang

On the D-Stability of Linear and Nonlinear Positive Switched Systems

Systems of Linear Equations

Tools for the analysis and design of communication networks with Markovian dynamics

I. GROUPS: BASIC DEFINITIONS AND EXAMPLES

BANACH AND HILBERT SPACE REVIEW

MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set.

ON LIMIT LAWS FOR CENTRAL ORDER STATISTICS UNDER POWER NORMALIZATION. E. I. Pancheva, A. Gacovska-Barandovska

This exposition of linear programming

Bounds on the spectral radius of a Hadamard product of nonnegative or positive semidefinite matrices

Elasticity Theory Basics

Optimal proportional reinsurance and dividend pay-out for insurance companies with switching reserves

Further Study on Strong Lagrangian Duality Property for Invex Programs via Penalty Functions 1

Linear Algebra I. Ronald van Luijk, 2012

Lecture L3 - Vectors, Matrices and Coordinate Transformations

Bipan Hazarika ON ACCELERATION CONVERGENCE OF MULTIPLE SEQUENCES. 1. Introduction

THE DIMENSION OF A VECTOR SPACE

10.2 Series and Convergence

Some Research Problems in Uncertainty Theory

1.2 Solving a System of Linear Equations

2.3 Convex Constrained Optimization Problems

SYMMETRIC EIGENFACES MILI I. SHAH

Discussion on the paper Hypotheses testing by convex optimization by A. Goldenschluger, A. Juditsky and A. Nemirovski.

Prime Numbers and Irreducible Polynomials

Wald s Identity. by Jeffery Hein. Dartmouth College, Math 100

SPECTRAL POLYNOMIAL ALGORITHMS FOR COMPUTING BI-DIAGONAL REPRESENTATIONS FOR PHASE TYPE DISTRIBUTIONS AND MATRIX-EXPONENTIAL DISTRIBUTIONS

Matrix Calculations: Applications of Eigenvalues and Eigenvectors; Inner Products

7 Gaussian Elimination and LU Factorization

Math 312 Homework 1 Solutions

Adaptive Online Gradient Descent

MATH1231 Algebra, 2015 Chapter 7: Linear maps

Transcription:

Finding Stationary Probability Vector of a Transition Probability Tensor Arising from a Higher-order Markov Chain Xutao Li Michael Ng Yunming Ye 8 February 2011 Abstract. In this paper we develop a new model and propose an iterative method to calculate stationary probability vector of a transition probability tensor arising from a higher-order Markov chain. Existence and uniqueness of such stationary probability vector are studied. We also discuss and compare the results of the new model with those by the eigenvector method for a nonnegative tensor. Numerical examples for ranking and probability estimation are given to illustrate the effectiveness of the proposed model and method. Key words. stationary probability distribution, transition probability, nonnegative tensor, iterative method, higher-order Markov chains. 1 Introduction 1.1 Markov Chains Markov chain concerns about a sequence of random variables, which correspond to the states of a certain system, in such a way that the state at one time epoch depends only on the one in the previous time epoch. We consider a stochastic process {X t, t = 0, 1, 2,...} that takes on a finite set S. For simplicity, we assume that S = {1, 2,, n}. An element in S is called a state of the process. Department of Computer Science, Shenzhen Graduate School, Harbin Institute of Technology. E-mail: xutaolee08@gmail.com. The Corresponding Author. Department of Mathematics, The Hong Kong Baptist University, Hong Kong. E-mail: mng@math.hkbu.edu.hk. Research supported in part by RGC Grants and HKBU FRGs. Department of Computer Science, Shenzhen Graduate School, Harbin Institute of Technology. E-mail: yeyunming@hit.edu.cn. 1

2 Definition 1.1 Suppose there is a fixed probability P ij independent of time such that Prob(X t+1 = i X t = j, X t 1 = i t 1,..., X 0 = i 0 ) = Prob(X t+1 = i X t = j) = P ij where i, j, i 0, i 1,..., i t 1 S. Then this is called a Markov chain process. The probability P ij represents the probability that the process will make a transition to state i given that currently the process is state j. Clearly one has P ij 0, P ij = 1, j = 1,..., n. i=1 The matrix P containing P ij is called the one-step transition probability matrix of the process. Let X t = [x 1 (t), x 2 (t),, x n (t)] T be the probability distribution of the states in a Markov chain process at the tth transition. It is easy to check that X t+1 = P X t and X t+1 = P t+1 X 0. If we let then lim X t = X = [ x 1, x 2,, x n ] T or lim x i (t) = x i, t t X = lim t X t = lim t P X t 1 = P X. (1.1) A vector X is said to be a stationary probability distribution of a finite Markov chain having k states with X 0, n i=1 x i = 1 and P X = X. We remark that the stationary probability distribution can also be given by the normalized eigenvector associated with the largest eigenvalue of P being equal to 1. Theorem 1.1 If P is an irreducible and primitive, then there exist X > 0, X R n such that P X = X, and for any initial distribution X 0, (1.1) satisfies. Interested readers can consult the book by Ross [9] for more detailed information. 1.2 Higher-order Markov Chains There are many situations that one would like to employ higher-order Markov chain models as a mathematical tool and a number of applications can be found in the literature, see for instance in [3] and therein references. The m-th order Markov chain model is used to fit the observed data through the calculation of the m-th order transition probabilities: 0 p i1,i 2,,i m = Prob(X t = i 1 X t 1 = i 2,..., X t m+1 = i m ) 1 (1.2) where i 1, i 2,..., i m S, and p i1,i 2,,i m = 1. (1.3) i 1 =1

3 It is clear when m = 2, the situation reduces to a standard Markov chain discussed in the previous subsection. In [3], several methods are employed to approximate higher-order Markov chains. For instance, the entry p i1,i 2,,i m is approximated by a linear combination of p ij,i j for some j and j, and therefore a higher-order Markov chain can be approximated by a linear combination of transition matrices. In this paper, we are interested in finding a stationary probability vector for a higher-order Markov chain, which is similar to (1.1) and the results in Theorem 1.1. 1.3 Nonnegative Tensor Let R be the real field. We consider an m-order n-dimensional tensor A consisting of n m entries in R: A = (A i1 i 2...i m ), A i1 i 2...i m R, 1 i 1, i 2,..., i m n. A is called non-negative (or, respectively, positive) if A i1 i 2...i m 0 (or, respectively, A i1 i 2...i m > 0). To an n-dimensional column vector X = [x 1, x 2,..., x n ] T R n, we define an n-dimensional column vector: AX m 1 := A ii2...i m x i2 x im (1.4) i 2,...,i m =1 1 i n Definition 1.2 An m-order n-dimensional tensor A is called reducible if there exists a nonempty proper index subset I {1, 2,..., n} such that A i1 i 2...i m = 0, i 1 I, i 2,..., i m / I. If A is not reducible, then we call A irreducible. Definition 1.3 Let A be an m-order n-dimensional tensor and C be the set of all complex numbers. Assume that AX m 1 is not identical to zero. We say (λ, X) C (C n \{0}) is an eigenvalue-eigenvector of A if Here, X [α] = [x α 1, x α 2,..., x α n] T. AX m 1 = λx [m 1]. (1.5) This definition was introduced by Qi [8] when m is even and A is symmetric. Independently, Lim [6] gave such a definition but restricted X to be a real vector and λ to be a real number. In [2], Chang et al. showed the following Perron-Frobenius Theorem for non-negative tensors. Theorem 1.2 If A is an irreducible non-negative tensor of order m and dimension n, then there exist λ 0 > 0 and X > 0, X R n such that A X m 1 = λ 0 X[m 1]. (1.6) Moreover, if λ is an eigenvalue with a non-negative eigenvector, then λ = λ 0. If λ is an eigenvalue of A, then λ λ 0.

4 For the m-th order Markov chain in (1.2) and (1.3), we can consider it as an m- order n-dimensional tensor P consisting of n m entries in between 0 and 1. By using Theorem 1.2, Ng et al. [7] employed the normalized eigenvector associated with the largest eigenvalue of an irreducible non-negative tensor to be the probability distribution of P. Corollary 1.1 If P is an irreducible non-negative tensor of order m and dimension n with (1.2) and (1.3), then there exist λ 0 > 0 and a vector X > 0, X R n such that P X m 1 = λ 0 X[m 1]. (1.7) In particular, we can normalize X such that the summation of its entries is equal to 1, and therefore a probability distribution of the m-th order Markov chain can be defined. However, we may not call it to be a stationary probability distribution as λ 0 may not be equal 1, see the examples in [7]. It is different from a second-order irreducible non-negative tensor (i.e., a Markov chain matrix P ) that the largest eigenvalue of a Markov chain matrix is always equal to one [1], the corresponding eigenvector X > 0 and is normalized with P X = X, see Theorem 1.1 According to the above discussion, we find that there is a gap in between a transition probability matrix and a transition probability tensor. The main aim of this paper is to develop a new model and propose an iterative method in Section 2 to calculate stationary probability vector of a transition probability tensor arising from a higher-order Markov chain. In Section 3, we will compare the results of the new model with those by the eigenvector method for a nonnegative tensor in Corollary 2. Numerical examples are given to illustrate the effectiveness of the proposed model and method. Finally, the concluding remarks are given in Section 4. 2 The Proposed Model We note from (1.1) that a stationary probability distribution of a Markov chain is given P X = X, i.e., the input probability vector after the transition process computation is still equal to the input probability vector. Our idea is to compute an m-th order stationary probability distribution Φ m (i 1, i 2,, i m ) of different states (independent of t) of a transition probability tensor by the same requirement, i.e., Φ 1 (i 1 ) = i 2,,i m =1 Prob(X t = i 1 X t 1 = i 2,..., X t m+1 = i m )Φ m 1 (i 2, i 3,, i m ) (2.8) where Φ 1 (i 1 ) is the first order stationary probability distribution of Φ m defined as follows: Φ 1 (i 1 ) := Φ m (i 1, i 2,, i m ) i 2,,i m=1

5 and Φ m 1 (i 2, i 3,, i m ) is an (m 1)-th order stationary probability distribution of Φ m defined as follows: Φ m 1 (i 2, i 3,, i m ) = Φ m (i 1, i 2,, i m ). i 1 =1 In general, solutions of Φ m in (2.8) may be difficult to obtain. In this paper, one simple form of Φ m to be considered is the product of the first order stationary probability distributions of different states. More precisely, we propose to employ a product form of Φ as follows: m Φ m (i 1, i 2, i 3,, i m ) = Φ 1 (i j ). j=1 It follows that and (2.8) becomes m Φ m 1 (i 2, i 3,, i m ) = Φ 1 (i j ) j=2 Φ 1 (i 1 ) = i 2,,i m =1 m Prob(X t = i 1 X t 1 = i 2,..., X t m+1 = i m ) Φ 1 (i j ) (2.9) It is interesting to note that under the tensor operation, the right hand of (2.9) is the same as PX m with X = [x 1, x 2,, x k ] T and x i = Φ 1 (i). Now we would like to solve a tensor equation: PX m 1 = X (2.10) in order to obtain the stationary probability distribution for a higher order Markov chain. It is clear that (2.8) is different from (1.6) as the right-hand side of (2.10) is not equal to λ 0 X [m 1]. It is straightforward to show that when x i 0 and i x i = 1, we have [PX m 1 ] i 0 for 1 i n and n i=1 [PX m 1 ] i = 1, we can intercept that the probability vector is preserved after the transition probability calculation. Next we state the main results of this paper as follows. Theorem 2.3 If P is an irreducible non-negative tensor of order m and dimension n with (1.2) and (1.3), then there exists a vector X > 0, X R n such that and n i=1 [ X] i = 1. j=2 P X m 1 = X (2.11) Proof: The problem can be reduced to a fixed point problem as follows. Let Ω = {(x 1, x 2,, x n ) R n x i 0, 1 i n, n i=1 x i = 1}. It is clear that Ω is a closed convex set. We define the following map T : Ω Ω: [T (X)] i = [PX m 1 ] i, 1 i n, (2.12)

6 where [PX m 1 ] i is the i-th component of PX m 1. It is clear that T is well-defined and continuous. According to the Brouwer Fixed Point Theorem, there exists X Ω such that T ( X) = X. Next we would like to show that X is positive. Assume that X is not positive, i.e., there exist some entries of X are zero. Let I = {i [ X] i = 0}. It is obvious that I is a proper subset of {1, 2,, n}. Let δ = min{[ X] j j / I}. We must have δ > 0. Since X satisfies P X m 1 = X, we have It follows that δ m 1 i 2,,i m =1 i 2,,i m / I p i,i2,,i m [ X] i2 [ X] im = [ X] i = 0, i I. p i,i2,,i m i 2,,i m / I p i,i2,,i m [ X] i2 [ X] im 0, i I. Hence we have p i,i2,,i m = 0 for all i I and for all i 2, i m / I, i.e., P is reducible. This is a contradiction, and the results follow. In [4], it has been given a general condition which guarantees the uniqueness of the fixed point in the Brouwer Fixed Point Theorem, namely, (i) 1 is not an eigenvalue of the Jacobian matrix of the mapping, and (ii) for each point in the boundary of the domain of the mapping, it is not a fixed point. In our case, we have shown in Theorem 2 that all the fixed points of T are positive when P is irreducible, i.e., they do not lie on the boundary Ω of Ω. The Jacobian matrix DT (x) of T is an n-by-n matrix where its (i 1, j)-th entry is given by i 2,, i m = 1, the summation except i j p i1,i 2,,i m x i2 x im corresponding to the derivative of the i 1 -th entry of PX m with respect to x j. To conclude, we have the following theorem. Theorem 2.4 Suppose P is an irreducible non-negative tensor of order m and dimension n with (1.2) and (1.3). If 1 is not the eigenvalue of DT (x) for all x Ω/ Ω, then then X in Theorem 2.3 is unique. According to Theorems 2.3 and 2.4, we consider and propose a simple iterative algorithm for computing a stationary probability distribution of a transition probability tensor arising from a higher order Markov chain. The Iterative Algorithm: 1. Input X 0 is any n-vector with n i=1 [X 0 ] i = 1;

7 2. Set t = 1; (a) Compute X t = PX m 1 t 1 ; (b) Set t = t + 1; (c) If X t = X t 1, then stop, otherwise goto Step 2(a). The iteration stops whenever X t = X t 1 which yields the solution X = X t as PX m 1 t 1 = X t = X t 1 is a fixed point. If the iteration does not terminate in a finite number of iterations, there exists a subsequence {X tj } converges to X by using the fact that Ω is closed in R n. As X is unique (Theorem 2.4), it implies that X t will converge to X. 3 Numerical Results 3.1 Example 1 In the first example, we compute stationary probability vectors of transition probability tensors arising from higher-order Markov chains in [7]. This example concerns an in-house problem of production planning and inventory control from a large softdrink company. Due to the limitation of storage space and the overall growing sales demand, the company needs to study the sales volume for each product so that it can make better use of the storage space in its central warehouse. According to the level of sales volume, each product can be characterized by one of the following states at each time epoch: state 1: very slow-moving (very low sales volume); state 2: slow-moving; state 3: standard; state 4: fast-moving; state 5: very fast-moving (very high sales volume). The stationary probability vector of each product would help the company in making its production planning. Figure 1 shows the change of the states of products A, B, C and D. Based on the change sequences, we consider the third-order Markov chains and construct the third-order transition probability tensors to compute the stationary probability vectors. Table 1 shows the results computed by the proposed iterative method. As a comparison, we also show the results computed by the eigenvector method in [7] in Table 2. With the results, we can show some advantages of the proposed method over the eigenvector method when dealing with higher-order Markov chains. First, we can see our proposed method results in a stationary probability vector while the eigenvector method only results in a probability distribution, which is not called a stationary probability vector because in the method the largest eigenvalue is not necessarily equal to 1 and the corresponding eigenvector needs to be normalized for obtaining the probabilities. The fact can be verified by the results of the largest eigenvalues λ in Table 2. Next we see in Figure 1 that the computed stationary probability vectors seem better than those by the eigenvector method. Given the sequences of states in Figure

8 5 Product A 5 Product B 4 4 3 3 2 2 1 0 50 100 150 200 250 1 0 50 100 150 200 250 300 5 Product C 5 Product D 4 4 3 3 2 2 1 0 20 40 60 80 100 120 140 1 0 50 100 150 200 250 Figure 1: The change of the states of products A, B, C and D. Product A Product B Product C Product D state 1 0.0262 0.0216 0.1683 0.0089 state 2 0.7782 0.6691 0.3454 0.6556 state 3 0.1233 0.1735 0.3090 0.2038 state 4 0.0506 0.0975 0.1522 0.0764 state 5 0.0217 0.038p3 0.0251 0.0553 Table 1: Computed Results by the Proposed Iterative Method. Product A Product B Product C Product D λ 4.2900 4.6171 4.5247 4.7198 probability 0.1139 0.1364 0.1918 0.1214 distribution 0.3509 0.3062 0.2687 0.2568 for five 0.2089 0.1985 0.2603 0.2416 states 0.1714 0.1939 0.1910 0.2064 0.1550 0.1650 0.0882 0.1737 Table 2: Computed Results by the Eigenvector Method in [7].

9 Product A Product B Product C Product D state 1 0.0106 0 0.1972 0 state 2 0.6915 0.7026 0.3310 0.7163 state 3 0.1702 0.1830 0.2746 0.1869 state 4 0.0745 0.0915 0.1761 0.0519 state 5 0.0532 0.0229 0.0211 0.0450 Table 3: Estimated Stationary Probability Vectors for Products A, B, C and D. Product A Product B Product C Product D l 1 -norm 0.2046 0.0860 0.1056 0.1213 l 2 -norm 0.1073 0.0442 0.0530 0.0690 l -norm 0.0867 0.0335 0.0344 0.0607 Table 4: Distances between Results of the Proposed Method and the Estimated ones. 1, we can estimate the probability of each state as follows: ˆp i = f i 5 j=1 f j, i = 1, 2,..., 5 where ˆp i represents the estimated probability of the state i and f i represents the times that state i being observed in the sequence. Then we obtain the estimated probability vectors for product A, B, C and D, shown as in Table 3. We calculate the distances between the probability vectors by both methods and the estimated ones respectively. The results are shown in Tables 4 and 5. We see from both tables that the results of the proposed method are much closer to the estimated probability vectors than those of the eigenvector method, which indicates the proposed method gives better probability vectors than the eigenvector method does. 3.2 Example 2 In this example, we would like to study the citation ranking among some researchers in the field of data mining by computing the stationary probability vector of a transition probability tensor constructed by their citations. The idea is similar to the PageRank algorithm where the stationary probability vector is computed based on Product A Product B Product C Product D l 1 -norm 0.6813 0.7928 0.1640 0.9178 l 2 -norm 0.3846 0.4546 0.0940 0.5187 l -norm 0.3406 0.3964 0.0671 0.4595 Table 5: Distances between Results of the Eigenvector Method and the Estimated ones.

10 the hyperlink of website among webpages via a column-stochastic matrix, see for instance [5]. In the test, we select fifteen famous researchers in the field of data mining. Their names shown as in Table 6. Then we construct a third-order fifteen dimensions transition probability tensor P based on their citation information in the The ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD). Our idea is as follows: (1) For each researcher, we select ten most-cited papers from his publications in the KDD conference. (2) Given any three indices i 1, i 2 and i 3 (1 i 1, i 2, i 3 15), we compute how much citations the i 1 -th researcher contributes to the i 3 -th researcher via the i 2 -th researcher and form a tensor A as the following formula: 10 10 c(i 1, j, i 2 ) c(i 2, j, i 3 ) j=1 j=1 A i3 i 2 i 1 = 10 10 k(i 1, j) k(i 2, j) j=1 j=1 where the function c(α, β, γ) means the number of times that the γ-th researcher in the reference list of the β-th paper of the α-th researcher, and the function k(α, β) means the total number of references in the β-th paper of the α-th researcher. (3) Normalize the tensor A into a transition probability tensor P. In the construction, we do not consider any self-citation, i.e., we set P] i,i,i = 0 for all i. The purpose is to avoid the ambiguous high transition probability by selfcitation. Table 7 shows the stationary probability vector computed by the proposed method. The result can characterize the citation ranking among these fifteen researchers. In our results, we find that Jiawei Han and Philip S. Yu (their stationary probabilities are 0.1420 and 0.1350 respectively) have higher ranks in citation than the other thirteen researchers in the field of data mining. To validate the results, we show in Table 8 the number of citations links among fifteen researchers. We note that there are at most 15 15 1 = 224 possible links from other researchers to a particular researcher. We see from Table 8 that Jiawei Han and Philip S. Yu have more citation links than the other researchers. Indeed, we compare the ranking of stationary probabilities in Table 7 and the number of citation links in Table 8, and find that they are consistent except for five researchers: Ravi Kumar (#8), Eamonn J. Keogh (#10), Mohammed Javeed Zaki (#12), Hui Xiong (#13) and Jieping Ye (#14). In the ranking of stationary probabilities, they are #8 > #13 > #10 > #12 > #14. while in the ranking of citation links, they are #13 > #12 > #10 > #8 > #14. It can be explained by their different citation fractions in the transition probability tensor.

11 ID Name ID Name ID Name 1 Christos Faloutsos 6 Jian Pei 11 Vipin Kumar 2 Jiawei Han 7 Martin Ester 12 Mohammed Javeed Zaki 3 Philip S. Yu 8 Ravi Kumar 13 Hui Xiong 4 Padhraic Smyth 9 Bing Liu 14 Jieping Ye 5 Heikki Mannila 10 Eamonn J. Keogh 15 Charu C. Aggarwal Table 6: The Fifteen Selected Researchers. ID Stat. Prob. ID Stat.Prob. ID Stat. Prob. 1 0.0700 6 0.0869 11 0.0910 2 0.1420 7 0.0234 12 0.0295 3 0.1350 8 0.0348 13 0.0342 4 0.0591 9 0.0484 14 0.0268 5 0.0972 10 0.0306 15 0.0911 Table 7: The Stationary Probability Vector Computed by the Proposed Method. ID Number of Links ID Number of Links ID Number of Links 1 67 6 72 11 76 2 96 7 8 12 23 3 96 8 17 13 24 4 55 9 55 14 14 5 90 10 20 15 90 Table 8: The Number of Citation Links from the Others Researcher to a Particular Researcher.

12 4 Concluding Remarks In this paper, we have developed a new model and presented an iterative method for computing stationary probability vector of a transition probability tensor arising from a higher-order Markov chain. We have proved the proposed iterative method converges to a unique stationary probability vector for given any initial probability vector. Numerical results have shown the effectiveness of the model and the proposed method. The proposed model would be useful to many applications in ranking [5] and probability estimation, see for instance [3]. Theoretically, we need the assumption in Theorem 2.4 in order to show that X is unique. We have tested many different initial probability distributions in the examples in Section 3, the same resulting probability vector is obtained. We conjecture that the assumption in the theorem may be true for irreducible transition probability tensors with some additional properties. We do not have a proof yet, so we leave it as an open problem for future analysis. Acknowledgement We would like to thank Professors Shmuel Friedland and Lek-Heng Lim for their suggestions and comments when we presented our work at the JRI Workshop on Eigenvalues of Nonnegative Tensors at The Hong Kong Polytechnic University on 18 December 2010. References [1] A. Berman and R. Plemmons, Nonnegative Matrices in the Mathematical Sciences, Classics in Applied Mathematics, SIAM, 1994. [2] K. C. Chang, K. Pearson and T. Zhang, Perron-Frobenius theorem for nonnegative tensors, Commu. Math. Sci., 6 (2008) 507-520. [3] W. Ching and M. Ng, Markov Chains: Models, Algorithms and Applications, International Series on Operations Research and Management Science, Springer, 2006. [4] R. Kellogg, Uniqueness in the Schauder Fixed Point Theorem, Proceedings of the American Mathematical Society, 60(1976), pp. 207 210. [5] A. N. Langville, C. D. Meyer, Deeper inside PageRank, Internet Mathematics, 1(2005), pp. 335 380. [6] L.-H. Lim, Singular values and eigenvalues of tensors: a variational approach, Proceedings of the IEEE International Workshop on Computational Advances in Multi-Sensor Adaptive Processing (CAMSAP 05), 1 (2005) 129-132.

13 [7] M. Ng, L. Qi and G. Zhou, Finding the largest eigenvalue of a non-negative tensor, SIAM J. Matrix Anal. Appl., 31 (2009) 1090-1099. [8] L. Qi, Eigenvalues of a real supersymmetric tensor, Journal of Symbolic Computation, 40 (2005) 1302-1324. [9] S. Ross, Introduction to Probability Models, Academic Press, 2003.