Academia.edu no longer supports Internet Explorer.
To browse Academia.edu and the wider internet faster and more securely, please take a few seconds to upgrade your browser.
2022, arXiv (Cornell University)
Markov chains are a class of probabilistic models that have achieved widespread application in the quantitative sciences. This is in part due to their versatility, but is compounded by the ease with which they can be probed analytically. This tutorial provides an in-depth introduction to Markov chains, and explores their connection to graphs and random walks. We utilize tools from linear algebra and graph theory to describe the transition matrices of different types of Markov chains, with a particular focus on exploring properties of the eigenvalues and eigenvectors corresponding to these matrices. The results presented are relevant to a number of methods in machine learning and data mining, which we describe at various stages. Rather than being a novel academic study in its own right, this text presents a collection of known results, together with some new concepts. Moreover, the tutorial focuses on offering intuition to readers rather than formal understanding, and only assumes basic exposure to concepts from linear algebra and probability theory. It is therefore accessible to students and researchers from a wide variety of disciplines.
Springer Series in Synergetics, 2011
Journal of Theoretical Probability - J THEOR PROBABILITY, 2001
We consider random walks with small fixed steps inside of edges of a graph <img src="/fulltext-image.asp?format=htmlnonpaginated&src=R47027122T5P5217_html\10959_2004_Article_300626_TeX2GIFIE1.gif" border="0" alt=" $${\mathcal{G}}$$ " />, prescribing a natural rule of probabilities of jumps over a vertex. We show that after an appropriate rescaling such random walks weakly converge to the natural Brownian motion on <img src="/fulltext-image.asp?format=htmlnonpaginated&src=R47027122T5P5217_html\10959_2004_Article_300626_TeX2GIFIE2.gif" border="0" alt=" $${\mathcal{G}}$$ " /> constructed in Ref. 1.
2006
, except for brief excerpts in connection with reviews or scholarly analysis. Use in connection with any form of information storage and retrieval, electronic adaptation, computer software, or by similar or dissimilar methodology now know or hereafter developed is forbidden. The use in this publication of trade names, trademarks, service marks and similar terms, even if the are not identified as such, is not to be taken as an expression of opinion as to whether or not they are subject to proprietary rights.
Probability Theory and Related Fields, 2004
Starting with a sequence of i.i.d. [uniform] random variables with m possible values, we consider the overlapping Markov chain formed by sliding a window of size k through the i.i.d. sequence. We study the limiting covariance matrix B k of this Markov chain and give algorithms for constructing the eigenvectors of B k. We also discuss the applicability of the results in strengthening Pearson's χ 2 test as well as the relation to approximate entropy and the usefulness in the area of testing the hypothesis of uniformity of random number generators.
2002
We obtain results on the sensitivity of the invariant measure and other statistical quantities of a Markov chain with respect to perturbations of the transition matrix. We use graph-theoretic techniques, in contrast with the matrix analysis techniques previously used.
In this paper a new proof is given for the supermodularity of information content. Using the decomposability of the information content an algorithm is given for discovering the Markov network graph structure endowed by the pairwise Markov property of a given probability distribution. A discrete probability distribution is given for which the equivalence of Hammersley-Clifford theorem is fulfilled although some of the possible vector realizations are taken on with zero probability. Our algorithm for discovering the pairwise Markov network is illustrated on this example, too.
2019
The paper is devoted to studies of perturbed Markov chains commonly used for description of information networks. In such models, the matrix of transition probabilities for the corresponding Markov chain is usually regularised by adding a special damping matrix multiplied by a small damping (perturbation) parameter $\varepsilon$. We give effective upper bounds for the rate of approximation for stationary distributions of unperturbed Markov chains by stationary distributions of perturbed Markov chains with regularised matrices of transition probabilities, asymptotic expansions for approximating stationary distributions with respect to damping parameter, as well as explicit upper bounds for the rate of convergence in ergodic theorems for $n$-step transition probabilities in triangular array mode, where perturbation parameter $\varepsilon \to 0$ and $n \to \infty$, simultaneously. The results of numerical experiments are also presented
Journal of Physics A: Mathematical and General, 2005
Random walks on graphs are widely used in all sciences to describe a great variety of phenomena where dynamical random processes are affected by topology. In recent years, relevant mathematical results have been obtained in this field, and new ideas have been introduced, which can be fruitfully extended to different areas and disciplines. Here we aim at giving a brief but comprehensive perspective of these progresses, with a particular emphasis on physical aspects. Contents 1 Introduction 2 Mathematical description of graphs 3 The random walk problem 4 The generating functions 5 Random walks on finite graphs 6 Infinite graphs 7 Random walks on infinite graphs 8 Recurrence and transience: the type problem 9 The local spectral dimension 10 Averages on infinite graphs 11 The type problem on the average 1 12 The average spectral dimension 21 13 A survey of analytical results on specific networks 23 13.1 Renormalization techniques. .
The European Physical Journal Special Topics, 2010
Markov chains provide us with a powerful tool for studying the structure of graphs and databases in details. We review the method of generalized inverses for Markov chains and apply it for the analysis of urban structures, evolution of languages, and musical compositions. We also discuss a generalization of Lévy flights over large complex networks and study the interplay between the nonlinearity of diffusion process and the topological structure of the network.
IEEE Transactions on Circuits and Systems I-regular Papers, 2009
The Annals of Statistics, 1997
Undirected graphs and acyclic digraphs (ADGs), as well as their mutual extension to chain graphs, are widely used to describe dependences among variables in multivariate distributions. In particular, the likelihood functions of ADG models admit convenient recursive factorizations that often allow explicit maximum likelihood estimates and that are well suited to building Bayesian networks for expert systems. Whereas the undirected graph associated with a dependence model is uniquely determined, there may, however, be many ADGs that determine the same dependence (= Markov) model. Thus, the family of all ADGs with a given set of vertices is naturally partitioned into Markov-equivalence classes, each class being associated with a unique statistical model. Statistical procedures, such as model selection or model averaging, that fail to take into account these equivalence classes, may incur substantial computational or other inefficiencies. Here it is shown that each Markov-equivalence class is uniquely determined by a single chain graph, the essential graph, that is itself Markov-equivalent simultaneously to all ADGs in the equivalence class. Essential graphs are characterized, a polynomial-time algorithm for their construction is given, and their applications to model selection and other statistical questions are described. 1. Introduction. The use of directed graphs to represent possible dependencies among statistical variables dates back to Wright (1921) and has generated considerable research activity in the social and natural sciences. Since 1980, particular attention has been directed at graphical Markov models specified by conditional independence relations among the variables, i.e., by the Markov properties determined by the graph. Both directed and undirected graphs have found extensive applications, the latter in such areas as spatial statistics and image analysis. The recent books by Whittaker (1990) and Lauritzen (1995) conveniently summarize the statistical perspective on these developments.
Random Structures and Algorithms, 1996
Linear Algebra and its Applications, 2004
For an irreducible stochastic matrix T , we consider a certain condition number c(T ), which measures the stability of the corresponding stationary distribution when T is perturbed. We characterize the strongly connected directed graphs D such that c(T ) is bounded as T ranges over S D , the set of stochastic matrices whose directed graph is contained in D. For those digraphs D for which c(T ) is bounded, we find the maximum value of c(T ) as T ranges over S D .
Journal of Chemical Physics, 2020
Markov chains can accurately model the state-to-state dynamics of a wide range of complex systems, but the underlying transition matrix is ill-conditioned when the dynamics feature a separation of timescales. Graph transformation (GT) provides a numerically stable method to compute exact mean first passage times (MF-PTs) between states, which are the usual dynamical observables, in continuous-time Markov chains (CTMCs). Here, we generalize the GT algorithm to discrete-time Markov chains (DTMCs), which are commonly estimated from simulation data, for example in the Markov State Model approach. We then consider the dimensionality reduction of CTMCs and DTMCs, which aids model interpretation and facilitates more expensive computations, including sampling of pathways. We perform a detailed numerical analysis of existing methods to compute the optimal reduced CTMC, given a partitioning of the network into metastable communities (macrostates) of nodes (microstates). We show that approaches based on linear algebra encounter numerical problems that arise from the requisite metastability. We propose an alternative approach, using GT to compute the matrix of intermicrostate MFPTs in the original Markov chain, from which a matrix of weighted intermacrostate MFPTs can be obtained. We also propose an approximation to the weighted-MFPT matrix in the strongly metastable limit. Inversion of the weighted-MFPT matrix, which is better conditioned than the matrices that must be inverted in alternative dimensionality reduction schemes, then yields the optimal reduced Markov chain. The superior numerical stability of the GT approach therefore enables us to realize optimal Markovian coarse-graining of systems with rare event dynamics.
2015
This thesis has two primary areas of focus. First we study connection graphs, which are weighted graphs in which each edge is associated with a d-dimensional rotation matrix for some fixed dimension d, in addition to a scalar weight. Second, we study non-backtracking random walks on graphs, which are random walks with the additional constraint that they cannot return to the immediately previous state at any given step. Our work in connection graphs is centered on the notion of consistency, that is, the product of rotations moving from one vertex to another is independent of the path taken, and a generalization called epsilon-consistency. We present higher dimensional versions of the combinatorial Laplacian matrix and normalized Laplacian matrix from spectral graph theory, and give results characterizing the consistency of a connection graph in terms of the spectra of these matrices. We generalize several tools from classical spectral graph theory, such as PageRank and effective resi...
arXiv (Cornell University), 2019
Complex networks or graphs are ubiquitous in sciences and engineering: biological networks, brain networks, transportation networks, social networks, and the World Wide Web, to name a few. Spectral graph theory provides a set of useful techniques and models for understanding 'patterns of interconnectedness' in a graph. Our prime focus in this paper is on the following question: Is there a unified explanation and description of the fundamental spectral graph methods? There are at least two reasons to be interested in this question. Firstly, to gain a much deeper and refined understanding of the basic foundational principles, and secondly, to derive rich consequences with practical significance for algorithm design. However, despite half a century of research, this question remains one of the most formidable open issues, if not the core problem in modern network science. The achievement of this paper is to take a step towards answering this question by discovering a simple, yet universal statistical logic of spectral graph analysis. The prescribed viewpoint appears to be good enough to accommodate almost all existing spectral graph techniques as a consequence of just one single formalism and algorithm.
IEEE Transactions on Information Theory
We study the following learning problem with dependent data: Observing a trajectory of length n from a stationary Markov chain with k states, the goal is to predict the next state. For 3 ≤ k ≤ O(√ n), using techniques from universal compression, the optimal prediction risk in Kullback-Leibler divergence is shown to be Θ(k 2 n log n k 2), in contrast to the optimal rate of Θ(log log n n) for k = 2 previously shown in [FOPS16]. These rates, slower than the parametric rate of O(k 2 n), can be attributed to the memory in the data, as the spectral gap of the Markov chain can be arbitrarily small. To quantify the memory effect, we study irreducible reversible chains with a prescribed spectral gap. In addition to characterizing the optimal prediction risk for two states, we show that, as long as the spectral gap is not excessively small, the prediction risk in the Markov model is O(k 2 n), which coincides with that of an iid model with the same number of parameters. Contents
Linear Algebra and its Applications, 2016
Computational procedures for the stationary probability distribution, the group inverse of the Markovian kernel and the mean first passage times of a finite irreducible Markov chain, are developed using perturbations. The derivation of these expressions involves the solution of systems of linear equations and, structurally, inevitably the inverses of matrices. By using a perturbation technique, starting from a simple base where no such derivations are formally required, we update a sequence of matrices, formed by linking the solution procedures via generalized matrix inverses and utilising matrix and vector multiplications. Four different algorithms are given, some modifications are discussed, and numerical comparisons made using a test example. The derivations are based upon the ideas outlined in Hunter,
Loading Preview
Sorry, preview is currently unavailable. You can download the paper by clicking the button above.