LECTURE 4. Last time: Lecture outline



Similar documents
Random access protocols for channel access. Markov chains and their stability. Laurent Massoulié.

Ex. 2.1 (Davide Basilio Bartolini)

Gambling with Information Theory

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES

IEOR 6711: Stochastic Models, I Fall 2012, Professor Whitt, Final Exam SOLUTIONS

Gambling and Data Compression

M/M/1 and M/M/m Queueing Systems

Master s Theory Exam Spring 2006

Modeling and Performance Evaluation of Computer Systems Security Operation 1

1. (First passage/hitting times/gambler s ruin problem:) Suppose that X has a discrete state space and let i be a fixed state. Let

HOMEWORK 5 SOLUTIONS. n!f n (1) lim. ln x n! + xn x. 1 = G n 1 (x). (2) k + 1 n. (n 1)!

Markov Chains. Chapter Stochastic Processes

Solutions to Homework 6

1 Gambler s Ruin Problem

Introduction to Markov Chain Monte Carlo

Homework set 4 - Solutions

Application of Markov chain analysis to trend prediction of stock indices Milan Svoboda 1, Ladislav Lukáš 2

Lecture 4: BK inequality 27th August and 6th September, 2007

Probability and Random Variables. Generation of random variables (r.v.)

Reinforcement Learning

Hydrodynamic Limits of Randomized Load Balancing Networks

Reinforcement Learning

Lectures on Stochastic Processes. William G. Faris

The Cost of Offline Binary Search Tree Algorithms and the Complexity of the Request Sequence

Chapter 2 Discrete-Time Markov Models

The Chinese Restaurant Process

6.02 Fall 2012 Lecture #5

Load Balancing and Switch Scheduling

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 14 10/27/2008 MOMENT GENERATING FUNCTIONS

Big Data Technology Motivating NoSQL Databases: Computing Page Importance Metrics at Crawl Time

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh

On the mathematical theory of splitting and Russian roulette

THE NUMBER OF GRAPHS AND A RANDOM GRAPH WITH A GIVEN DEGREE SEQUENCE. Alexander Barvinok

Optimal Hiring of Cloud Servers A. Stephen McGough, Isi Mitrani. EPEW 2014, Florence

10.2 Series and Convergence

Random graphs with a given degree sequence

Cooperative Multiple Access for Wireless Networks: Protocols Design and Stability Analysis

Introduction to Probability

INTEGRATED OPTIMIZATION OF SAFETY STOCK

On Directed Information and Gambling

The Exponential Distribution

A linear combination is a sum of scalars times quantities. Such expressions arise quite frequently and have the form

Fast Fourier Transform: Theory and Algorithms

DATA ANALYSIS II. Matrix Algorithms

Message-passing sequential detection of multiple change points in networks

Review Horse Race Gambling and Side Information Dependent horse races and the entropy rate. Gambling. Besma Smida. ES250: Lecture 9.

VERTICES OF GIVEN DEGREE IN SERIES-PARALLEL GRAPHS

1 Short Introduction to Time Series

How To Solve A Sequential Mca Problem

Perron vector Optimization applied to search engines

Transportation Polytopes: a Twenty year Update

Universal hashing. In other words, the probability of a collision for two different keys x and y given a hash function randomly chosen from H is 1/m.

Asymptotics for ruin probabilities in a discrete-time risk model with dependent financial and insurance risks

Continued Fractions and the Euclidean Algorithm

5 Directed acyclic graphs

Performance Analysis of a Telephone System with both Patient and Impatient Customers

Network Protocol Design and Evaluation

Probability Generating Functions

1.5 / 1 -- Communication Networks II (Görg) Transforms

Estimating an ARMA Process

Proximal mapping via network optimization

How will the programme be delivered (e.g. inter-institutional, summerschools, lectures, placement, rotations, on-line etc.):

6.231 Dynamic Programming and Stochastic Control Fall 2008

A Network Flow Approach in Cloud Computing

SPECTRAL POLYNOMIAL ALGORITHMS FOR COMPUTING BI-DIAGONAL REPRESENTATIONS FOR PHASE TYPE DISTRIBUTIONS AND MATRIX-EXPONENTIAL DISTRIBUTIONS

INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS

UNIFORM ASYMPTOTICS FOR DISCOUNTED AGGREGATE CLAIMS IN DEPENDENT RISK MODELS

The Basics of Graphical Models

The Random Waypoint Mobility Model with Uniform Node Spatial Distribution

Notes on Probability Theory

Monte Carlo-based statistical methods (MASM11/FMS091)

Tools for the analysis and design of communication networks with Markovian dynamics

Pablo A. Ferrari Antonio Galves. Construction of Stochastic Processes, Coupling and Regeneration

Lecture 18: Applications of Dynamic Programming Steven Skiena. Department of Computer Science State University of New York Stony Brook, NY

Stationary random graphs on Z with prescribed iid degrees and finite mean connections

On the Traffic Capacity of Cellular Data Networks. 1 Introduction. T. Bonald 1,2, A. Proutière 1,2

Voice Service Support over Cognitive Radio Networks

The PageRank Citation Ranking: Bring Order to the Web

Analysis of a Production/Inventory System with Multiple Retailers

ALGORITHMIC TRADING WITH MARKOV CHAINS

Lecture 13 - Basic Number Theory.

9.2 Summation Notation

3.2 Roulette and Markov Chains

Markov Chains for the RISK Board Game Revisited. Introduction. The Markov Chain. Jason A. Osborne North Carolina State University Raleigh, NC 27695

Information Theory and Stock Market

Monte Carlo and Empirical Methods for Stochastic Inference (MASM11/FMS091)

Transcription:

LECTURE 4 Last time: Types of convergence Weak Law of Large Numbers Strong Law of Large Numbers Asymptotic Equipartition Property Lecture outline Stochastic processes Markov chains Entropy rate Random walks on graphs Hidden Markov models Reading: Chapter 4.

Stochastic processes A stochastic process is an indexed sequence or r.v.s X 0, X 1,... characterized by the joint PMF P X0,X 1,...,X n (x 0, x 1,..., x n ), (x 0, x 1,..., x n ) X n for n = 0, 1,.... A stochastic process is stationary if P X0,X 1,...,X n (x 0, x 1,..., x n ) = P Xl,X l+1,...,x l+n (x 0, x 1,..., x n ) for every shift l and all (x 0, x 1,..., x n ) X n.

Stochastic processes A discrete stochastic process is a Markov chain if P Xn X 0,...,X n 1 xn x 0,..., x n 1 = PXn X n 1 xn x n 1 for n = 1, 2,... and all (x 0, x 1,..., x n ) X n. We deal with time invariant Markov chains X n : state after n transitions belongs to a finite set, e.g., {1,..., m} X 0 is either given or random (given current state, the past does not matter) p i,j = P(X n+1 = j X n = i) = P(X n+1 = j X n = i, X n 1,..., X 0 ) Markov chain is characterized by probability transition matrix P = [p i,j ]

Review of Markov chains State occupancy probabilities, given initial state i: r i,j (n) = P(X n = j X 0 = i) Key recursion: m r i,j (n) = r i,k (n 1)p k,j k=1 With random initial state: m P(X n = j) = P(X 0 = i)r i,j (n) i=1 Does r ij converge to something? Does the limit depend on initial state?

Review of Markov chains Recurrent and transient states. State i is recurrent if: starting from i, and from wherever you can go, there is a way of returning to i. If not recurrent, called transient. Recurrent class collection of recurrent states that communicate to each other and to no other state. A recurrent state is periodic if: there is an integer d > 1 such that r i,i (k) = 0 when k is not an integer multiple of d Assume a single class of recurrent states, aperiodic. Then, lim r i,j (n) = π j n where π j does not depend on the initial conditions lim P(X n = j n X 0 ) = π j

π 1,..., π m can be found as the unique solution of the balance equations π j = π k p k,j k together with πj = 1 j

Entropy rate The entropy rate of a stochastic process is if it exists 1 lim H(X n ) n n For a stationary stochastic process, the entropy rate exists and is equal to lim H(X n X n 1 ) n since conditioning decreases entropy and by stationarity, it holds that H(X n+1 X n ) H(X n+1 X 2 n ) = H(X n X n 1 ) so it reaches a limit (decreasing non-negative sequence) Chain rule n 1 1 n H(Xn ) = H(Xi X i 1 ) n i=1 since the elements in the sum on the RHS reach a limit, that is the limit of the LHS

Entropy rate Markov chain entropy rate: lim H(X n X n 1 ) n = lim H(X n X n 1 ) n = H(X 2 X 1 ) = pi,jπ i log(p i,j ) i,j

Random walk on graph Consider undirected graph G = (N, E, W) where N, E, W are the nodes, edges and weights. With each edge there is an associated edge weight W i,j W i,j = W j,i W i = Wi,j W = j W i,j i,j:j>i 2W = Wi i

Random walk on graph We call a random walk the Markov chain in which the states are the nodes of the graph W i,j p i,j = Wi W i π i = 2W Check: i π i = 1 and i π i p i,j = W i 2W W i,j i W i W i,j = i 2W W j = 2W = π j

Random walk on graph H(X 2 X 1 ) = πi pi,j log(p i,j ) i j = W i W i,j Wi,j log i 2W j W i W i = W i,j Wi,j log i,j 2W W i = W i,j Wi,j log i,j 2W 2W W i,j Wi + log i,j 2W 2W W i,j Wi,j = log + W i Wi log 2W 2W 2W 2W i,j Entropy rate is difference of two entropies i Note: time reversibility for Markov chain that can be represented as random walk on undirected weighted graph

Hidden Markov models Consider an ALOHA wireless model M users sharing the same radio channel to transmit packets to a base station During each time slot, a packet arrives to a user s queue with probability p, independently of the other M 1 users Also, at the beginning of each time slot, if a user has at least one packet in its queue, it will transmit a packet with probability q, independently of all other users If two packets collide at the receiver, they are not successfully transmitted and remain in their respective queues

Hidden Markov models Let X i = (N[1] i, N[2] i,..., N[M] ) denote the random vector at time i where N[m] i is the number of packets that are in user m s queue at time i. X i is a Markov chain. Consider the random vector Y i = (Z[1], Z[2],..., Z[M]) where Z[m] i = 1 if user m transmits during time slot i and Z[i] = 0 otherwise Is Y i Markov?

Hidden Markov processes Let X i, X 2,... be a stationary Markov chain and let Y i = φ(x i ) be a process, each term of which is a function of the corresponding state in the Markov chain Y 1, Y 2,... form a hidden Markov chain, which is not always a Markov chain, but is still stationary What is its entropy rate?

Hidden Markov processes We suspect that the effect of initial information should decay H(Y n Y n 1 ) H(Y n Y n 1, X 1 ) = I(X 1 ; Y n Y n 1 ) should go to 0 Indeed, n lim I (X 1 ; Y n ) = n lim n i=1 I X1 ; Y i Y i 1 = I X 1 ; Y i Y i 1 i=1 since we have an infinite sum in which the terms are non-negative and which is upper bounded by H(X 1 ), the terms must tend to 0

MIT OpenCourseWare http://ocw.mit.edu 6.441 Information Theory Spring 2010 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.