Maximal Likelihood Estimator

Introduction

Let’s say we observe a realization of X0,X1,,XNX_{0},X_{1},\ldots,X_{N} from a \mboxMarkov(λ,P)\mbox{Markov}(\lambda,P). How do we estimate the transition matrix PP? Well lets define some things:

Likelihood

Definition (Log-likelihood)

Given N+1N+1 RVs X0,,XNX_{0},\ldots,X_{N} from \mboxMarkov(λ,P)\mbox{Markov}(\lambda,P) the log-likelihood LL is defined as log(L)=log(λX0)+i,jS(k=0N11{Xk=i,Xk+1=j})log(pij)\log(L)=\log(\lambda_{X_{0}})+\sum\limits_{i,j\in S}\left(\sum\limits_{k=0}^{N-1}\mathbb{1}_{\{X_{k}=i,X_{k+1}=j\}} \right)\log(p_{ij})

Definition (Maxmal likelihood estimator)

For i,jSi,j\in S the maximal likelihood estimator for PP is P^ij=k=0N11{Xk=i,Xk+1=j}k=0N11{Xk=i}\hat P_{ij}=\frac{\sum\limits_{k=0}^{N-1}\mathbb{1}_{\{X_{k}=i,X_{k+1}=j\}}}{\sum\limits_{k=0}^{N-1}\mathbb{1}_{\{X_{k}=i\}}}

Lemma

Assume PP is irreducible and positive recurrent. Then for any i,jSi,j\in S P(limnP^ij=Pij)=1P\left(\lim_{n\to\infty}\hat P_{ij}=P_{ij}\right)=1