I'm trying to understand hidden Markov model (HMM). Here is the material which I studied.
It states that there are two assumptions in HMM (page 3):
- $P( q_i | q_1, ..., q_{i-1} ) = P( q_i | q_{i-1} )$
- $P(o_i | q_1, ..., q_T, o_1, ..., o_T) = P(o_i | q_i)$
where $q_i$ denotes the i'th state, $o_i$ denotes the i'th observation, T is the length of sequence.
And in forward-backward algorithm, we need to define the backward probability (page 12, eq A.15):
- $\beta_t(i) = P(o_{t+1}, ..., o_{T} | q_t=i)$
Here explains that backward probability is "the probability of emitting the remaining sequence from t+1 until the end of time after being at hidden state i at time t".
My question is about the assumption (2) and $\beta_t(i)$. Assumption (2) says that the i'th observation only depends on the i'th state. Backward probability only considers $o_{t+1}, ..., o_T$, so they only depend on $q_{t+1}, ..., q_T$, right? Thus I don't know why conditional on $q_i=i$ is needed in backward probability.
In other words, why can't we state that:
- $\beta_t(i)=P(o_{t+1}, ..., o_{T} | q_t=i) = P(o_{t+1}, ..., o_{T})$
Very thanks!