Jump to content

Talk:Forward algorithm

Page contents not supported in other languages.
From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by Avidelego (talk | contribs) at 18:45, 14 April 2010 (Created page with 'Why does this page forward to Viterbi?? The Forward algorithm and Viterbi are different algorithms with different purposes. ====Filtering==== The Forward algorith...'). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.
(diff) ← Previous revision | Latest revision (diff) | Newer revision → (diff)

Why does this page forward to Viterbi?? The Forward algorithm and Viterbi are different algorithms with different purposes.

Filtering

The Forward algorithm (in the context of a Hidden Markov Model) is used to calculate a belief state, which is the probability of a state at a certain time, given the history of evidence. This process is also known as filtering. For an HMM such as this one:

Temporal evolution of a hidden Markov model
Temporal evolution of a hidden Markov model

this probability is written as . (here I'm using subscripts to abbreviate as .) A Belief state can be calculated at each time step, but doing this does not, in a strict sense, produce the most likely state sequence, but rather the most likely state at each time step, given the previous history.

Smoothing

In order to take into account future history (say if you want to improve your estimate for past times), you then run the Backward algorithm, a complement of the Forward. This is called smoothing. Mathematically, we would say that the Forward/Backward algorithm computes for all . So you use the full F/B algorithm to take into account all evidence.

Decoding

What if we want the most likely sequence? Enter Viterbi. This algorithm computes the most likely state sequence given the history of observations, that is, the state sequence that maximizes .

While it may be conceptually difficult to understand the difference between the state sequence that Viterbi estimates versus the evolving estimate that you get from the Forward algorithm, the bottom line is that they are different.


If you want to learn more about this I recommend Russel and Norvig's Artificial Intelligence, a Modern Approach. No, it's not free online anywhere, but it's worth owning a copy if you are a computer science/autonomy student, or have a research interest in artificial intelligence. They explain this stuff succinctly and completely starting on page 541 of the 2003 edition. Avidelego (talk) 18:45, 14 April 2010 (UTC)[reply]