Analysis: An Analysis Of The Viterbi Algorithm

882 Words2 Pages

3.6 The Viterbi Algorithm (HMM)
The Viterbi algorithm analyzes English text. Probabilities are assigned to each word in the context of the sentence. A Hidden Markov Model for English syntax is used in which the probability of the word is dependent on the previous word or words. The probability of word or words followed by a word in the sentence the probability is calculated for bi-gram, tri-gram and 4-gram.Depending on the length of the sentence the probability is calculated for n-grams [1].
3.6.1 What are N-grams:
N-grams of text are extensively used in text mining and natural language processing tasks. They are basically a set of co-occurring words within a given window and when computing the n-gram we move forward. We can move X-words forward in more advanced scenarios. When N=1, it is referred as unigram, when N=2, it is bi-gram, when N=3, it is tri-gram. When N>3, it is 4-gram, 5-gram and so on. …show more content…

The Viterbi Algorithm estimates the sequence of states , that maximizes ) for a given sequence of observations (feature vectors) derived from the M-words in a sentence. The observations x0 and xM+1 as well as the states Z0 and ZM+1 designate the beginning and the end of the sentence.
From Bayes rule we obtain (1)
Since P(Z) is independent of the sequence maximized over all to discriminant function needs to be maximized over all sequences

Open Document