Capture the same
information
The importance of a tokens position in the context of the search term
The sequential order of tokens
Different in complexity
Bigram model
Simplified Markov model
with each token as a state
Captures token
sequential information by bigram probabilities
PHMM model
More complex aggregated token sequential
information by hidden state
transition probabilities
Experimental results show
PHMM is less sensitive to model
length
PHMM may benefit more by using more
training data