Professional Documents
Culture Documents
Transition
Prob.
Output Prob.
This is a possible hidden Markov model for the protein ACCY. The protein is
probability that an amino acid occurs in a particular state, and the numbers next to
the directed arcs show probabilities, which connect the states. The probability of
ACCY is shown as a highlighted path through the model. There are three kinds of
states represented by three different shapes. The squares are called match states,
and the amino acids emitted from them form the conserved primary structure of a
protein. These amino acids are the same as those in the common ancestor or, if
not, are the result of substitutions. The diamond shapes are insert states and emit
amino acids that result from insertions. The circles are special, silent states known
as delete states and model deletions. These type of HMMs are called Protein
Any sequence can be represented by a path through the model. The probability of
any sequence, given the model, is computed by multiplying the emission and
transition probabilities along the path. A path through the model represented by
is 0.3, and the probability of C being emitted in position 2 is 0.6. The probability of
.4*.3*.46*.6*.97*.5*.015*.73*.01*1 = 1.76x10-6.
1) Scoring Problem
It can be seen that one method of calculating the probability of the observed
sequence would be to find each possible sequence of the hidden states, and sum
I0
I1
M1
M3
M2
Consider the HMM shown above. In this figure several paths exist for the protein
sequence ACCY.
1
The Forward algorithm employs a matrix, shown below. The columns of the
matrix are indexed by the states in the model, and the rows are indexed by the
sequence. The elements of the matrix are initialized to zero and then computed
1. The probability that the amino acid A was generated by state I0 is computed
and entered as the first element of the matrix. This is .4*.3 = .12
the most likely transition to state M1 from state I0) and in state I1 (multiplied
by the most likely transition to state I1 from state I0) are entered into the
The probability of the sequence is found by summing the probabilities in the last
column.
2) Alignment Problem
sequence the most likely sequence of underlying hidden states that might have
2
generated it. This is the alignment problem and the Viterbi Algorithm is used to
maximum rather than a sum is calculated. The most likely path through the model
Once the most probable path through the model is known, the probability of a
sequence given the model can be computed by multiplying all probabilities along
the path.
3) Training Problem
Another tricky problem is how to create an HMM in the first place, given a
probabilities from a set of related training sequences. This is done by using the
3
The algorithm proceeds by making an initial guess of the parameters (which may
well be entirely wrong) and then refining it by assessing its worth, and attempting
to reduce the errors it provokes when fitted to the given data. In this sense, it is
4
2. Applications of HMM’s
In this section, we will delve into greater depth at specific problems in the area of
We introduce here the gene-prediction HMMs that can be used to predict the
structure of the gene. Our objective is to find the coding and non-coding regions of
methods
other processes
5
As shown in the diagram above, a string of DNA nucleotides containing a gene will
In the process of transcription, only the exons are left to form the protein sequence
as depicted below.
6
Many problems in biological sequence analysis have a grammatical structure
. HMMs are very useful in modeling grammar. The input to such a HMM is the
genomic DNA sequence and the output, in the simplest case is a parse tree of
Shown below is a simple model for unspliced genes that recognizes the start
codon, stop codon (only one of the three possible stop codons are shown) and the
coding/non-coding regions. This model has been trained with a test set of gene
data.
DNA ? We simply use the Viterbi algorithm to find the most probable path through the
model
7
8
2.2. Protein- Profile HMMs
statistical model of a protein family which is called a profile. The idea is, given a
single amino acid target sequence of unknown structure, we want to infer the
set of related proteins. This HMM in a natural way can model positional dependant
gap penalties.
The basic topology of a profile HMM is shown above. Each position, or module, in
Matching
Deletion
states Insertion states
states
the model has three states. A state shown as a rectangular box is a match state
letters between two alignment positions, and a state shown by a circle models a
are connected, as shown by lines. For each of these lines there is an associated
9
`transition probability', which is the probability of going from one state to the
other.
The match state represents a consensus amino acid for this position in the protein
family. The delete state is a non-emitting state, and represents skipping this
consensus position in the multiple alignment. Finally, the insert state models the
(http://pfam.wustl.edu).
Building profiles from a family of proteins(or DNA) a profile HMM can be made for
searching a database for other members of the family. As we have seen before in
the section on HMM problems, profile HMM’s can also be used for the following
Scoring a sequence
Given a HMM for a protein family and some unknown sequences, we are trying to
find a path through the model where the new sequence fits in or we are tying to
states to each residue in the sequence. There are many such alignments and the
Vitterbi’s algorithm is used to give the probability of the sequence for that
alignment.
10
Creating Multiple sequence alignment
unaligned sequences. By taking a close look at the alignment, we can see the
history of evolution. One great advantage of HMMs is that they can be estimated
from sequences, without having to align the sequences first. The sequences used to
estimate or train the model are called the training sequences, and any reserved
sequences used to evaluate the model are called the test sequences. The model
estimation is done with the forward-backward algorithm, also known as the Baum-
training sequences.
The method is to train the four HMMs of secondary structure – helix, sheet, turn
and other – by training sequences. The Baum-Welch method is used to train the
HMMs. So, the HMM of helix is able to produce helix-like sequences with high
probabilities. Now, these HMMs can be used to predict the secondary structure of
probabilities of these HMMs outputting the test sequence. The sequence has the
secondary structure whose HMM showed the highest probability to output the
sequence.
11
3. HMM implementation
HMMER - http://hmmer.wustl.edu/
4. Advantages of HMMs
position.
5. Limitations of HMMs
Linear Model
So, they are unable to capture higher order correlations among amino-acids.
12
Standard Machine Learning Problems
In the training problem, we need to watch out for local maxima and so model may
not converge to a truly optimal parameter set for a given training set. Secondly,
since the model is only as good as your training set, this may lead to over-fitting.
member protein family when one exists to better the parameterization of the
Model architecture
architectures that can fit the observed data. Is this the best architecture to use?
Can one use protein structure knowledge to make better architecture decisions, or,
in limited regions, to learn the architecture directly from the data? Will these
Biological mechanism
In gene prediction, the HMM’s may be getting close to replicating the same sort of
accuracy as the biological machine (the HMM’s have the additional task of finding
the gene in the genomic DNA context, which is not handled by the biological
machine that processes the RNA). What constraints does our statistical model
13
place on the biological mechanism— in particular, can we consider a biological
7. References
3. Krogh, A., M. Brown, I.S. Mian, K. Sjolander, and D. Haussler. Hidden Markov
1989.
6. Baldi, P., Chauvin, Y., Hunkapiller, T. and McClure, M.A. (1993) Hidden Markov
Neural Information Processing Systems 5, Eds. S.J. Hanson, J.D. Cowan and C.
7. Baldi, P., Chauvin, Y., Hunkapiller, T., and McClure, M.A. (1994) Hidden Markov
generalized hidden markov model for the recognition of human genes in DNA,
14
Procedings of the Fourth International Conference on Intelligent Systems for
Biosciences, 12:95-107.1996
http://www.cse.ucsc.edu/research/compbio/html_format_papers/hughkrogh96/c
abios.html
http://www.research.ibm.com/journal/rd/453/birney.html
13.HMMer - http://hmmer.wustl.edu/
15