SlideShare a Scribd company logo
Maximum Likelihood
Likelihood
The likelihood is the probability of the data given the
model.
If we flip a coin and get a head and we think the coin is
unbiased, then the probability of observing this head is 0.5.
If we think the coin is biased so that we expect to get a head
80% of the time, then the likelihood of observing this datum (a
head) is 0.8.
The likelihood of making some observation is entirely
dependent on the model that underlies our assumption.
The datum has not changed, our model has. Therefore under
the new model the likelihood of observing the datum has
changed.
Likelihood
Maximum Likelihood (ML)
ML assumes a explicit model of sequence evolution. This is
justifiable, since molecular sequence data can be shown to
have arisen according to a stochastic process.
ML attempts to answer the question:
What is the probability that I would observe these data (a
multiple sequence alignment) given a particular model of
evolution (a tree and a process)?
Likelihood calculations
In molecular phylogenetics, the data are an alignment of sequences
We optimize parameters and branch lengths to get the maximum likelihood
Each site has a likelihood
The total likelihood is the product of the site likelihoods
The maximum likelihood tree is the tree topology that gives the highest
(optimized) likelihood under the given model.
We use reversible models, so the position of the root does not matter.
What is the probability of observing a G nucleotide?
If we have a DNA sequence of 1 nucleotide in length and the identity of this
nucleotide is G, what is the likelihood that we would observe this G?
In the same way as the coin-flipping observation, the likelihood of observing
this G is dependent on the model of sequence evolution that is thought to
underlie the data.
Model 1: frequency of G = 0.4 => likelihood(G) = 0.4
Model 2: frequency of G = 0.1 => likelihood(G) = 0.1
Model 3: frequency of G = 0.25 => likelihood(G) = 0.25
What about longer sequences?
If we consider a gene of length 2
gene 1 GA
The the probability of observing this gene is the product of the
probabilities of observing each character
Model frequency of G = 0.4 frequencyof A= 0.15
p(G) = 0.4 p(A) =0.15
Likelihood (GA) = 0.4 x 0.15 = 0.06
…or even longer sequences?
gene 1 GACTAGCTAGACAGATACGAATTAC
Model simple base frequency model
p(A)=0.15; p(C)=0.2; p(G)=0.4; p(T)=0.25;
(the sum of all probabilities must equal 1)
Likelihood (gene 1) = 0.000000000000000018452813
Note about models
You might notice that our model of base frequency is not the
optimal model for our observed data.
If we had used the following model
p(A)=0.4; p(C) =0.2; p(G)= 0.2; p(T) = 0.2;
The likelihood of observing the gene is
L (gene 1) = 0.000000000000335544320000
L (gene 1) = 0.000000000000000018452813
The datum has not changed, our model has. Therefore under
the new model the likelihood of observing the datum has
changed.
Increase in model sophistication
It is no longer possible to simply invoke a model that
encompasses base composition, we must also include the
mechanism of sequence change and stasis.
There are two parts to this model - the tree and the process
(the latter is confusingly referred to as the model, although
both parts really compose the model).
Different Branch Lengths
For very short branch lengths, the probability of a character staying the
same is high and the probability of it changing is low.
For longer branch lengths, the probability of character change becomes
higher and the probability of staying the same is lower.
The previous calculations are based on the assumption that the branch
length describes one Certain Evolutionary Distance or CED.
If we want to consider a branch length that is twice as long (2 CED), then
we can multiply the substitution matrix by itself (matrix2
).
I (A) II (C)
I (A) II (C)
v = 0.1
v = 1.0
v = µt
µ = mutation rate
t = time
ximum Likelihood
Two trees each consisting of single branch
Jukes-Cantor model
I (A) II (C)
I (A) II (C)
v = 0.1
v = 1.0
Ι AACC
ΙΙ CACT
1 j
N
1 C G G A C A C G T T T A C
2 C A G A C A C C T C T A C
3 C G G A T A A G T T A A C
4 C G G A T A G C C T A G C
1
42
3
1
C
2
C
4
G
3
A
5
6
L(j) = p
C C A G
A
A
C C A G
C
A
C C A G
T
T
+ p + … + p
L(j) = p
C C A G
A
A
C C A G
C
A
C C A G
T
T
+ p + … + p
N
L = L(1) • L(2) • … L(N) = ΠL(j)j = 1
N
lnL = lnL(1) + lnL(2) + … L(N) = Σ lnL(j)j = 1
Likelihood of the alignment at various branch lengths
0
0,00002
0,00004
0,00006
0,00008
0,0001
0,00012
0,00014
0,00016
0,00018
0,0002
0 0,1 0,2 0,3 0,4 0,5 0,6
Strengths of ML
• Does not try to make an observation of sequence change and then a
correction for superimposed substitutions. There is no need to
‘correct’ for anything, the models take care of superimposed
substitutions.
• Accurate branch lengths.
• Each site has a likelihood.
• If the model is correct, we should retrieve the correct tree (If we have
long-enough sequences and a sophisticated-enough model).
• You can use a model that fits the data.
• ML uses all the data (no selection of sites based on informativeness,
all sites are informative).
• ML can not only tell you about the phylogeny of the sequences, but
also the process of evolution that led to the observations of today’s
sequences.
Weaknesses of ML
• Can be inconsistent if we use models that are not accurate.
• Model might not be sophisticated enough
• Very computationally-intensive. Might not be possible to
examine all models (substitution matrices, tree topologies).
Models
• You can use models that:
Deal with different transition/transversion ratios.
Deal with unequal base composition.
Deal with heterogeneity of rates across sites.
Deal with heterogeneity of the substitution process (different rates
across lineages, different rates at different parts of the tree).
• The more free parameters, the better your model fits your data (good).
• The more free parameters, the higher the variance of the estimate (bad).
Choosing a Model
Don’t assume a model, rather find a model that fits your data.
Models often have “free” parameters. These can be fixed to a
reasonable value, or estimated by ML.
The more free parameters, the better the fit (higher the likelihood) of
the model to the data. (Good!)
The more free parameters, the higher the variance, and the less
power to discriminate among competing hypotheses. (Bad!)
We do not want to over-fit the model to the data
What is the best way to fit a line (a model) through these points?
How to tell if adding (or removing) a certain parameter is a good idea?
• Use statistics
• The null hypothesis is that the presence or absence of the parameter makes no difference
• In order to assess signifcance you need a null distribution
We have some DNA data, and a tree. Evaluate the data with 3 different
models.
model ln likelihood ∆
JC -2348.68
K2P -2256.73 91.95
GTR -2254.94 1.79
Evaluations with more complex models have higher likelihoods
The K2P model has 1 more parameter than the JC model
The GTR model has 4 more parameters than the K2P model
Are the extra parameters worth adding?
JC vs K2P K2P vs GTR
We have generated many true null hypothesis data sets and evaluated them under the JC
model and the K2P model. 95% of the differences are under 2.The statistic for our original
data set was 91.95, and so it is highly significant. In this case it is worthwhile to add the extra
parameter (tRatio).
We have generated many true null hypothesis data sets and evaluated them under the K2P
model and the GTR model. The statistic for our original data set was 1.79, and so it is not
signifcant. In this case it is not worthwhile to add the extra parameters.
You can use the χ2
approximation to assess
significance of adding parameters
Bayesian Inference
Maximum likelihood
Search for tree that maximizes the chance of
seeing the data (P (Data | Tree))
Bayesian Inference
Search for tree that maximizes the chance of
seeing the tree given the data (P (Tree | Data))
Bayesian Phylogenetics
Maximize the posterior probability of a tree given the aligned DNA
sequences
Two steps
- Definition of the posterior probabilities of trees (Bayes’ Rule)
- Approximation of the posterior probabilities of trees
Markov chain Monte Carlo (MCMC) methods
90 10
yesian Inference
yesian Inference
Data mining maximumlikelihood
Markov Chain Monte Carlo Methods
Posterior probabilities of trees are complex joint probabilities
that cannot be calculated analytically.
Instead, the posterior probabilities of trees are approximated
with Markov Chain Monte Carlo (MCMC) methods that sample
trees from their posterior probability distribution.
MCMC
A way of sampling / touring a set of solutions,biased
by their likelihood
1 Make a random solution N1 the current solution
2 Pick another solution N2
3 If Likelihood (N1 < N2) then replace N1 with N2
4 Else if Random (Likelihood (N2) / Likelihood (N1)) then replace
N1 with N2
5 Sample (record) the current solution
6 Repeat from step 2
Data mining maximumlikelihood
Data mining maximumlikelihood
Data mining maximumlikelihood
Data mining maximumlikelihood
Data mining maximumlikelihood
Data mining maximumlikelihood
Data mining maximumlikelihood
Data mining maximumlikelihood
yesian Inference
yesian Inference

More Related Content

PDF
Avoid Overfitting with Regularization
PDF
Data Science - Part III - EDA & Model Selection
PDF
Haoying1999
PPT
Quicksort
PDF
Logistic regression
PDF
A Preference Model on Adaptive Affinity Propagation
Avoid Overfitting with Regularization
Data Science - Part III - EDA & Model Selection
Haoying1999
Quicksort
Logistic regression
A Preference Model on Adaptive Affinity Propagation

What's hot (6)

PDF
A Method for Constructing Non-Isosceles Triangular Fuzzy Numbers Using Freque...
PDF
Random Forest / Bootstrap Aggregation
PPT
Quicksort
PDF
Summary statistics
PPTX
Quicksort algorithm
PDF
Simplicial closure and higher-order link prediction --- SIAMNS18
A Method for Constructing Non-Isosceles Triangular Fuzzy Numbers Using Freque...
Random Forest / Bootstrap Aggregation
Quicksort
Summary statistics
Quicksort algorithm
Simplicial closure and higher-order link prediction --- SIAMNS18
Ad

Viewers also liked (13)

PPT
Presentation Travel CRM (English)
PPTX
Network
PDF
The applications of blockchain and crypto currencies
PPTX
Alex and dela
PPT
Exception
PDF
Babasaheb - Rahuri (1)
PPTX
Qué es jdbc
PPTX
S I T E & I T S E N V I R O N M E N T S
PPT
Stacks queues lists
PPTX
Digital Marketing Guide for Hotels
PPTX
Optimizing shared caches in chip multiprocessors
PPT
Abstract class
PPTX
Extending burp with python
Presentation Travel CRM (English)
Network
The applications of blockchain and crypto currencies
Alex and dela
Exception
Babasaheb - Rahuri (1)
Qué es jdbc
S I T E & I T S E N V I R O N M E N T S
Stacks queues lists
Digital Marketing Guide for Hotels
Optimizing shared caches in chip multiprocessors
Abstract class
Extending burp with python
Ad

Similar to Data mining maximumlikelihood (20)

PPTX
Into to prob_prog_hari
PPTX
MyStataLab Assignment Help
PPT
Intro to Model Selection
PPTX
Ders 1 mean mod media st dev.pptx
PPTX
Probability distribution Function & Decision Trees in machine learning
PDF
Explore ml day 2
DOCX
Maximum likelihood estimation from uncertain
PDF
Cost Optimized Design Technique for Pseudo-Random Numbers in Cellular Automata
PPTX
Data Science Interview Questions | Data Science Interview Questions And Answe...
PPTX
03 Data Mining Techniques
PDF
Kaggle digits analysis_final_fc
PDF
ML MODULE 2.pdf
PPTX
Machine Learning with Python unit-2.pptx
PPTX
CPSC 531: System Modeling and Simulation.pptx
PPTX
Into to prob_prog_hari (2)
PPTX
Data mining Part 1
PDF
report
PDF
Data Science - Part V - Decision Trees & Random Forests
PPT
Other classification methods in data mining
PDF
Bel ventutorial hetero
Into to prob_prog_hari
MyStataLab Assignment Help
Intro to Model Selection
Ders 1 mean mod media st dev.pptx
Probability distribution Function & Decision Trees in machine learning
Explore ml day 2
Maximum likelihood estimation from uncertain
Cost Optimized Design Technique for Pseudo-Random Numbers in Cellular Automata
Data Science Interview Questions | Data Science Interview Questions And Answe...
03 Data Mining Techniques
Kaggle digits analysis_final_fc
ML MODULE 2.pdf
Machine Learning with Python unit-2.pptx
CPSC 531: System Modeling and Simulation.pptx
Into to prob_prog_hari (2)
Data mining Part 1
report
Data Science - Part V - Decision Trees & Random Forests
Other classification methods in data mining
Bel ventutorial hetero

More from Luis Goldster (20)

PPTX
Ruby on rails evaluation
PPTX
Design patterns
PPT
Lisp and scheme i
PPT
Ado.net &amp; data persistence frameworks
PPTX
Multithreading models.ppt
PPTX
Business analytics and data mining
PPTX
Big picture of data mining
PPTX
Data mining and knowledge discovery
PPTX
Cache recap
PPTX
Directory based cache coherence
PPTX
Hardware managed cache
PPTX
How analysis services caching works
PPT
Abstract data types
PPTX
Optimizing shared caches in chip multiprocessors
PPTX
Api crash
PPTX
Object model
PPTX
Abstraction file
PPTX
Object oriented analysis
PPT
Abstract class
PPTX
Concurrency with java
Ruby on rails evaluation
Design patterns
Lisp and scheme i
Ado.net &amp; data persistence frameworks
Multithreading models.ppt
Business analytics and data mining
Big picture of data mining
Data mining and knowledge discovery
Cache recap
Directory based cache coherence
Hardware managed cache
How analysis services caching works
Abstract data types
Optimizing shared caches in chip multiprocessors
Api crash
Object model
Abstraction file
Object oriented analysis
Abstract class
Concurrency with java

Recently uploaded (20)

PDF
Building High-Performance Oracle Teams: Strategic Staffing for Database Manag...
PDF
NewMind AI Weekly Chronicles - August'25 Week I
PDF
How Onsite IT Support Drives Business Efficiency, Security, and Growth.pdf
PPTX
CroxyProxy Instagram Access id login.pptx
PPTX
PA Analog/Digital System: The Backbone of Modern Surveillance and Communication
PDF
madgavkar20181017ppt McKinsey Presentation.pdf
PPTX
ABU RAUP TUGAS TIK kelas 8 hjhgjhgg.pptx
PDF
Reimagining Insurance: Connected Data for Confident Decisions.pdf
PDF
CIFDAQ's Teaching Thursday: Moving Averages Made Simple
PDF
Software Development Methodologies in 2025
PDF
Google’s NotebookLM Unveils Video Overviews
PDF
Shreyas Phanse Resume: Experienced Backend Engineer | Java • Spring Boot • Ka...
PPTX
Comunidade Salesforce São Paulo - Desmistificando o Omnistudio (Vlocity)
PDF
NewMind AI Weekly Chronicles - July'25 - Week IV
PPTX
Understanding_Digital_Forensics_Presentation.pptx
PDF
This slide provides an overview Technology
PDF
SparkLabs Primer on Artificial Intelligence 2025
PDF
A Day in the Life of Location Data - Turning Where into How.pdf
PPTX
breach-and-attack-simulation-cybersecurity-india-chennai-defenderrabbit-2025....
PDF
Security features in Dell, HP, and Lenovo PC systems: A research-based compar...
Building High-Performance Oracle Teams: Strategic Staffing for Database Manag...
NewMind AI Weekly Chronicles - August'25 Week I
How Onsite IT Support Drives Business Efficiency, Security, and Growth.pdf
CroxyProxy Instagram Access id login.pptx
PA Analog/Digital System: The Backbone of Modern Surveillance and Communication
madgavkar20181017ppt McKinsey Presentation.pdf
ABU RAUP TUGAS TIK kelas 8 hjhgjhgg.pptx
Reimagining Insurance: Connected Data for Confident Decisions.pdf
CIFDAQ's Teaching Thursday: Moving Averages Made Simple
Software Development Methodologies in 2025
Google’s NotebookLM Unveils Video Overviews
Shreyas Phanse Resume: Experienced Backend Engineer | Java • Spring Boot • Ka...
Comunidade Salesforce São Paulo - Desmistificando o Omnistudio (Vlocity)
NewMind AI Weekly Chronicles - July'25 - Week IV
Understanding_Digital_Forensics_Presentation.pptx
This slide provides an overview Technology
SparkLabs Primer on Artificial Intelligence 2025
A Day in the Life of Location Data - Turning Where into How.pdf
breach-and-attack-simulation-cybersecurity-india-chennai-defenderrabbit-2025....
Security features in Dell, HP, and Lenovo PC systems: A research-based compar...

Data mining maximumlikelihood

  • 2. Likelihood The likelihood is the probability of the data given the model.
  • 3. If we flip a coin and get a head and we think the coin is unbiased, then the probability of observing this head is 0.5. If we think the coin is biased so that we expect to get a head 80% of the time, then the likelihood of observing this datum (a head) is 0.8. The likelihood of making some observation is entirely dependent on the model that underlies our assumption. The datum has not changed, our model has. Therefore under the new model the likelihood of observing the datum has changed. Likelihood
  • 4. Maximum Likelihood (ML) ML assumes a explicit model of sequence evolution. This is justifiable, since molecular sequence data can be shown to have arisen according to a stochastic process. ML attempts to answer the question: What is the probability that I would observe these data (a multiple sequence alignment) given a particular model of evolution (a tree and a process)?
  • 5. Likelihood calculations In molecular phylogenetics, the data are an alignment of sequences We optimize parameters and branch lengths to get the maximum likelihood Each site has a likelihood The total likelihood is the product of the site likelihoods The maximum likelihood tree is the tree topology that gives the highest (optimized) likelihood under the given model. We use reversible models, so the position of the root does not matter.
  • 6. What is the probability of observing a G nucleotide? If we have a DNA sequence of 1 nucleotide in length and the identity of this nucleotide is G, what is the likelihood that we would observe this G? In the same way as the coin-flipping observation, the likelihood of observing this G is dependent on the model of sequence evolution that is thought to underlie the data. Model 1: frequency of G = 0.4 => likelihood(G) = 0.4 Model 2: frequency of G = 0.1 => likelihood(G) = 0.1 Model 3: frequency of G = 0.25 => likelihood(G) = 0.25
  • 7. What about longer sequences? If we consider a gene of length 2 gene 1 GA The the probability of observing this gene is the product of the probabilities of observing each character Model frequency of G = 0.4 frequencyof A= 0.15 p(G) = 0.4 p(A) =0.15 Likelihood (GA) = 0.4 x 0.15 = 0.06
  • 8. …or even longer sequences? gene 1 GACTAGCTAGACAGATACGAATTAC Model simple base frequency model p(A)=0.15; p(C)=0.2; p(G)=0.4; p(T)=0.25; (the sum of all probabilities must equal 1) Likelihood (gene 1) = 0.000000000000000018452813
  • 9. Note about models You might notice that our model of base frequency is not the optimal model for our observed data. If we had used the following model p(A)=0.4; p(C) =0.2; p(G)= 0.2; p(T) = 0.2; The likelihood of observing the gene is L (gene 1) = 0.000000000000335544320000 L (gene 1) = 0.000000000000000018452813 The datum has not changed, our model has. Therefore under the new model the likelihood of observing the datum has changed.
  • 10. Increase in model sophistication It is no longer possible to simply invoke a model that encompasses base composition, we must also include the mechanism of sequence change and stasis. There are two parts to this model - the tree and the process (the latter is confusingly referred to as the model, although both parts really compose the model).
  • 11. Different Branch Lengths For very short branch lengths, the probability of a character staying the same is high and the probability of it changing is low. For longer branch lengths, the probability of character change becomes higher and the probability of staying the same is lower. The previous calculations are based on the assumption that the branch length describes one Certain Evolutionary Distance or CED. If we want to consider a branch length that is twice as long (2 CED), then we can multiply the substitution matrix by itself (matrix2 ).
  • 12. I (A) II (C) I (A) II (C) v = 0.1 v = 1.0 v = µt µ = mutation rate t = time ximum Likelihood Two trees each consisting of single branch
  • 13. Jukes-Cantor model I (A) II (C) I (A) II (C) v = 0.1 v = 1.0
  • 15. 1 j N 1 C G G A C A C G T T T A C 2 C A G A C A C C T C T A C 3 C G G A T A A G T T A A C 4 C G G A T A G C C T A G C 1 42 3 1 C 2 C 4 G 3 A 5 6 L(j) = p C C A G A A C C A G C A C C A G T T + p + … + p
  • 16. L(j) = p C C A G A A C C A G C A C C A G T T + p + … + p N L = L(1) • L(2) • … L(N) = ΠL(j)j = 1 N lnL = lnL(1) + lnL(2) + … L(N) = Σ lnL(j)j = 1
  • 17. Likelihood of the alignment at various branch lengths 0 0,00002 0,00004 0,00006 0,00008 0,0001 0,00012 0,00014 0,00016 0,00018 0,0002 0 0,1 0,2 0,3 0,4 0,5 0,6
  • 18. Strengths of ML • Does not try to make an observation of sequence change and then a correction for superimposed substitutions. There is no need to ‘correct’ for anything, the models take care of superimposed substitutions. • Accurate branch lengths. • Each site has a likelihood. • If the model is correct, we should retrieve the correct tree (If we have long-enough sequences and a sophisticated-enough model). • You can use a model that fits the data. • ML uses all the data (no selection of sites based on informativeness, all sites are informative). • ML can not only tell you about the phylogeny of the sequences, but also the process of evolution that led to the observations of today’s sequences.
  • 19. Weaknesses of ML • Can be inconsistent if we use models that are not accurate. • Model might not be sophisticated enough • Very computationally-intensive. Might not be possible to examine all models (substitution matrices, tree topologies).
  • 20. Models • You can use models that: Deal with different transition/transversion ratios. Deal with unequal base composition. Deal with heterogeneity of rates across sites. Deal with heterogeneity of the substitution process (different rates across lineages, different rates at different parts of the tree). • The more free parameters, the better your model fits your data (good). • The more free parameters, the higher the variance of the estimate (bad).
  • 21. Choosing a Model Don’t assume a model, rather find a model that fits your data. Models often have “free” parameters. These can be fixed to a reasonable value, or estimated by ML. The more free parameters, the better the fit (higher the likelihood) of the model to the data. (Good!) The more free parameters, the higher the variance, and the less power to discriminate among competing hypotheses. (Bad!) We do not want to over-fit the model to the data
  • 22. What is the best way to fit a line (a model) through these points? How to tell if adding (or removing) a certain parameter is a good idea? • Use statistics • The null hypothesis is that the presence or absence of the parameter makes no difference • In order to assess signifcance you need a null distribution
  • 23. We have some DNA data, and a tree. Evaluate the data with 3 different models. model ln likelihood ∆ JC -2348.68 K2P -2256.73 91.95 GTR -2254.94 1.79 Evaluations with more complex models have higher likelihoods The K2P model has 1 more parameter than the JC model The GTR model has 4 more parameters than the K2P model Are the extra parameters worth adding?
  • 24. JC vs K2P K2P vs GTR We have generated many true null hypothesis data sets and evaluated them under the JC model and the K2P model. 95% of the differences are under 2.The statistic for our original data set was 91.95, and so it is highly significant. In this case it is worthwhile to add the extra parameter (tRatio). We have generated many true null hypothesis data sets and evaluated them under the K2P model and the GTR model. The statistic for our original data set was 1.79, and so it is not signifcant. In this case it is not worthwhile to add the extra parameters. You can use the χ2 approximation to assess significance of adding parameters
  • 26. Maximum likelihood Search for tree that maximizes the chance of seeing the data (P (Data | Tree)) Bayesian Inference Search for tree that maximizes the chance of seeing the tree given the data (P (Tree | Data))
  • 27. Bayesian Phylogenetics Maximize the posterior probability of a tree given the aligned DNA sequences Two steps - Definition of the posterior probabilities of trees (Bayes’ Rule) - Approximation of the posterior probabilities of trees Markov chain Monte Carlo (MCMC) methods
  • 31. Markov Chain Monte Carlo Methods Posterior probabilities of trees are complex joint probabilities that cannot be calculated analytically. Instead, the posterior probabilities of trees are approximated with Markov Chain Monte Carlo (MCMC) methods that sample trees from their posterior probability distribution.
  • 32. MCMC A way of sampling / touring a set of solutions,biased by their likelihood 1 Make a random solution N1 the current solution 2 Pick another solution N2 3 If Likelihood (N1 < N2) then replace N1 with N2 4 Else if Random (Likelihood (N2) / Likelihood (N1)) then replace N1 with N2 5 Sample (record) the current solution 6 Repeat from step 2

Editor's Notes

  • #11: Even though we tend to refer to the tree and the model separately, they are in fact both parts of the model.