Professional Documents
Culture Documents
Machine Learning
UNIT-II
Syllabus
Bayes' theorem is also known as Bayes' Rule or Bayes' law, which is used to determine the
probability of a hypothesis with prior knowledge. It depends on the conditional probability.
Where,
• P(c|x) is Posterior probability: Probability of hypothesis ‘c’ on the observed event ‘x’.
• P(x|c) is Likelihood probability: Probability of the evidence given that the probability of a hypothesis is true.
Working of Bayes' Classifier can be understood with the help of the below example:
Applying Bayes’
theorem:
P(Yes|Sunny)= P(Sunny|Yes)*P(Yes)/P(Sunny) P(No|Sunny)= P(Sunny|No)*P(No)/P(Sunny)
• The extra pieces of knowledge that we do not have access to are named as the
unobservable variables.
• In the coin tossing example, the only observable variable is the outcome of the toss.
• Denoting the unobservable by Z and the observable as x.
𝑥=𝑓 𝑍
Where f(·) is the deterministic function that defines the outcome from the
unobservable pieces of knowledge
We cannot model the process in this way, we define the outcome X as a random
variable drawn from a probability distribution P(X = 𝑥 ) that specifies the
process.
Bayesian Decision Theory: Introduction
• Let us say X = 1 denotes that the outcome of a toss is heads and X = 0 denotes tails.
• Such X are Bernoulli-distributed where the parameter of the distribution 𝑝0 is
the probability that the outcome is heads:
• If we don’t know P(X) and want to estimate this from a given sample.
• We have a sample, X, containing examples drawn from the probability distribution
of the observables 𝑥 𝑡 , denoted as p(x).
• The aim is to build an approximator to it, 𝑝Ƹ 𝑥 , using the sample X.
• The sample contains the outcomes of the past N tosses. Then using X, we
can estimate 𝑝0, which is the parameter that uniquely specifies the
distribution.
• Our estimate of 𝑝0 is:
Bayesian Decision Theory: Introduction