Background
We want to find out the relation between two events. For example, when one thing happens, what is the probability that another thing happens. For this, we introduce the Bayesian Learning.
Bayesian Theorem
- Assumptions are mutually exclusive.
- H Hypothesis Space is fully exhaustive.
stands for one of the sample set for all the possible data. and are independent of each other. can be ignored when only comparing different assumptions.
likelihood - log likelihood
Max A Posterior (MAP)
From the equation, we can see that
Maximum Likelihood (ML)
If we are completely unaware of the probability distribution, or we know that all the assumptions happened are with equal probability, then MAP is equivalent to ML.
The relation between ML and Least Square Estimation (LSE)
Let’s express the training set as
In the LSE, we want to find an expression f to let
We assume that
So, we can get:
We can see that
Naive Bayes
Naive Bayesian Assumption
We assume that each
Naive Bayesian Classifier
If MAP’s attributes also satisfy independence, then