202412111029
Status: #idea
Tags: Generative Models
State: #nascient
Naive Bayes
It is a cousin of the Discriminant Analysis models such as Linear Discriminant Analysis (LDA) and Quadratic Discriminant Analysis (QDA).
While they attempt to solve the same problem, that is generating the posterior probability through estimation of the prior and likelihood of the classes, the Naive Bayes makes an assumption that is arguably stronger than even its cousins.
While Linear Discriminant Analysis (LDA) makes the assumptions of normality of
By that we mean that while we would expect that
In other words,
Plugging it into Bayes theorem as for the other models we get:
Though it's simplifying assumption, Naive Bayes increases the bias as we essentially don't have to fit covariance matrices, but decreases the variance. This means that we would expect it to come into its own in cases where
Naive Bayes success can be puzzling as after all, in most contexts, we'd expect the assumption to be entirely false.
Still it can be explained through the lens of the Bias-Variance Tradeoff, thanks to the assumptions it makes we require less parameters to fit at the cost of bias, but considering the fact that for a rigorous estimation of the likelihoods would require ungodly amounts of data which in most cases are not even available, this bias ends up smaller than the great variance that would arise if other methods were used simply because generally the lack of data prevents us from actually estimating the parameters.
As a result, even though the assumptions are really strong, it often gives pretty good results.
References
olio