We show that the `naive Bayes' classifier which assumes independent covariates greatly outperforms the Fisher linear discriminant rule under broad conditions when the number of variables grows faster than the number of observations, in the classical problem of discriminating between two normal populations. We also introduce a class of rules spanning the range between independence and arbitrary dependence. These rules are shown to achieve Bayes consistency for the Gaussian `coloured noise' model and to adapt to a spectrum of convergence rates, which we conjecture to be minimax.
"Some theory for Fisher's linear discriminant function, `naive Bayes', and some alternatives when there are many more variables than observations." Bernoulli 10 (6) 989 - 1010, December 2004. https://doi.org/10.3150/bj/1106314847