Abstract
Even with a finite set of training samples, the performance of a Bayesian classifier can not be degraded by increasing the number of features, as long as the old features are recoverable from the new features. This is true even for the general Bayesian classifiers investigated by qq Hughes, a result which contradicts previous interpretations of Hughes' model. The reasons for these difficulties are discussed. It would appear that the peaking behavior of practical classifiers is caused principally by their nonoptimal use of the features.

This publication has 5 references indexed in Scilit: