Error entropy in classification problems: a univariate data analysis.

Neural Comput

Instituto de Engenharia Biomédica, Laboratório Sinal e Imagen Biomédica, 4200-465, Porto, Portugal.

Published: September 2006

Entropy-based cost functions are enjoying a growing attractiveness in unsupervised and supervised classification tasks. Better performances in terms both of error rate and speed of convergence have been reported. In this letter, we study the principle of error entropy minimization (EEM) from a theoretical point of view. We use Shannon's entropy and study univariate data splitting in two-class problems. In this setting, the error variable is a discrete random variable, leading to a not too complicated mathematical analysis of the error entropy. We start by showing that for uniformly distributed data, there is equivalence between the EEM split and the optimal classifier. In a more general setting, we prove the necessary conditions for this equivalence and show the existence of class configurations where the optimal classifier corresponds to maximum error entropy. The presented theoretical results provide practical guidelines that are illustrated with a set of experiments with both real and simulated data sets, where the effectiveness of EEM is compared with the usual mean square error minimization.

Download full-text PDF

Source
http://dx.doi.org/10.1162/neco.2006.18.9.2036DOI Listing

Publication Analysis

Top Keywords

error entropy
16
univariate data
8
optimal classifier
8
error
7
entropy classification
4
classification problems
4
problems univariate
4
data
4
data analysis
4
analysis entropy-based
4

Similar Publications

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!