Testing modified confusion entropy as split criterion for decision trees
View/ Open
Date
2019-05-29Author
Gonzalo de Sá, Alexander
Metadata
Show full item recordAbstract
In 2010, a new performance measure to evaluate the results obtained by algorithms of data classification was presented, Confusion Entropy (CEN). This render measure is able to achieve a greater discrimination than Accuracy focusing on the distribution across different classes of both correctly and wrongly classified instances, but it is not able to work correctly in cases of binary classification. Recently, an enhancement has been proposed to correct its behaviour in those cases, the Modified Confusion Entropy (MCEN).
In this work, we propose a new algorithm, MCENTree. This algorithm uses MCEN as splitting criterion to build a decision tree model instead of CEN, as proposed in the CENTree algorithm in the literature.
We make a comparison among a classic J48, CENTree and the new algorithm MCENTree in terms of Accuracy, CEN and MCEN performance measures, and we analyze how the undesired behaviour of CEN affects the results of the algorithms and how MCEN shows a good behaviour in terms of results: while MCENTree gives correct results in a statistical range [0,1], CENTree sometimes gives non monotonous and out of range results in binary class classification.