Research on Machine Learning

Multi-class boosting with asymmetric binary weak-learners. We introduce a multi-class generalization of AdaBoost with binary weak-learners. We use a vectorial codification to represent class labels and a multi-class exponential loss function to evaluate classifier responses. This representation produces a set of margin values that provide a range of punishments for failures and rewards for successes. Moreover, the stage-wise optimization of this model introduces an asymmetric boosting procedure whose costs depend on the number of classes separated by each weak-learner. In this way the boosting procedure takes into account class imbalances when building the ensemble. The experiments performed compare this new approach favorably to AdaBoost.MH, GentleBoost and the SAMME algorithms.
Related Publications: PR'2014
Multiclass cost-sensitive boosting: BAdaCost. We introduce a cost-sensitive multi-class Boosting algorithm (BAdaCost) based on a generalization of the Boosting margin, termed multi-class cost-sensitive margin. It can be used, for example, to address the class imbalance by the introduction of a cost matrix that weighs more hevily the costs of confused classes.
Related Publications: IBPRIA'2015