ArnetMiner: Extraction and mining of academic social networks
Jie Tang, Jing Zhang, et al.
KDD 2008
We propose a family of novel cost-sensitive boosting methods for multi-class classification by applying the theory of gradient boosting to p-norm based cost functionals. We establish theoretical guarantees including proof of convergence and convergence rates for the proposed methods. Our theoretical treatment provides interpretations for some of the existing algorithms in terms of the proposed family, including a generalization of the costing algorithm, DSE and GBSE-t, and the Average Cost method. We also experimentally evaluate the performance of our new algorithms against existing methods of cost sensitive boosting, including AdaCost, CSB2, and AdaBoost.M2 with cost-sensitive weight initialization. We show that our proposed scheme generally achieves superior results in terms of cost minimization and, with the use of higher order p-norm loss in certain cases, consistently outperforms the comparison methods, thus establishing its empirical advantage. Copyright 2008 ACM.
Jie Tang, Jing Zhang, et al.
KDD 2008
Aurélie C. Lozano, Naoki Abe, et al.
KDD 2009
Edwin Pednault, Naoki Abe, et al.
KDD 2002
Andrew Arnold, Yan Liu, et al.
KDD 2007