Diverse classifier ensemble creation based on heuristic dataset modification |
| |
Authors: | Hamid Jamalinia Saber Khalouei Vahideh Rezaie Samad Nejatian Karamolah Bagheri-Fard |
| |
Affiliation: | 1. Department of Computer Engineering, Islamic Azad University, Nourabad Mamasani, Iran;2. Department of Computer Engineering, Islamic Azad University, Yasooj, Iran;3. Department of Mathematics, Islamic Azad University, Yasooj, Iran;4. Young Researchers and Elite Club, Islamic Azad University, Yasooj, Iran;5. Young Researchers and Elite Club, Islamic Azad University, Yasooj, Iran;6. Department of Electrical Engineering, Islamic Azad University, Yasooj, Iran |
| |
Abstract: | Bagging and Boosting are two main ensemble approaches consolidating the decisions of several hypotheses. The diversity of the ensemble members is considered to be a significant element to obtain generalization error. Here, an inventive method called EBAGTS (ensemble-based artificially generated training samples) is proposed to generate ensembles. It manipulates training examples in three ways in order to build various hypotheses straightforwardly: drawing a sub-sample from training set, reducing/raising error-prone training instances, and reducing/raising local instances around error-prone regions. The proposed method is a straightforward, generic framework utilizing any base classifier as its ensemble members to assemble a powerfully built combinational classifier. Decision-tree classifier and multilayer perceptron classifier as some basic classifiers have been employed in the experiments to indicate the proposed method accomplish higher predictive accuracy compared to meta-learning algorithms like Boosting and Bagging. Furthermore, EBAGTS outperforms Boosting more impressively as the training data set gets broader. It is illustrated that EBAGTS can fulfill better performance comparing to the state of the art. |
| |
Keywords: | Classification combining classifier diversity artificially created data |
|
|