Abstract—Usage of more training data with label information gives more success for classification of datasets in machine learning. But in real life, obtaining data with label information is a cost-effective and long-lasting process. Herein, active learning algorithms are emerged. Active learning algorithms aim to maintain current success rate with fewer samples in train set or increase total success of model in training process. Active learning is not only functional for regular learning methods but also can be used in ensemble learning algorithms with specified techniques. In this study, two different active learning algorithms based on class probabilities of the samples are tested on five datasets classification. Ensemble learning methods are used as classification model. Comparative results presented as graphically and numerically.
Index Terms—Active learning, adaboost, bagging, decision tree, ensemble learning, machine learning.
The authors are with the Department of Computer Engineering, Yildiz Technical University, İstanbul, 34720 TR (e-mail: firstname.lastname@example.org, email@example.com).
Cite:Hamza Osman İlhan and Mehmet Fatih Amasyal, "Comparing Informative Sample Selection Strategies in Classification Ensembles," International Journal of Machine Learning and Computing vol.4, no. 1, pp. 79-84, 2014.