Hothorn, T and Lausen, B (2005) Bundling classifiers by bagging trees. Computational Statistics & Data Analysis, 49 (4). pp. 1068-1078. DOI https://doi.org/10.1016/j.csda.2004.06.019
Hothorn, T and Lausen, B (2005) Bundling classifiers by bagging trees. Computational Statistics & Data Analysis, 49 (4). pp. 1068-1078. DOI https://doi.org/10.1016/j.csda.2004.06.019
Hothorn, T and Lausen, B (2005) Bundling classifiers by bagging trees. Computational Statistics & Data Analysis, 49 (4). pp. 1068-1078. DOI https://doi.org/10.1016/j.csda.2004.06.019
Abstract
The quest of selecting the best classifier for a discriminant analysis problem is often rather difficult. A combination of different types of classifiers promises to lead to improved predictive models compared to selecting one of the competitors. An additional learning sample, for example the out-of-bag sample, is used for the training of arbitrary classifiers. Classification trees are employed to bundle their predictions for the bootstrap sample. Consequently, a combined classifier is developed. Benchmark experiments show that the combined classifier is superior to any of the single classifiers in many applications.
Item Type: | Article |
---|---|
Uncontrolled Keywords: | Bagging; Ensemble-methods; Method selection; Error rate estimation |
Subjects: | H Social Sciences > HA Statistics |
Divisions: | Faculty of Science and Health Faculty of Science and Health > Mathematics, Statistics and Actuarial Science, School of |
SWORD Depositor: | Unnamed user with email elements@essex.ac.uk |
Depositing User: | Unnamed user with email elements@essex.ac.uk |
Date Deposited: | 31 May 2012 12:58 |
Last Modified: | 24 Oct 2024 17:57 |
URI: | http://repository.essex.ac.uk/id/eprint/2460 |