Title: Trimmed bagging
Authors: Croux, Christophe ×
Joossens, Kristel
Lemmens, Aurélie #
Issue Date: 2007
Publisher: North-Holland Pub. Co.
Series Title: Computational statistics & data analysis vol:52 issue:1 pages:362-368
Abstract: Bagging has been found to be successful in increasing the predictive performance of unstable classifiers. Bagging draws bootstrap samples from the training sample, applies the classifier to each bootstrap sample, and then averages over all obtained classification rules. The idea of
trimmed bagging is to exclude the bootstrapped classification rules that yield the highest error rates, as estimated by the out-of-bag error rate, and to aggregate over the remaining ones. In this note we explore the
potential benefits of trimmed bagging. On the basis of numerical experiments, we conclude that trimmed bagging performs comparably to standard bagging when applied to unstable classifiers as decision trees, but yields better results when applied to more stable base classifiers, like
support vector machines.
ISSN: 0167-9473
Publication status: published
KU Leuven publication type: IT
Appears in Collections:Research Center for Operations Research and Business Statistics (ORSTAT), Leuven
× corresponding author
# (joint) last author

Files in This Item:
File Description Status SizeFormat
trimmedbagging.pdfTrimmed bagging Published 125KbAdobe PDFView/Open


All items in Lirias are protected by copyright, with all rights reserved.

© Web of science