@inproceedings{9ee9763678be44189e561b511c501028,
title = "Subsampling the concurrent AdaBoost algorithm: An efficient approach for large datasets",
abstract = "In this work we propose a subsampled version of the Concurrent AdaBoost algorithm in order to deal with large datasets in an efficient way. The proposal is based on a concurrent computing approach focused on improving the distribution weight estimation in the algorithm, hence obtaining better capacity of generalization. On each round, we train in parallel several weak hypotheses, and using a weighted ensemble we update the distribution weights of the following boosting rounds. Instead of creating resamples of size equal to the original dataset, we subsample the datasets in order to obtain a speed-up in the training phase. We validate our proposal with different resampling sizes using 3 datasets, obtaining promising results and showing that the size of the resamples does not affect considerably the performance of the algorithm, but the execution time improves greatly.",
keywords = "Classification, Concurrent AdaBoost, Large data sets classification, Machine learning, Subsampling",
author = "H{\'e}ctor Allende-Cid and Diego Acu{\~n}a and H{\'e}ctor Allende",
note = "Publisher Copyright: {\textcopyright} Springer International Publishing AG 2017.; 21st Iberoamerican Congress on Pattern Recognition, CIARP 2016 ; Conference date: 08-11-2016 Through 11-11-2016",
year = "2017",
doi = "10.1007/978-3-319-52277-7_39",
language = "English",
isbn = "9783319522760",
series = "Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)",
publisher = "Springer Verlag",
pages = "318--325",
editor = "Cesar Beltran-Castanon and Fazel Famili and Ingela Nystrom",
booktitle = "Progress in Pattern Recognition, Image Analysis, Computer Vision, and Applications - 21st Iberoamerican Congress, CIARP 2016, Proceedings",
}