Random prism: An alternative to random forests

13Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Ensemble learning techniques generate multiple classifiers, so called base classifiers, whose combined classification results are used in order to increase the overall classification accuracy. In most ensemble classifiers the base classifiers are based on the Top Down Induction of Decision Trees (TDIDT) approach. However, an alternative approach for the induction of rule based classifiers is the Prism family of algorithms. Prism algorithms produce modular classification rules that do not necessarily fit into a decision tree structure. Prism classification rulesets achieve a comparable and sometimes higher classification accuracy compared with decision tree classifiers, if the data is noisy and large. Yet Prism still suffers from overfitting on noisy and large datasets. In practice ensemble techniques tend to reduce the overfitting, however there exists no ensemble learner for modular classification rule inducers such as the Prism family of algorithms. This article describes the first development of an ensemble learner based on the Prism family of algorithms in order to enhance Prism's classification accuracy by reducing overfitting. © Springer-Verlag London Limited 2011.

Cite

CITATION STYLE

APA

Stahl, F., & Bramer, M. (2011). Random prism: An alternative to random forests. In Res. and Dev. in Intelligent Syst. XXVIII: Incorporating Applications and Innovations in Intel. Sys. XIX - AI 2011, 31st SGAI Int. Conf. on Innovative Techniques and Applications of Artificial Intel. (pp. 5–18). Springer London. https://doi.org/10.1007/978-1-4471-2318-7_1

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free