Optimizing decision trees using multi-objective particle swarm optimization

23Citations
Citations of this article
17Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Although conceptually quite simple, decision trees are still among the most popular classifiers applied to real-world problems. Their popularity is due to a number of factors -core among these is their ease of comprehension, robust performance and fast data processing capabilities. Additionally feature selection is implicit within the decision tree structure. This chapter introduces the basic ideas behind decision trees, focusing on decision trees which only consider a rule relating to a single feature at a node (therefore making recursive axis-parallel slices in feature space to form their classification boundaries). The use of particle swarm optimization (PSO) to train near optimal decision trees is discussed, and PSO is applied both in a single objective formulation (minimizing misclassification cost), and multi-objective formulation (trading off misclassification rates across classes). Empirical results are presented on popular classification data sets from the well-known UCI machine learning repository, and PSO is demonstrated as being fully capable of acting as an optimizer for trees on these problems. Results additionally support the argument that multi-objectification of a problem can improve uni-objective search in classification problems. © 2009 Springer-Verlag Berlin Heidelberg.

Cite

CITATION STYLE

APA

Fieldsend, J. E. (2009). Optimizing decision trees using multi-objective particle swarm optimization. Studies in Computational Intelligence, 242, 93–114. https://doi.org/10.1007/978-3-642-03625-5_5

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free