Random Prism: a noise-tolerant alternative to Random Forests

Random Prism: a noise-tolerant alternative to Random Forests

0.00 Avg rating0 Votes
Article ID: iaor201522571
Volume: 31
Issue: 5
Start Page Number: 411
End Page Number: 420
Publication Date: Nov 2014
Journal: Expert Systems
Authors: ,
Keywords: learning
Abstract:

Ensemble learning can be used to increase the overall classification accuracy of a classifier by generating multiple base classifiers and combining their classification results. A frequently used family of base classifiers for ensemble learning are decision trees. However, alternative approaches can potentially be used, such as the Prism family of algorithms that also induces classification rules. Compared with decision trees, Prism algorithms generate modular classification rules that cannot necessarily be represented in the form of a decision tree. Prism algorithms produce a similar classification accuracy compared with decision trees. However, in some cases, for example, if there is noise in the training and test data, Prism algorithms can outperform decision trees by achieving a higher classification accuracy. However, Prism still tends to overfit on noisy data; hence, ensemble learners have been adopted in this work to reduce the overfitting. This paper describes the development of an ensemble learner using a member of the Prism family as the base classifier to reduce the overfitting of Prism algorithms on noisy datasets. The developed ensemble classifier is compared with a stand‐alone Prism classifier in terms of classification accuracy and resistance to noise.

Reviews

Required fields are marked *. Your email address will not be published.