Decision Tree Integration Using Dynamic Regions of Competence
A vital aspect of the Multiple Classifier Systems construction process is the base model integration. For example, the Random Forest approach used the majority voting rule to fuse the base classifiers obtained by bagging the training dataset. In this paper we propose the algorithm that uses partitio...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2020-10-01
|
Series: | Entropy |
Subjects: | |
Online Access: | https://www.mdpi.com/1099-4300/22/10/1129 |
id |
doaj-2ad7901cfe8741faa056c78feda7cccc |
---|---|
record_format |
Article |
spelling |
doaj-2ad7901cfe8741faa056c78feda7cccc2020-11-25T03:26:58ZengMDPI AGEntropy1099-43002020-10-01221129112910.3390/e22101129Decision Tree Integration Using Dynamic Regions of CompetenceJędrzej Biedrzycki0Robert Burduk1Department of Systems and Computer Networks, Wroclaw University of Science and Technology, 50-370 Wroclaw, PolandDepartment of Systems and Computer Networks, Wroclaw University of Science and Technology, 50-370 Wroclaw, PolandA vital aspect of the Multiple Classifier Systems construction process is the base model integration. For example, the Random Forest approach used the majority voting rule to fuse the base classifiers obtained by bagging the training dataset. In this paper we propose the algorithm that uses partitioning the feature space whose split is determined by the decision rules of each decision tree node which is the base classification model. After dividing the feature space, the centroid of each new subspace is determined. This centroids are used in order to determine the weights needed in the integration phase based on the weighted majority voting rule. The proposal was compared with other Multiple Classifier Systems approaches. The experiments regarding multiple open-source benchmarking datasets demonstrate the effectiveness of our method. To discuss the results of our experiments, we use micro and macro-average classification performance measures.https://www.mdpi.com/1099-4300/22/10/1129decision treerandom forestmajority votingclassifier ensembleclassifier integration |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Jędrzej Biedrzycki Robert Burduk |
spellingShingle |
Jędrzej Biedrzycki Robert Burduk Decision Tree Integration Using Dynamic Regions of Competence Entropy decision tree random forest majority voting classifier ensemble classifier integration |
author_facet |
Jędrzej Biedrzycki Robert Burduk |
author_sort |
Jędrzej Biedrzycki |
title |
Decision Tree Integration Using Dynamic Regions of Competence |
title_short |
Decision Tree Integration Using Dynamic Regions of Competence |
title_full |
Decision Tree Integration Using Dynamic Regions of Competence |
title_fullStr |
Decision Tree Integration Using Dynamic Regions of Competence |
title_full_unstemmed |
Decision Tree Integration Using Dynamic Regions of Competence |
title_sort |
decision tree integration using dynamic regions of competence |
publisher |
MDPI AG |
series |
Entropy |
issn |
1099-4300 |
publishDate |
2020-10-01 |
description |
A vital aspect of the Multiple Classifier Systems construction process is the base model integration. For example, the Random Forest approach used the majority voting rule to fuse the base classifiers obtained by bagging the training dataset. In this paper we propose the algorithm that uses partitioning the feature space whose split is determined by the decision rules of each decision tree node which is the base classification model. After dividing the feature space, the centroid of each new subspace is determined. This centroids are used in order to determine the weights needed in the integration phase based on the weighted majority voting rule. The proposal was compared with other Multiple Classifier Systems approaches. The experiments regarding multiple open-source benchmarking datasets demonstrate the effectiveness of our method. To discuss the results of our experiments, we use micro and macro-average classification performance measures. |
topic |
decision tree random forest majority voting classifier ensemble classifier integration |
url |
https://www.mdpi.com/1099-4300/22/10/1129 |
work_keys_str_mv |
AT jedrzejbiedrzycki decisiontreeintegrationusingdynamicregionsofcompetence AT robertburduk decisiontreeintegrationusingdynamicregionsofcompetence |
_version_ |
1724590169274712064 |