Weighted Scoring in Geometric Space for Decision Tree Ensemble

In order to improve the classification performance of a single classification model, Multiple Classifier Systems (MCS) are used. One of the most common techniques utilizing multiple decision trees is the random forest, where diversity between base classifiers is obtained by bagging the training data...

Full description

Bibliographic Details
Main Authors: Jedrzej Biedrzycki, Robert Burduk
Format: Article
Language:English
Published: IEEE 2020-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/9079594/
id doaj-26d0ba758c9a48c181568a7c3e51f3de
record_format Article
spelling doaj-26d0ba758c9a48c181568a7c3e51f3de2021-03-30T01:44:22ZengIEEEIEEE Access2169-35362020-01-018821008210710.1109/ACCESS.2020.29907219079594Weighted Scoring in Geometric Space for Decision Tree EnsembleJedrzej Biedrzycki0https://orcid.org/0000-0002-4924-1759Robert Burduk1https://orcid.org/0000-0002-3506-6611Faculty of Electronics, Wroclaw University of Science and Technology, Wroclaw, PolandFaculty of Electronics, Wroclaw University of Science and Technology, Wroclaw, PolandIn order to improve the classification performance of a single classification model, Multiple Classifier Systems (MCS) are used. One of the most common techniques utilizing multiple decision trees is the random forest, where diversity between base classifiers is obtained by bagging the training dataset. In this paper, we propose the algorithm that uses horizontal partitioning the learning set and uses decision trees as base models to obtain decision regions. In the proposed approach feature space is divided into disjoint subspace. Additionally, the location of the subspace centroids, as well as the size and location of decision regions, are used in order to determine the weights needed in the last process of creating MCS, i.e. in the integration phase. The proposed algorithm was evaluated employing multiple open-source benchmarking datasets, compared using accuracy and Matthews correlation coefficient performance measures with two existing MCS methods - random forest and majority voting. The statistical analysis confirms an improvement in recognition compared to the random forest. In addition, we proved that for infinitely dense space division proposed algorithm is equivalent to majority voting.https://ieeexplore.ieee.org/document/9079594/Decision treeensemble classifiermajority votingmultiple classifier systemrandom forest
collection DOAJ
language English
format Article
sources DOAJ
author Jedrzej Biedrzycki
Robert Burduk
spellingShingle Jedrzej Biedrzycki
Robert Burduk
Weighted Scoring in Geometric Space for Decision Tree Ensemble
IEEE Access
Decision tree
ensemble classifier
majority voting
multiple classifier system
random forest
author_facet Jedrzej Biedrzycki
Robert Burduk
author_sort Jedrzej Biedrzycki
title Weighted Scoring in Geometric Space for Decision Tree Ensemble
title_short Weighted Scoring in Geometric Space for Decision Tree Ensemble
title_full Weighted Scoring in Geometric Space for Decision Tree Ensemble
title_fullStr Weighted Scoring in Geometric Space for Decision Tree Ensemble
title_full_unstemmed Weighted Scoring in Geometric Space for Decision Tree Ensemble
title_sort weighted scoring in geometric space for decision tree ensemble
publisher IEEE
series IEEE Access
issn 2169-3536
publishDate 2020-01-01
description In order to improve the classification performance of a single classification model, Multiple Classifier Systems (MCS) are used. One of the most common techniques utilizing multiple decision trees is the random forest, where diversity between base classifiers is obtained by bagging the training dataset. In this paper, we propose the algorithm that uses horizontal partitioning the learning set and uses decision trees as base models to obtain decision regions. In the proposed approach feature space is divided into disjoint subspace. Additionally, the location of the subspace centroids, as well as the size and location of decision regions, are used in order to determine the weights needed in the last process of creating MCS, i.e. in the integration phase. The proposed algorithm was evaluated employing multiple open-source benchmarking datasets, compared using accuracy and Matthews correlation coefficient performance measures with two existing MCS methods - random forest and majority voting. The statistical analysis confirms an improvement in recognition compared to the random forest. In addition, we proved that for infinitely dense space division proposed algorithm is equivalent to majority voting.
topic Decision tree
ensemble classifier
majority voting
multiple classifier system
random forest
url https://ieeexplore.ieee.org/document/9079594/
work_keys_str_mv AT jedrzejbiedrzycki weightedscoringingeometricspacefordecisiontreeensemble
AT robertburduk weightedscoringingeometricspacefordecisiontreeensemble
_version_ 1724186412230639616