A Dendritic Neuron Model with Adaptive Synapses Trained by Differential Evolution Algorithm

A dendritic neuron model with adaptive synapses (DMASs) based on differential evolution (DE) algorithm training is proposed. According to the signal transmission order, a DNM can be divided into four parts: the synaptic layer, dendritic layer, membrane layer, and somatic cell layer. It can be conver...

Full description

Bibliographic Details
Main Authors: Zhe Wang, Shangce Gao, Jiaxin Wang, Haichuan Yang, Yuki Todo
Format: Article
Language:English
Published: Hindawi Limited 2020-01-01
Series:Computational Intelligence and Neuroscience
Online Access:http://dx.doi.org/10.1155/2020/2710561
id doaj-8e4069081f1447b4a38d45af7a5aa0d6
record_format Article
spelling doaj-8e4069081f1447b4a38d45af7a5aa0d62020-11-25T01:19:54ZengHindawi LimitedComputational Intelligence and Neuroscience1687-52651687-52732020-01-01202010.1155/2020/27105612710561A Dendritic Neuron Model with Adaptive Synapses Trained by Differential Evolution AlgorithmZhe Wang0Shangce Gao1Jiaxin Wang2Haichuan Yang3Yuki Todo4Faculty of Engineering, University of Toyama, Toyama-Shi 930-8555, JapanFaculty of Engineering, University of Toyama, Toyama-Shi 930-8555, JapanFaculty of Engineering, University of Toyama, Toyama-Shi 930-8555, JapanFaculty of Engineering, University of Toyama, Toyama-Shi 930-8555, JapanSchool of Electrical and Computer Engineering, Kanazawa University, Kanazawa-Shi 920-1192, JapanA dendritic neuron model with adaptive synapses (DMASs) based on differential evolution (DE) algorithm training is proposed. According to the signal transmission order, a DNM can be divided into four parts: the synaptic layer, dendritic layer, membrane layer, and somatic cell layer. It can be converted to a logic circuit that is easily implemented on hardware by removing useless synapses and dendrites after training. This logic circuit can be designed to solve complex nonlinear problems using only four basic logical devices: comparators, AND (conjunction), OR (disjunction), and NOT (negation). To obtain a faster and better solution, we adopt the most popular DE for DMAS training. We have chosen five classification datasets from the UCI Machine Learning Repository for an experiment. We analyze and discuss the experimental results in terms of the correct rate, convergence rate, ROC curve, and the cross-validation and then compare the results with a dendritic neuron model trained by the backpropagation algorithm (BP-DNM) and a neural network trained by the backpropagation algorithm (BPNN). The analysis results show that the DE-DMAS shows better performance in all aspects.http://dx.doi.org/10.1155/2020/2710561
collection DOAJ
language English
format Article
sources DOAJ
author Zhe Wang
Shangce Gao
Jiaxin Wang
Haichuan Yang
Yuki Todo
spellingShingle Zhe Wang
Shangce Gao
Jiaxin Wang
Haichuan Yang
Yuki Todo
A Dendritic Neuron Model with Adaptive Synapses Trained by Differential Evolution Algorithm
Computational Intelligence and Neuroscience
author_facet Zhe Wang
Shangce Gao
Jiaxin Wang
Haichuan Yang
Yuki Todo
author_sort Zhe Wang
title A Dendritic Neuron Model with Adaptive Synapses Trained by Differential Evolution Algorithm
title_short A Dendritic Neuron Model with Adaptive Synapses Trained by Differential Evolution Algorithm
title_full A Dendritic Neuron Model with Adaptive Synapses Trained by Differential Evolution Algorithm
title_fullStr A Dendritic Neuron Model with Adaptive Synapses Trained by Differential Evolution Algorithm
title_full_unstemmed A Dendritic Neuron Model with Adaptive Synapses Trained by Differential Evolution Algorithm
title_sort dendritic neuron model with adaptive synapses trained by differential evolution algorithm
publisher Hindawi Limited
series Computational Intelligence and Neuroscience
issn 1687-5265
1687-5273
publishDate 2020-01-01
description A dendritic neuron model with adaptive synapses (DMASs) based on differential evolution (DE) algorithm training is proposed. According to the signal transmission order, a DNM can be divided into four parts: the synaptic layer, dendritic layer, membrane layer, and somatic cell layer. It can be converted to a logic circuit that is easily implemented on hardware by removing useless synapses and dendrites after training. This logic circuit can be designed to solve complex nonlinear problems using only four basic logical devices: comparators, AND (conjunction), OR (disjunction), and NOT (negation). To obtain a faster and better solution, we adopt the most popular DE for DMAS training. We have chosen five classification datasets from the UCI Machine Learning Repository for an experiment. We analyze and discuss the experimental results in terms of the correct rate, convergence rate, ROC curve, and the cross-validation and then compare the results with a dendritic neuron model trained by the backpropagation algorithm (BP-DNM) and a neural network trained by the backpropagation algorithm (BPNN). The analysis results show that the DE-DMAS shows better performance in all aspects.
url http://dx.doi.org/10.1155/2020/2710561
work_keys_str_mv AT zhewang adendriticneuronmodelwithadaptivesynapsestrainedbydifferentialevolutionalgorithm
AT shangcegao adendriticneuronmodelwithadaptivesynapsestrainedbydifferentialevolutionalgorithm
AT jiaxinwang adendriticneuronmodelwithadaptivesynapsestrainedbydifferentialevolutionalgorithm
AT haichuanyang adendriticneuronmodelwithadaptivesynapsestrainedbydifferentialevolutionalgorithm
AT yukitodo adendriticneuronmodelwithadaptivesynapsestrainedbydifferentialevolutionalgorithm
AT zhewang dendriticneuronmodelwithadaptivesynapsestrainedbydifferentialevolutionalgorithm
AT shangcegao dendriticneuronmodelwithadaptivesynapsestrainedbydifferentialevolutionalgorithm
AT jiaxinwang dendriticneuronmodelwithadaptivesynapsestrainedbydifferentialevolutionalgorithm
AT haichuanyang dendriticneuronmodelwithadaptivesynapsestrainedbydifferentialevolutionalgorithm
AT yukitodo dendriticneuronmodelwithadaptivesynapsestrainedbydifferentialevolutionalgorithm
_version_ 1715796856768823296