Systematic Boolean Satisfiability Programming in Radial Basis Function Neural Network

Radial Basis Function Neural Network (RBFNN) is a class of Artificial Neural Network (ANN) that contains hidden layer processing units (neurons) with nonlinear, radially symmetric activation functions. Consequently, RBFNN has extensively suffered from significant computational error and difficulties...

Full description

Bibliographic Details
Main Authors: Mohd. Asyraf Mansor, Siti Zulaikha Mohd Jamaludin, Mohd Shareduwan Mohd Kasihmuddin, Shehab Abdulhabib Alzaeemi, Md Faisal Md Basir, Saratha Sathasivam
Format: Article
Language:English
Published: MDPI AG 2020-02-01
Series:Processes
Subjects:
Online Access:https://www.mdpi.com/2227-9717/8/2/214
id doaj-7fd6c462c9404124b216fb8aecf9d943
record_format Article
spelling doaj-7fd6c462c9404124b216fb8aecf9d9432020-11-25T01:30:16ZengMDPI AGProcesses2227-97172020-02-018221410.3390/pr8020214pr8020214Systematic Boolean Satisfiability Programming in Radial Basis Function Neural NetworkMohd. Asyraf Mansor0Siti Zulaikha Mohd Jamaludin1Mohd Shareduwan Mohd Kasihmuddin2Shehab Abdulhabib Alzaeemi3Md Faisal Md Basir4Saratha Sathasivam5School of Distance Education, Universiti Sains Malaysia, Penang 11800 USM, MalaysiaSchool of Mathematical Sciences, Universiti Sains Malaysia, Penang 11800 USM, MalaysiaSchool of Mathematical Sciences, Universiti Sains Malaysia, Penang 11800 USM, MalaysiaSchool of Mathematical Sciences, Universiti Sains Malaysia, Penang 11800 USM, MalaysiaDepartment of Mathematical Sciences, Faculty of Science, Universiti Teknologi Malaysia, Johor Bahru 81310, MalaysiaSchool of Mathematical Sciences, Universiti Sains Malaysia, Penang 11800 USM, MalaysiaRadial Basis Function Neural Network (RBFNN) is a class of Artificial Neural Network (ANN) that contains hidden layer processing units (neurons) with nonlinear, radially symmetric activation functions. Consequently, RBFNN has extensively suffered from significant computational error and difficulties in approximating the optimal hidden neuron, especially when dealing with Boolean Satisfiability logical rule. In this paper, we present a comprehensive investigation of the potential effect of systematic Satisfiability programming as a logical rule, namely 2 Satisfiability (2SAT) to optimize the output weights and parameters in RBFNN. The 2SAT logical rule has extensively applied in various disciplines, ranging from industrial automation to the complex management system. The core impetus of this study is to investigate the effectiveness of 2SAT logical rule in reducing the computational burden for RBFNN by obtaining the parameters in RBFNN. The comparison is made between RBFNN and the existing method, based on the Hopfield Neural Network (HNN) in searching for the optimal neuron state by utilizing different numbers of neurons. The comparison was made with the HNN as a benchmark to validate the final output of our proposed RBFNN with 2SAT logical rule. Note that the final output in HNN is represented in terms of the quality of the final states produced at the end of the simulation. The simulation dynamic was carried out by using the simulated data, randomly generated by the program. In terms of 2SAT logical rule, simulation revealed that RBFNN has two advantages over HNN model: RBFNN can obtain the correct final neuron state with the lowest error and does not require any approximation for the number of hidden layers. Furthermore, this study provides a new paradigm in the field feed-forward neural network by implementing a more systematic propositional logic rule.https://www.mdpi.com/2227-9717/8/2/214radial basis function neural networkhopfield neural networksatisfiabilityoptimizationlogic programming
collection DOAJ
language English
format Article
sources DOAJ
author Mohd. Asyraf Mansor
Siti Zulaikha Mohd Jamaludin
Mohd Shareduwan Mohd Kasihmuddin
Shehab Abdulhabib Alzaeemi
Md Faisal Md Basir
Saratha Sathasivam
spellingShingle Mohd. Asyraf Mansor
Siti Zulaikha Mohd Jamaludin
Mohd Shareduwan Mohd Kasihmuddin
Shehab Abdulhabib Alzaeemi
Md Faisal Md Basir
Saratha Sathasivam
Systematic Boolean Satisfiability Programming in Radial Basis Function Neural Network
Processes
radial basis function neural network
hopfield neural network
satisfiability
optimization
logic programming
author_facet Mohd. Asyraf Mansor
Siti Zulaikha Mohd Jamaludin
Mohd Shareduwan Mohd Kasihmuddin
Shehab Abdulhabib Alzaeemi
Md Faisal Md Basir
Saratha Sathasivam
author_sort Mohd. Asyraf Mansor
title Systematic Boolean Satisfiability Programming in Radial Basis Function Neural Network
title_short Systematic Boolean Satisfiability Programming in Radial Basis Function Neural Network
title_full Systematic Boolean Satisfiability Programming in Radial Basis Function Neural Network
title_fullStr Systematic Boolean Satisfiability Programming in Radial Basis Function Neural Network
title_full_unstemmed Systematic Boolean Satisfiability Programming in Radial Basis Function Neural Network
title_sort systematic boolean satisfiability programming in radial basis function neural network
publisher MDPI AG
series Processes
issn 2227-9717
publishDate 2020-02-01
description Radial Basis Function Neural Network (RBFNN) is a class of Artificial Neural Network (ANN) that contains hidden layer processing units (neurons) with nonlinear, radially symmetric activation functions. Consequently, RBFNN has extensively suffered from significant computational error and difficulties in approximating the optimal hidden neuron, especially when dealing with Boolean Satisfiability logical rule. In this paper, we present a comprehensive investigation of the potential effect of systematic Satisfiability programming as a logical rule, namely 2 Satisfiability (2SAT) to optimize the output weights and parameters in RBFNN. The 2SAT logical rule has extensively applied in various disciplines, ranging from industrial automation to the complex management system. The core impetus of this study is to investigate the effectiveness of 2SAT logical rule in reducing the computational burden for RBFNN by obtaining the parameters in RBFNN. The comparison is made between RBFNN and the existing method, based on the Hopfield Neural Network (HNN) in searching for the optimal neuron state by utilizing different numbers of neurons. The comparison was made with the HNN as a benchmark to validate the final output of our proposed RBFNN with 2SAT logical rule. Note that the final output in HNN is represented in terms of the quality of the final states produced at the end of the simulation. The simulation dynamic was carried out by using the simulated data, randomly generated by the program. In terms of 2SAT logical rule, simulation revealed that RBFNN has two advantages over HNN model: RBFNN can obtain the correct final neuron state with the lowest error and does not require any approximation for the number of hidden layers. Furthermore, this study provides a new paradigm in the field feed-forward neural network by implementing a more systematic propositional logic rule.
topic radial basis function neural network
hopfield neural network
satisfiability
optimization
logic programming
url https://www.mdpi.com/2227-9717/8/2/214
work_keys_str_mv AT mohdasyrafmansor systematicbooleansatisfiabilityprogramminginradialbasisfunctionneuralnetwork
AT sitizulaikhamohdjamaludin systematicbooleansatisfiabilityprogramminginradialbasisfunctionneuralnetwork
AT mohdshareduwanmohdkasihmuddin systematicbooleansatisfiabilityprogramminginradialbasisfunctionneuralnetwork
AT shehababdulhabibalzaeemi systematicbooleansatisfiabilityprogramminginradialbasisfunctionneuralnetwork
AT mdfaisalmdbasir systematicbooleansatisfiabilityprogramminginradialbasisfunctionneuralnetwork
AT sarathasathasivam systematicbooleansatisfiabilityprogramminginradialbasisfunctionneuralnetwork
_version_ 1725092454748651520