Convolutional Recurrent Deep Learning Model for Sentence Classification

As the amount of unstructured text data that humanity produces overall and on the Internet grows, so does the need to intelligently to process it and extract different types of knowledge from it. Convolutional neural networks (CNNs) and recurrent neural networks (RNNs) have been applied to natural l...

Full description

Bibliographic Details
Main Authors: Abdalraouf Hassan, Ausif Mahmood
Format: Article
Language:English
Published: IEEE 2018-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/8314136/
id doaj-d04f7698f9fa434193885af27bcd7839
record_format Article
spelling doaj-d04f7698f9fa434193885af27bcd78392021-03-29T21:00:25ZengIEEEIEEE Access2169-35362018-01-016139491395710.1109/ACCESS.2018.28148188314136Convolutional Recurrent Deep Learning Model for Sentence ClassificationAbdalraouf Hassan0https://orcid.org/0000-0002-1852-4449Ausif Mahmood1Department of Computer Science and Engineering, University of Bridgeport, Bridgeport, CT, USADepartment of Computer Science and Engineering, University of Bridgeport, Bridgeport, CT, USAAs the amount of unstructured text data that humanity produces overall and on the Internet grows, so does the need to intelligently to process it and extract different types of knowledge from it. Convolutional neural networks (CNNs) and recurrent neural networks (RNNs) have been applied to natural language processing systems with comparative, remarkable results. The CNN is a noble approach to extract higher level features that are invariant to local translation. However, it requires stacking multiple convolutional layers in order to capture long-term dependencies, due to the locality of the convolutional and pooling layers. In this paper, we describe a joint CNN and RNN framework to overcome this problem. Briefly, we use an unsupervised neural language model to train initial word embeddings that are further tuned by our deep learning network, then, the pre-trained parameters of the network are used to initialize the model. At a final stage, the proposed framework combines former information with a set of feature maps learned by a convolutional layer with long-term dependencies learned via long-short-term memory. Empirically, we show that our approach, with slight hyperparameter tuning and static vectors, achieves outstanding results on multiple sentiment analysis benchmarks. Our approach outperforms several existing approaches in term of accuracy; our results are also competitive with the state-of-the-art results on the Stanford Large Movie Review data set with 93.3% accuracy, and the Stanford Sentiment Treebank data set with 48.8% fine-grained and 89.2% binary accuracy, respectively. Our approach has a significant role in reducing the number of parameters and constructing the convolutional layer followed by the recurrent layer as a substitute for the pooling layer. Our results show that we were able to reduce the loss of detailed, local information and capture long-term dependencies with an efficient framework that has fewer parameters and a high level of performance.https://ieeexplore.ieee.org/document/8314136/Convolutional neural networkrecurrent neural networknatural language processingdeep learningsentiment analysislong-term dependencies
collection DOAJ
language English
format Article
sources DOAJ
author Abdalraouf Hassan
Ausif Mahmood
spellingShingle Abdalraouf Hassan
Ausif Mahmood
Convolutional Recurrent Deep Learning Model for Sentence Classification
IEEE Access
Convolutional neural network
recurrent neural network
natural language processing
deep learning
sentiment analysis
long-term dependencies
author_facet Abdalraouf Hassan
Ausif Mahmood
author_sort Abdalraouf Hassan
title Convolutional Recurrent Deep Learning Model for Sentence Classification
title_short Convolutional Recurrent Deep Learning Model for Sentence Classification
title_full Convolutional Recurrent Deep Learning Model for Sentence Classification
title_fullStr Convolutional Recurrent Deep Learning Model for Sentence Classification
title_full_unstemmed Convolutional Recurrent Deep Learning Model for Sentence Classification
title_sort convolutional recurrent deep learning model for sentence classification
publisher IEEE
series IEEE Access
issn 2169-3536
publishDate 2018-01-01
description As the amount of unstructured text data that humanity produces overall and on the Internet grows, so does the need to intelligently to process it and extract different types of knowledge from it. Convolutional neural networks (CNNs) and recurrent neural networks (RNNs) have been applied to natural language processing systems with comparative, remarkable results. The CNN is a noble approach to extract higher level features that are invariant to local translation. However, it requires stacking multiple convolutional layers in order to capture long-term dependencies, due to the locality of the convolutional and pooling layers. In this paper, we describe a joint CNN and RNN framework to overcome this problem. Briefly, we use an unsupervised neural language model to train initial word embeddings that are further tuned by our deep learning network, then, the pre-trained parameters of the network are used to initialize the model. At a final stage, the proposed framework combines former information with a set of feature maps learned by a convolutional layer with long-term dependencies learned via long-short-term memory. Empirically, we show that our approach, with slight hyperparameter tuning and static vectors, achieves outstanding results on multiple sentiment analysis benchmarks. Our approach outperforms several existing approaches in term of accuracy; our results are also competitive with the state-of-the-art results on the Stanford Large Movie Review data set with 93.3% accuracy, and the Stanford Sentiment Treebank data set with 48.8% fine-grained and 89.2% binary accuracy, respectively. Our approach has a significant role in reducing the number of parameters and constructing the convolutional layer followed by the recurrent layer as a substitute for the pooling layer. Our results show that we were able to reduce the loss of detailed, local information and capture long-term dependencies with an efficient framework that has fewer parameters and a high level of performance.
topic Convolutional neural network
recurrent neural network
natural language processing
deep learning
sentiment analysis
long-term dependencies
url https://ieeexplore.ieee.org/document/8314136/
work_keys_str_mv AT abdalraoufhassan convolutionalrecurrentdeeplearningmodelforsentenceclassification
AT ausifmahmood convolutionalrecurrentdeeplearningmodelforsentenceclassification
_version_ 1724193772350210048