Convolutional Recurrent Deep Learning Model for Sentence Classification
As the amount of unstructured text data that humanity produces overall and on the Internet grows, so does the need to intelligently to process it and extract different types of knowledge from it. Convolutional neural networks (CNNs) and recurrent neural networks (RNNs) have been applied to natural l...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2018-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/8314136/ |
id |
doaj-d04f7698f9fa434193885af27bcd7839 |
---|---|
record_format |
Article |
spelling |
doaj-d04f7698f9fa434193885af27bcd78392021-03-29T21:00:25ZengIEEEIEEE Access2169-35362018-01-016139491395710.1109/ACCESS.2018.28148188314136Convolutional Recurrent Deep Learning Model for Sentence ClassificationAbdalraouf Hassan0https://orcid.org/0000-0002-1852-4449Ausif Mahmood1Department of Computer Science and Engineering, University of Bridgeport, Bridgeport, CT, USADepartment of Computer Science and Engineering, University of Bridgeport, Bridgeport, CT, USAAs the amount of unstructured text data that humanity produces overall and on the Internet grows, so does the need to intelligently to process it and extract different types of knowledge from it. Convolutional neural networks (CNNs) and recurrent neural networks (RNNs) have been applied to natural language processing systems with comparative, remarkable results. The CNN is a noble approach to extract higher level features that are invariant to local translation. However, it requires stacking multiple convolutional layers in order to capture long-term dependencies, due to the locality of the convolutional and pooling layers. In this paper, we describe a joint CNN and RNN framework to overcome this problem. Briefly, we use an unsupervised neural language model to train initial word embeddings that are further tuned by our deep learning network, then, the pre-trained parameters of the network are used to initialize the model. At a final stage, the proposed framework combines former information with a set of feature maps learned by a convolutional layer with long-term dependencies learned via long-short-term memory. Empirically, we show that our approach, with slight hyperparameter tuning and static vectors, achieves outstanding results on multiple sentiment analysis benchmarks. Our approach outperforms several existing approaches in term of accuracy; our results are also competitive with the state-of-the-art results on the Stanford Large Movie Review data set with 93.3% accuracy, and the Stanford Sentiment Treebank data set with 48.8% fine-grained and 89.2% binary accuracy, respectively. Our approach has a significant role in reducing the number of parameters and constructing the convolutional layer followed by the recurrent layer as a substitute for the pooling layer. Our results show that we were able to reduce the loss of detailed, local information and capture long-term dependencies with an efficient framework that has fewer parameters and a high level of performance.https://ieeexplore.ieee.org/document/8314136/Convolutional neural networkrecurrent neural networknatural language processingdeep learningsentiment analysislong-term dependencies |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Abdalraouf Hassan Ausif Mahmood |
spellingShingle |
Abdalraouf Hassan Ausif Mahmood Convolutional Recurrent Deep Learning Model for Sentence Classification IEEE Access Convolutional neural network recurrent neural network natural language processing deep learning sentiment analysis long-term dependencies |
author_facet |
Abdalraouf Hassan Ausif Mahmood |
author_sort |
Abdalraouf Hassan |
title |
Convolutional Recurrent Deep Learning Model for Sentence Classification |
title_short |
Convolutional Recurrent Deep Learning Model for Sentence Classification |
title_full |
Convolutional Recurrent Deep Learning Model for Sentence Classification |
title_fullStr |
Convolutional Recurrent Deep Learning Model for Sentence Classification |
title_full_unstemmed |
Convolutional Recurrent Deep Learning Model for Sentence Classification |
title_sort |
convolutional recurrent deep learning model for sentence classification |
publisher |
IEEE |
series |
IEEE Access |
issn |
2169-3536 |
publishDate |
2018-01-01 |
description |
As the amount of unstructured text data that humanity produces overall and on the Internet grows, so does the need to intelligently to process it and extract different types of knowledge from it. Convolutional neural networks (CNNs) and recurrent neural networks (RNNs) have been applied to natural language processing systems with comparative, remarkable results. The CNN is a noble approach to extract higher level features that are invariant to local translation. However, it requires stacking multiple convolutional layers in order to capture long-term dependencies, due to the locality of the convolutional and pooling layers. In this paper, we describe a joint CNN and RNN framework to overcome this problem. Briefly, we use an unsupervised neural language model to train initial word embeddings that are further tuned by our deep learning network, then, the pre-trained parameters of the network are used to initialize the model. At a final stage, the proposed framework combines former information with a set of feature maps learned by a convolutional layer with long-term dependencies learned via long-short-term memory. Empirically, we show that our approach, with slight hyperparameter tuning and static vectors, achieves outstanding results on multiple sentiment analysis benchmarks. Our approach outperforms several existing approaches in term of accuracy; our results are also competitive with the state-of-the-art results on the Stanford Large Movie Review data set with 93.3% accuracy, and the Stanford Sentiment Treebank data set with 48.8% fine-grained and 89.2% binary accuracy, respectively. Our approach has a significant role in reducing the number of parameters and constructing the convolutional layer followed by the recurrent layer as a substitute for the pooling layer. Our results show that we were able to reduce the loss of detailed, local information and capture long-term dependencies with an efficient framework that has fewer parameters and a high level of performance. |
topic |
Convolutional neural network recurrent neural network natural language processing deep learning sentiment analysis long-term dependencies |
url |
https://ieeexplore.ieee.org/document/8314136/ |
work_keys_str_mv |
AT abdalraoufhassan convolutionalrecurrentdeeplearningmodelforsentenceclassification AT ausifmahmood convolutionalrecurrentdeeplearningmodelforsentenceclassification |
_version_ |
1724193772350210048 |