Convolutional Recurrent Neural Networks for Text Classification

Convolutional Recurrent Neural Networks for Text Classification

Shengfei Lyu, Jiaqi Liu
Copyright: © 2021 |Pages: 18
DOI: 10.4018/JDM.2021100105
OnDemand:
(Individual Articles)
Available
$37.50
No Current Special Offers
TOTAL SAVINGS: $37.50

Abstract

Recurrent neural network (RNN) and convolutional neural network (CNN) are two prevailing architectures used in text classification. Traditional approaches combine the strengths of these two networks by straightly streamlining them or linking features extracted from them. In this article, a novel approach is proposed to maintain the strengths of RNN and CNN to a great extent. In the proposed approach, a bi-directional RNN encodes each word into forward and backward hidden states. Then, a neural tensor layer is used to fuse bi-directional hidden states to get word representations. Meanwhile, a convolutional neural network is utilized to learn the importance of each word for text classification. Empirical experiments are conducted on several datasets for text classification. The superior performance of the proposed approach confirms its effectiveness.
Article Preview
Top

Introduction

Text classification is a basic task in Natural Language Processing (NLP). It aims to classify a text into an appropriate class. There are many practical applications of text classification, such as sentiment analysis (Pang & Lee, 2005) and topic categorization (Wang & Manning, 2012).

Traditional approaches, based on the text representation, extract features for a general classifier. For instance, in the model of bag-of-words, statistics on unigrams, bigrams, and k-grams are used as features (Pang et al., 2002; Wang & Manning, 2012). These traditional approaches either totally ignore the word order or constrain their focus on small word tuples, which results in inevitable information loss. Moreover, they suffer from the problems of data sparsity and high dimensionality.

In recent years, the development of deep neural networks has drawn attention in various NLP tasks with the help of the pre-trained word embeddings (Pennington et al., 2014; Xu et al., 2016; Liu et al., 2017). Words are embedded into an implicit space where each word is represented by a vector. Semantic relationships among words are preserved in the space by their reciprocal distances commonly. Word embeddings can not only alleviate the problem of data sparsity (Bengio et al., 2003), but also are demonstrated that they preserve meaningful syntactic and semantic regularities (Pennington et al. 2014). With the assistance of word embeddings, deep learning approaches, especially recurrent neural networks (RNN) (Elman 1990) and convolutional neural networks (CNN) (LeCun et al., 2003) are introduced to learn the semantic representation of the text. These approaches have promoted the development of text classification (Kalchbrenner et al., 2014; Jiang et al., 2018).

RNN is a natural way to process a text word by word by regarding it as a sequential data. RNN preserves a word’s information and its contextual information into a hidden unit. Long Short-term Memory (LSTM) units (Hochreiter & Schmidhuber, 1997) and its updated version, gated recurrent units (GRU) (Chung et al. 2014), can solve gradient vanishing and gradient explosion problems, which makes the two units become popular prototypes. Theoretically, arbitrary long sequences including long-term dependencies in texts can be modeled by RNN (Cho et al. 2014). Alternative popular neural network model, CNN has made impressive progress in modeling semantic information of texts (Kalchbrenner et al., 2014; Kim, 2014). CNN performs the feature mapping by a one-dimension (1D) convolution operator in a two-dimension (2D) matrix that is composed of concatenated embeddings of words in texts together. Then, a 1D pooling operation is applied to the convolutional features over the time domain to obtain a fixed-length vector as output. Both position-invariant and local features in texts can be captured by CNN due to its convolutional operation.

Complete Article List

Search this Journal:
Reset
Volume 35: 1 Issue (2024)
Volume 34: 3 Issues (2023)
Volume 33: 5 Issues (2022): 4 Released, 1 Forthcoming
Volume 32: 4 Issues (2021)
Volume 31: 4 Issues (2020)
Volume 30: 4 Issues (2019)
Volume 29: 4 Issues (2018)
Volume 28: 4 Issues (2017)
Volume 27: 4 Issues (2016)
Volume 26: 4 Issues (2015)
Volume 25: 4 Issues (2014)
Volume 24: 4 Issues (2013)
Volume 23: 4 Issues (2012)
Volume 22: 4 Issues (2011)
Volume 21: 4 Issues (2010)
Volume 20: 4 Issues (2009)
Volume 19: 4 Issues (2008)
Volume 18: 4 Issues (2007)
Volume 17: 4 Issues (2006)
Volume 16: 4 Issues (2005)
Volume 15: 4 Issues (2004)
Volume 14: 4 Issues (2003)
Volume 13: 4 Issues (2002)
Volume 12: 4 Issues (2001)
Volume 11: 4 Issues (2000)
Volume 10: 4 Issues (1999)
Volume 9: 4 Issues (1998)
Volume 8: 4 Issues (1997)
Volume 7: 4 Issues (1996)
Volume 6: 4 Issues (1995)
Volume 5: 4 Issues (1994)
Volume 4: 4 Issues (1993)
Volume 3: 4 Issues (1992)
Volume 2: 4 Issues (1991)
Volume 1: 2 Issues (1990)
View Complete Journal Contents Listing