Sunteți pe pagina 1din 4

Fake News Identification with

CNN and RNN Models


LSTM:

Long short-term memory (LSTM) is an artificial recurrent neural network (RNN)


architecture used in the field of deep learning. Unlike standard feed forward neural networks,
LSTM has feedback connections. It can not only process single data points (such as images), but
also entire sequences of data (such as speech or video). For example, LSTM is applicable to
tasks such as unsegmented, connected handwriting recognition or speech recognition. Bloomberg
Business Week wrote: "These powers make LSTM arguably the most commercial AI
achievement, used for everything from predicting diseases to composing music."

A common LSTM unit is composed of a cell, an input gate, an output gate and a forget
gate. The cell remembers values over arbitrary time intervals and the three gates regulate the
flow of information into and out of the cell.

LSTM networks are well-suited to classifying, processing and making predictions based
on time series data, since there can be lags of unknown duration between important events in a
time series. LSTMs were developed to deal with the exploding and vanishing gradient problems
that can be encountered when training traditional RNNs. Relative insensitivity to gap length is an
advantage of LSTM over RNNs, hidden Markov models and other sequence learning methods in
numerous applications.[citation needed]
CNN:

In deep learning, a convolutional neural network (CNN, or ConvNet) is a class of deep


neural networks, most commonly applied to analyzing visual imagery.

CNNs are regularized versions of multilayer perceptrons. Multilayer perceptrons usually


mean fully connected networks, that is, each neuron in one layer is connected to all neurons in
the next layer. The "fully-connectedness" of these networks makes them prone to over
fitting data. Typical ways of regularization include adding some form of magnitude measurement
of weights to the loss function. However, CNNs take a different approach towards regularization:
they take advantage of the hierarchical pattern in data and assemble more complex patterns using
smaller and simpler patterns. Therefore, on the scale of connectedness and complexity, CNNs are
on the lower extreme.
Hybrid Model:

Text classification is a classic task in the field of natural language processing. however, the
existing methods of text classification tasks still need to be improved because of the complex
abstraction of text semantic information and the strong relecvance of context. In this paper, we
combine the advantages of two traditional neural network model, Long Short-Term
Memory(LSTM) and Convolutional Neural Network(CNN). LSTM can effectively preserve the
characteristics of historical information in long text sequences, and extract local features of text
by using the structure of CNN. We proposes a hybrid model of LSTM and CNN, construct CNN
model on the top of LSTM, the text feature vector output from LSTM is further extracted by
CNN structure. The performance of the hybrid model is compared with that of other models in
the experiment. The experimental results show that the hybrid model can effectively improve the
accuracy of text classification.

S-ar putea să vă placă și