target="_blank" rel="nofollow" href="#fb3_img_img_d70dea5c-88e2-5d8f-abda-52bc98436b0e.png" alt="StartLayout 1st Row 1st Column bold-italic z overTilde equals g Subscript phi Baseline left-parenthesis bold-italic epsilon comma bold-italic x right-parenthesis comma bold-italic epsilon tilde p left-parenthesis bold-italic epsilon right-parenthesis 2nd Column Blank EndLayout"/>
where
(8)
where
6 Recurrent Neural Networks
6.1 Introduction
The previously introduced models have the same assumptions on the data, that is, the independence among the samples and the fixed input size. However, these assumptions may not be true in some cases, thus limiting the application of these models. For example, videos can have different lengths, and frames of the same video are not independent, and sentences of an chapter can have different lengths and are not independent.
A RNN is another modified DNN that is used primarily to handle sequential and time series data. In a RNN, the hidden layer of each input is a function of not just the input layer but also the previous hidden layers of the inputs before it. Therefore, it addresses the issues of dependence among samples and does not have any restriction on the input size. RNNs are used primarily in natural language processing applications, such as document modeling and speech recognition.
6.2 Architecture
As illustrated in Figure 7, a general neural network
RNN can work with sequence data, which has input as sequence or target as sequence or both. An input sequence data can be denoted as
Figure 7 Feedforward network.
Figure 8 Architecture of recurrent neural network (RNN).
Though multiple network blocks are shown on the right side of Figure