Sequence modeling is a task of modeling sequential data. Modeling sequence data is when you create a mathematical notion to understand and study sequential data, and use those understandings to generate, predict or classify the same for a specific application.

Sequential data has three properties:

- Elements in the sequence can repeat
- It follows order (contextual arrangement)
- Length of data varies (potentially infinitely)

Example of sequential data are:

- Text and sentences
- Audio (eg. speech and music)
- Motion pictures or videos
- Time-series data (eg. Stock market data)
- DNA sequence, protein structures
- Material composition
- Decision-making

Sequence data is difficult to model because of its properties, and it requires a different method. For instance, if sequential data is fed through a feed-forward network, it might not be able to model it well, because **sequential data has variable length**. The feed-forward network works well with fixed-size input, and doesn’t take structure into account well.

Convolutional neural networks, on the other hand, were created to process structures, or grids of data, such as an image. They can deal with long sequences of data, but are limited by the fact that **they can’t order the sequence correctly**.

So, **how do we build deep learning models that can model sequential data? **Modeling sequential data is not an easy task.

For instance, when it comes to modeling a supervised learning task, our approach is to feed the neural network with a pair of input (x) and output (y). During training, the model learns to map the input with the output by approximating a value closer to the original value.

*y ≈ f _{θ}(x) *

During each iteration, the model tries to correct the approximated output through a loss function, i.e. by taking the difference of the original or ground truth value with the predicted value, and simultaneously updating the parameter of the model through an optimization process. This way, it makes sure that the difference between them reduces, and the approximated value reaches as close to the original value as possible.

When we process sequential data, we try to model the input sequence. Unlike a supervised learning task, where we map the input with the output, in sequence modelling we try to model how **probable** the sequence is.

This gives machine learning or deep learning models the ability to generate likeable sequences, or an ability to estimate the likeliness of the sequence. The rest of the process of calculating the loss function and optimisation remains the same.

## How to model sequences

Let’s understand how to model sequences with an example.

In the following example, we’ll use sequences of english words (sentences) for modeling, because they inherit the same properties as what we discussed earlier.

### Modeling p(x)

Assuming that words in a sentence are **independent** to each other, we can use a corpus which tells us how probable each of the words in the English language is.

Once we know the probability of each word (from the corpus), we can then find the probability of the entire sentence by multiplying individual words with each other.

For instance, if we were to model the sentence “*Cryptocurrency is the next big thing*”, then it would look something like this:

p(“*Cryptocurrency*”)p(“*is*”)p(“*the*”)p(“*next*”)p(“*big*”)p(“*thing*”)

The above model can be described in a formula:

Each word is given a timestamp: t, t-1, t-2, t-n, which describes the position of an individual word.

But, it turns out that the model described above does not really capture the structure of the sequence. Why?

Because the probability of any particular word can be higher than the rest of the word. In our example, the probability of the word “*the**” *is higher than any other word, so the resultant sequence will be “*The the the the the the”*.

### Modeling p(x|context)

Although we could still modify the same model by introducing **conditional probability**, assuming that each word is now dependent on every other word rather than independent, we could now model the sequence in the following way: p(x_{T}) = p(x_{T} | x_{1}…., x_{T-1}).

The same sentence “*Cryptocurrency is the next big ______*” can now have a range of options to choose from. For example:

Stuff

0.0002

Thing

0.01

Coin

0.00003

*Essentially, conditional probability describes what the next word will be.*

But the above example can predict one word at a time; in order to predict a sequence of words we need to calculate the **joint probability from the conditionals**.

For instance,

**Cryptocurrency**

p(x1)

*Cryptocurrency is*

p(x2|x1)

*Cryptocurrency is the*

p(x3|x2,x1)

*Cryptocurrency is the next*

p(x4|x3,x2,x1)

*Cryptocurrency is the next big*

p(x5|x4,x3,x2,x1)

*Cryptocurrency is the next big thing*

p(x6|x5,x4,x3,x2,x1)

**the predicted words are written in bold*

So far we understood how to go about modeling a sequence of sentences using a corpus of words (which contains probability value), and then calculating **joint** and **conditional** probability of each word in the sentence we could select the appropriate word from the corpus and predict the next word in a sentence.

This type of approach works well with a few sentences, and **captures the structure of the data very well**. But when we deal with paragraphs, then we have to deal with **scalability**. When such models are introduced with large sentences, then processing power increases and efficiency decreases.

### N-grams

To counter the issue of scalability, NLP (natural language processing) researchers introduced the idea of **N-grams**, where you take into account an ‘n’ number of words for conditional and joint probability. For instance, if n is equal 2, then only the previous two words of the sentence will be used to calculate joint probability instead of the entire sentence.

**Cryptocurrency**

p(x1)

*Cryptocurrency is*

p(x2|x1)

*Cryptocurrency is the*

p(x3|x2,x1)

*Cryptocurrency is the next*

p(x4|x3,x2)

*Cryptocurrency is the next big*

p(x5|x4,x3)

*Cryptocurrency is the next big thing*

p(x6|x5,x4)

**the predicted words are written in bold*

This approach reduces the scalability issue, but not completely.

The disadvantages of N-grams are:

- Context of the sentence is lost if the sentence is long.
- Reduces the scalability issue by a small scale

### Context vectorizing

Context vectorizing is an approach where the input sequence is summarized to a vector such that that vector is then used to predict what the next word could be.

f h

f* _{θ}* summarizes the context in h such that:

*p(x _{t}|x_{1},….,x_{t-1}) ≈ p(x_{t}|h)*

Once we find the context vector **h**, we can then use a **second function** g which produces a **probability distribution**.

*h g _{θ}*

The advantages of context vectorizing are:

- Order is preserved
- Can operate in variable length of sequences
- It can learn hence differentiable (backpropagation)
- Context is preserved in short sentences or sequences.

So far we have seen what sequential data is and how to model it. In the next section, we will learn about RNNs and how they use context vectorizing to predict the next word.

## What are recurrent neural networks (RNNs)?

Recurrent neural networks are used to model sequential data with the time step index t, and incorporate the technique of context vectorizing.

Context vectoring acts as “memory” which captures information about what has been calculated so far, and enables RNNs to remember past information, where they’re able to preserve information of long and variable sequences. Because of that, RNNs can take one or multiple input vectors and produce one or multiple output vectors.

RNNs by design are similar to deep neural networks. They have input vectors, weight vectors, hidden states and output vectors. The hidden state captures the patterns or the context of a sequence into a summary vector. The outputs are influenced not just by weights applied on inputs like a regular neural network, but also by a “hidden” state vector representing the context based information on prior inputs, such that the same input could produce a different output depending on context of inputs in the sequence.

## How does RNN work?

As discussed RNNs have four major components:

- Input
- Hidden state
- Weights or parameters
- Output

To understand RNN better we need to see how the multi-layer perceptron works with sequential data.

### Multi-layer perceptron

In the multi-layer perceptron (MLP), we have an input layer, a hidden layer and an output layer. The input layer receives the input, passes it through the hidden layer where activations are applied, and then returns the output.

When multiple hidden layers are present, the activations are applied to each node of a first hidden layer, and then sent to the next hidden layer, and so forth until it reaches the last layer to produce the output. Each hidden layer is characterized by its own weights and biases, making them independent of each other.

But the relationship can’t be established between the two unless we combine these hidden layers together because we’re dealing with sequences. **If the hidden layers are independent of each other, then the structure and the order will be lost**. So, in order to combine these hidden layers together, we should have the same weights and bias for each of these hidden layers.

To achieve this continuous and structured relationship between the input and output, we need to feed the input sentence word by word into all the hidden layers in a sequence, simultaneously feeding the output of the previous neuron to the next.

This approach makes the overall process structured and contextual.

This is the inception of recurrent neural networks, where previous input combines with the current input, thereby preserving some relationship of the current input (x2) with the previous input (x1).

In essence, RNNs are a modified version of MLP, where the data is fed in each hidden layer.

### Recurrent neural network

In RNNs, x(t) is taken as the input to the network at time step t. The time step t in RNN indicates the order in which a word occurs in a sentence or sequence. The hidden state h(t) represents a **contextual vector** at time t and acts as “**memory**” of the network. Contextual vector h(t) is calculated based on the current input and the previous time step’s hidden state.

For instance, take the sentence “*Cryptocurrency is the next big thing”. *At time t1 “Cryptocurrency” will be fed into the network, at time t2 “is” be fed into the network, at time t3 “the” be fed into the network, and so forth.

We denote a hidden state using this formula:

*h _{t}= tanh(W_{h}h_{t-1} + W_{x}x_{t})*

When t = 1,

*h _{1}= tanh(W_{h}h_{0} + W_{x}x_{1})*, where x

_{1}is ‘Cryptocurrency’, and

**h**

_{0}is initialised as zeroWhen t = 2,

*h _{2}= tanh(W_{h}h_{1} + W_{x}x_{2})*, where x

_{1}is ‘is’.

When t = 3,

*h _{3}= tanh(W_{h}h_{2} + W_{x}x_{3})*, where x

_{2}is ‘the’.

The weight parameters for both hidden state and input are learnable, which means that during the training it will update itself using backpropagation.

If you notice the formula above you will observe that the parameters ={Wh,Wx,Wy} remain constant in each and every input. This is known as **parameter sharing**. Parameter sharing gives RNNs the ability to structure and retain contextual information even if the length of the sentence has variable length. Such sharing is important when we want to model a sequence where a data can reoccur quite often. **RNNs share the same weight across several time steps.**

The RNNs predict the output from the last hidden state along with output parameter Wy. Prediction is more of a classification task, where a **softmax** function is used to ensure the probability over all the possible words in the english sentence.

## Training RNNs

### Computing gradients: Backpropagation through time

So far we’ve seen how feed-forward works in RNNs, i.e. the inputs enter the network and move forward while sharing the same parameter across each time step. During backpropagation it has to go back through the time-step to update the parameters.

So how does it happen?

As discussed earlier, predicting output in RNN is a classification task, so we use **cross-entropy **to calculate the loss.

*L _{θ}(y,y’)_{t} = -y_{t}logy_{t}‘*

Where *θ={W _{h},W_{x},W_{y}}*

We know that the parameters are shareable, and it’s this ability of RNNs that makes all the difference in the network; it’s quite tricky to differentiate. In order to tackle this problem, the network needs to be expanded, or **unfolded**, so that the parameters could be differentiated across the network with respect to time step – hence the term backpropagation through time (BPTT).

From the above diagram you can see how an unfolded recurrent network looks like.

Since the neural network takes one word at a time, the loss calculated is based on per-word basis. The overall procedure looks like this:

You can see how the expansion in formula takes place.

For differentiation with respect to Wy, the procedure should look like this:

So, let’s summarize backpropagation:

- Loss is computed using the current output and the actual output using cross-entropy loss
- The network is unrolled, or expanded, so that the backpropagation could take place through hidden layers and also through time
- The weights parameters are updated for both recurrent neuron, i.e. hidden state, and the dense layers

### Issues with gradients

Two common problems that occur during the backpropagation of sequential data are vanishing and exploding gradients.

#### Vanishing gradients

When the differentiating vector goes to zero exponentially fast, which in turn makes it difficult for the network to learn some long period dependencies, the problem is **vanishing gradient**.

Coming to backpropagation in RNNs, we saw that every single neuron in the network participated in the calculation of the output with respect to the cost function. Because of that,** we have to make sure that the parameters are updated for every neuron to minimize the error, and this goes back to all the neurons in time.** So, you have to propagate all the way back through time to these neurons.

We also know that the **contextual vector, or the hidden state** **parameter**, is shared across the network to preserve order and continuity. During initialization, the parameter is assigned with a random number which is close to zero, and when the hidden state moves forward in time it

gets multiplied by itself over at different time steps, making the gradient Wh smaller and smaller, essentially zero to a point where it vanishes.

**The lower the gradient is, the harder it is for the network to update the weights, and if the gradient is zero, the weights will not be updated.**

#### Exploding gradients

Exploding gradients occur when large gradients accumulate due to an unstable process, and result in very large updates to the parameters.

In RNNs, exploding gradients can occur during backpropagation and result in very large gradients essentially making large updates to the network parameters. At an extreme, the values of weights can become so large that they become NaN values.

### Overcoming gradient issues

Gradient issues in RNNs can be solved with:

- Gradient clipping
- Gated networks

#### Gradient clipping

Gradient clipping is a technique used to avoid exploding gradients. It’s fair to assume that RNNs behave in an approximate linear fashion, which makes the gradient unstable.

In order to control the gradient, it’s clipped, or reshaped to a smaller value. There are two ways to clip gradients:

- Clip the gradient from a mini batch just before the parameter is updated
- Use a hyperparameter C which measures the norm ||g|| where g is the gradient. If ||g|| > C then gg.C/||g||

** ^{READ ALSO}**Understanding Gradient Clipping (and How It Can Fix Exploding Gradients Problem)

#### Gated networks

To tackle vanishing gradients, you can use new architectures with gated mechanisms. Architecture like long short term memory, and gated recurrent networks have been proven to solve vanishing gradients. We’ll dive into them in the next section.

## Summary of RNN (so far):

- Order of the sequence is preserved
- Able to map:
- Input sequence of variable to a fixed-size vector
- Input sequence of fixed-size vector to a sequence
- Input sequence to an output sequence of the same length.

- It’s differential (so it can learn)
- Context of the sequence is lost when dealing with long sequences due to gradient issues

## Types of RNN

So far, we’ve learned about:

- Sequence modeling
- What is it?
- How can we model sequences?
- Contextual vectorization

- Recurrent neural networks
- What are they?
- How do they work?
- Backpropagation through time
- Teacher forcing
- Issues with RNNs while training
- Overcoming gradient issues

Now, we’ll move on to different types of RNNs, while shining some light on the limitations of vanilla RNNs. We’ll discuss fives types of RNN:

- Bidirectional RNN
- Encoder-decoder sequence-to-sequence architecture
- LSTM
- GRU
- Convolutional recurrent neural network

### Bidirectional RNN

In sequence modeling, so far we assumed that our goal is to model the next output given a particular sequence of sentences. In an NLP task, there might be a situation where the context depends on the future sentence.

For instance:

- I am ___.
- I am ___ hungry.
- I am ___ hungry, and I can eat the whole pie.

In the above example, the answer can be ‘hungry’, ‘not’, ‘very’ respectively. But in order to understand the context, especially when the overall context depends on the future word, we need to design a model that takes the future word into consideration.

One drawback with RNNs is that they take into account the past and the present word in time, and not the future word. This makes RNNs a **unidirectional sequential network**, where information flows in one direction, usually a forward direction.

To overcome this issue we have to introduce a mechanism where:

- The information of the sequence travels backward
- The neural network considers all the three timestep past, present and future, and based upon it predicts an output.

A bidirectional RNN is a type of network that solves this problem. It’s an architecture where two independent RNNs are put together. While one works in the conventional manner, i.e. in the forward direction, the other works in the backward direction.

As you can see from the diagram above, both the networks output their individual output based on **past-present** (forward-directional RNNs) information and **future-present** (backward-directional RNNs) information at every time step.

### Encoder-decoder sequence-to-sequence architecture

The advantage of using RNNs in sequential modeling is that it can:

- Map an input sequence to a fixed-size vector
- Map fixed-size vector to a sequence
- Map an input sequence to an output sequence of the same length.

But let’s say we want to train a RNN to map an input sequence to an output sequence, not necessarily of the same length. This can come up especially when we want to translate from one language to another.

Encoder-decoder sequence-to-sequence is an architecture that deals with this type of problem. As the name suggests, it has two types of architecture: encoder and decoder.

Encoder RNN receives the input sequence of variable length, and processes it to return a vector or a sequence of vectors called the “context” vector *C*.

The decoder RNN is conditioned on a fixed-length vector to generate an output sequence. Also, the last hidden state of the encoder is the initial hidden state of the decoder.

### Long Short Term Memory (LSTM)

Long short term memory was built to **preserve long term contextual information of a sequence that the RNN lacks**.

The core idea of LSTM is to make sure the gradient flows for a long period of time, and doesn’t vanish or explode.

As we saw earlier, RNNs have a standard architecture where the hidden state formed some sort of a looping mechanism to preserve and share the information for every time step. Well, LSTM also has a looping mechanism. Instead of having a single neural network layer, there are four neural networks, interacting in a way to preserve and share long contextual information.

Let’s understand how the LSTM network works step-by-step:

**Cell state**: The**long-term memory**is usually called the cell state, denoted by Ct-1, and**recursive**in nature. Cell state allows previous information to be stored in the LSTM network. It’s modulated by the internal gates namely: forget and input gate.

**Forget gate**: The cell state is connected to the forget gate which dictates which information to preserve and which to forget. It does this through a sigmoid function. It looks at the initial state at ht-1and xt, and outputs a number between 0 and 1 for each cell state Ct-1. If the information returned by the forget gate is 1, it stores the information, and if it returns 0, it deletes the information.

**Input gate**: The input gate takes the same input xtfrom the same initial state ht-1and it passes through two non-linear activation functions: sigmoid (which returns the value between 0 and 1) and tanh (which returns the value between -1 and 1), which are**input gate**itand**candidate memory****C****`****t**. This helps the cell state to save new information and simultaneously update the cell state from Ct-1 to Ct. The idea of the input gate is to store new information in the cell state, along with the pruned memory from the forget gate.

**Output gate**: The output will be based on the cell state Ct, the initial hidden stateht-1and input xt. The hidden state and input will be passed through a sigmoid so that value is between 0 and 1. Simultaneously we pass cell state Ctthrough the tanh to get values between 1 and -1, and multiply it by the output of the sigmoid gate, so that we only output what’s relevant to the context.

In summary:

- The first
**sigmoid activation**function is the forget gate that ensures what information should be forgotten from the previous cell state (Ct-1). - The second
**sigmoid**and first**tanh**activation function is for the input gate that ensures what information should be saved to the cell state along with the previous pruned information. - The last sigmoid is the output gate and tanh from the cell that highlights which information should be going to the next hidden state.

### Gated Recurrent Unit (GRU)

The GRU is the newer generation of Recurrent Neural networks, and it’s the modified version of LSTM but with less complexity. GRUs use the hidden state to transfer information instead of cell state. It also only has two gates, a reset gate and update gate.

**Update Gate: **The update gate selects information that needs to be added and / or ignored. It’s similar to the LSTM input gate.

**Reset Gate: **The reset gate is used to decide how much past information to forget.

Since GRUs are smaller in operation, it’s faster than LSTM networks. And the reason both LSTMs and GRUs are successful, is because their gating mechanism:

- Preserves contextual information and long-term sequences.
- Avoids gradient issues.

The gating functions allow the network to modulate how much the gradient vanishes, and since it’s being copied four times, it takes different values at each time step. The values that they take on are learned functions of the current input and hidden state.

### Convolutional recurrent neural network

The Convolutional Recurrent Neural Networks is a combination of two neural networks: convolutional neural network and recurrent neural network. Both have their own unique properties that help them excel in what they do.

The CNNs are very good in extracting features and representation from any given data because of grid-like operation. On the other hand, the RNNs are very well suited for sequential data modeling, which in turn preserves order, structure and context.

The CRNN architectures are often used to generate better / optimal results towards audio signal processing. Since audio signals like music are very complicated (a lot of noises clustered together), it becomes fairly difficult to extract low dimensional patterns and representations. Why do we need low dimensional patterns?

Because high dimensional patterns don’t offer a better understanding of the structure of data. Using CNN, we can dig deeper into high dimensional sequential structure such that it gives us low dimensional representation, which can then be fed into the RNN, or even LSTM, which can be used to generate / predict another sequence.

CRNN is also used in images to sequence application example image captioning, also it can be used to generate images from a sentence, also known as sequence-to-image.

## Implementation of RNN in Keras / Tensorflow and Python

Let’s see how we can implement RNN with keras for **character text prediction**.

We’ll be using two libraries: numpy and tensorflow-keras.

```
import numpy as np
from tensorflow.keras.layers.recurrent import SimpleRNN
from tensorflow.keras.models import Sequential
from tensorflow.keras.layers import Dense, Activation
```

You can download the dataset here.

`INPUT_FILE = "data/wonderland.txt"`

### Data preprocessing

Data preprocessing is needed because the data contains ASCII characters, which might interfere with our modeling process and give incorrect results.

```
print("Extracting text from input")
file = open(INPUT_FILE, 'rb')
lines = []
for line in file:
line = line.strip().lower()
line = line.decode("ascii", "ignore")
if len(line) == 0:
continue
lines.append(line)
file.close()
text = " ".join(lines)
```

After the data is cleaned, we can create a look-up table. The look-up table basically creates a table of unique characters from the data. This table will convert the characters in the text to numbers, which will then be fed into the RNNs.

```
chars = set([c for c in text])
nb_chars = len(chars)
char2index = dict((c, i) for i, c in enumerate(chars))
index2char = dict((i, c) for i, c in enumerate(chars))
print(index2char)
```

```
{0: '!',
1: 's',
2: '7',
3: 'p',
4: 'x',
5: '.',
6: ']',
7: '3',
8: 'a',
9: 'z',
10: '9',
11: '5',
12: '6',
13: 'h',
14: ',',
15: 'c',
16: 'r',
17: '@',
18: 'k',
19: 'w',
20: 'd',
21: '(',
22: 'o',
23: 'f',.....}
```

Now, we’ll create inputs and labels. It’s important to understand that in sequence modeling, the input will start from index 0, where the label will start from index 1.

For instance, if we feed “cryptocurrency is the next big thing”, then:

cryptocurrency

is

is

the

the

next

next

big

big

thing

thing

**it’s worth mentioning that it’s a character level RNN, so instead of words you will be working with characters*

```
print("Creating input and label")
SEQ_LEN = 10 #we define the size of the character sequence in one time step
STEP = 1
print("The table is ready")
```

```
input_chars = []
label_chars = []
for i in range(0, len(text) - SEQ_LEN, STEP):
input_chars.append(text[i:i + SEQ_LEN])
label_chars.append(text[i + SEQ_LEN])
print(input_chars)
```

```
['project gu',
'roject gut',
'oject gute',
'ject guten',
'ect gutenb',
'ct gutenbe',
't gutenber',
' gutenberg',
'gutenbergs',...]
```

### One-hot-encoding

Transforming sequences to one-hot-encoded vectors.

```
print("Vectorizing input and label text...")
X = np.zeros((len(input_chars), SEQ_LEN, nb_chars), dtype=np.bool)
y = np.zeros((len(input_chars), nb_chars), dtype=np.bool)
for i, input_char in enumerate(input_chars):
for j, ch in enumerate(input_char):
X[i, j, char2index[ch]] = 1
y[i, char2index[label_chars[i]]] = 1
```

### Modeling the RNN

Defining the model:

- RNN will contain 128 hidden states
- Always remember that the dense layer should equal the number of unique characters in the text

```
HIDDEN_SIZE = 128
BATCH_SIZE = 128
NUM_ITERATIONS = 25
NUM_EPOCHS_PER_ITERATION = 1
NUM_PREDS_PER_EPOCH = 100
model = Sequential()
model.add(SimpleRNN(HIDDEN_SIZE, return_sequences=False,
input_shape=(SEQ_LEN, nb_chars),
unroll=True))
model.add(Dense(nb_chars))
model.add(Activation("softmax"))
model.compile(loss="categorical_crossentropy", optimizer="rmsprop")
```

### Training the model

```
for iteration in range(NUM_ITERATIONS):
print("=" * 50)
print("Iteration #: %d" % (iteration))
model.fit(X, y, batch_size=BATCH_SIZE, epochs=NUM_EPOCHS_PER_ITERATION)
# testing model
# randomly choose a row from input_chars, then use it to
# generate text from model for next 100 chars
test_idx = np.random.randint(len(input_chars))
test_chars = input_chars[test_idx]
print("Generating from seed: %s" % (test_chars))
print(test_chars, end="")
for i in range(NUM_PREDS_PER_EPOCH):
Xtest = np.zeros((1, SEQ_LEN, nb_chars))
for i, ch in enumerate(test_chars):
Xtest[0, i, char2index[ch]] = 1
pred = model.predict(Xtest, verbose=0)[0]
ypred = index2char[np.argmax(pred)]
print(ypred, end="")
# move forward with test_chars + ypred
test_chars = test_chars[1:] + ypred
print()
```

## Conclusion

In this article, we discussed:

- How sequences can be modelled efficiently to capture the structure and context.
- How context vector provides an efficient way to summarize a sequence
- How a multilayer perceptron can be transformed to an RNN by sequentially feeding the input directly into the hidden layer at a given time step.
- How parameter sharing can transform an independent neural network into a continuous neural network which can be differentiated and preserve the integrity of the sequence.
- The issues with the gradients and how it can be rectified.
- The different types of RNNs:
- Bidirectional RNNs : that takes into consideration the future sequence for context-based learning.
- Encoder-decoder RNNs: that uses two architectures to map an input sequence to an output sequence of variable length which a normal RNN could not do.
- LSTMs: it rectifies the issues with vanishing gradient by introducing gated mechanism and also preserves long-term dependencies or long-term contexts.
- GRU: same as LSTM but provide a simpler architecture.
- CRNN: uses the combined architecture of CNN and RNN for better extraction of high-dimensional features into low-dimensional features and to perform sequence modeling over the extracted low-dimensional respectively.

It’s a lot of stuff, so if you made it all the way to the end – thank you for reading! I hope this guide was helpful.

### Resources:

- https://towardsdatascience.com/introduction-to-sequence-modeling-problems-665817b7e583
- https://www.statisticshowto.com/probability-and-statistics/statistics-definitions/conditional-probability-definition-examples/
- https://kavita-ganesan.com/what-are-n-grams/#.YEyLImhKiUk
- https://towardsdatascience.com/recurrent-neural-networks-d4642c9bc7ce
- https://machinelearningmastery.com/neural-networks-crash-course/
- https://towardsdatascience.com/understanding-bidirectional-rnn-in-pytorch-5bd25a5dd66
- https://towardsdatascience.com/understanding-gru-networks-2ef37df6c9behttps://towardsdatascience.com/an-approach-towards-convolutional-recurrent-neural-networks-a2e6ce722b19
- https://www.tensorflow.org/guide/keras/rnn

**READ NEXT**

## How to Organize Deep Learning Projects – Examples of Best Practices

13 mins read | Author Nilesh Barla | Updated May 31st, 2021

For a successful deep learning project, you need a lot of iterations, a lot of time, and a lot of effort. To make this process less painful, you should try to use your resources to the max.

A good step-by-step workflow will help you do that. With it, your projects become **productive, reproducible,** and **understandable**.

In this article you’ll see how to structure work on deep learning projects — from the inception to deployment, monitoring the deployed model, and everything in between.

Along the way, we’ll use Neptune to run, monitor, and analyze your experiments. Neptune is a cool tool for increasing productivity in ML projects.

In this article you will learn:

- About the lifecycle of the project.
- Importance of defining an objective or goal of the project.
- Collecting data based on the requirements of the project.
- Model training and results exploration including:
- Establishing baselines for better results.
- Adopting techniques and approaches from the existing open-source state-of-the-art models research papers and code repositories.
- Experiment tracking and management management

- Model refinement techniques to avoid underfitting and overfitting like:
- Controlling hyperparameters
- Regularisation
- Pruning

- Testing and evaluating your project before deployment.
- Model deployment
- Project maintenance