Back to articles
Understanding Seq2Seq Neural Networks – Part 5: Decoding the Context Vector

Understanding Seq2Seq Neural Networks – Part 5: Decoding the Context Vector

via Dev.toRijul Rajesh

In the previous article, we stopped at the concept of the context vector . In this article, we will start by decoding the context vector . Connecting the Decoder The first thing we need to do is connect the long-term and short-term memories (the cell states and hidden states ) that form the context vector to a new set of LSTMs . Just like the encoder , the decoder will also have two layers , and each layer will have two LSTM cells . The LSTMs in the decoder are different from the ones in the encoder and have their own separate weights and biases . Using the Context Vector The context vector is used to initialize the long-term and short-term memories (the cell states and hidden states ) in the LSTMs of the decoder . This is important because it allows the decoder to start with the information learned from the input sentence . Goal of the Decoder The ultimate goal of the decoder is to convert the context vector into the output sentence . In simple terms, the encoder understands the input

Continue reading on Dev.to

Opens in a new tab

Read Full Article
2 views

Related Articles