Hi, in this blog I will be covering most of the sequence networks. Sequence networks can have either input as sequence or output as sequence or both(input and output) as sequences. We can sub-divide these sequence networks in the following three ways:
- Vec2Seq
- Seq2Vec
- Seq2Seq
The above generative model is called Recurrent Neural Network(RNN).
2. Seq2Vec(Sequence Classification):
In classification task, the output is class label,
=Cat(y|softmax(Wh_T)))
We get better results if we let the hidden states depends on past as well as future context
bidirectional RNN.
Then we define, hidden state at time
,
.
=Cat(y|W \cdot softmax(\bar{h})))
where 
3. Seq2Seq(Sequence Translation):
- Aligned case
- Unaligned case
where the initial state,
=f_{0}(x_{1}))
we can get more espressive models by stacking multiple hidden chains on top of each other. For layer
at time
:

We map a sequence of length
to a sequence of length
. In this we first encode the input to get context vector
using the last state of an RNN(or avg pooling over a biRNN). Then generate the output sequence using an RNN decoder
(encoder-decoder architecture).
Comments
Post a Comment