What is: Asymmetrical Bi-RNN?
Source | Asymmetrical Bi-RNN for pedestrian trajectory encoding |
Year | 2000 |
Data Source | CC BY-SA - https://paperswithcode.com |
An aspect of Bi-RNNs that could be undesirable is the architecture's symmetry in both time directions.
Bi-RNNs are often used in natural language processing, where the order of the words is almost exclusively determined by grammatical rules and not by temporal sequentiality. However, in some cases, the data has a preferred direction in time: the forward direction.
Another potential drawback of Bi-RNNs is that their output is simply the concatenation of two naive readings of the input in both directions. In consequence, Bi-RNNs never actually read an input by knowing what happens in the future. Conversely, the idea behind U-RNN, is to first do a backward pass, and then use during the forward pass information about the future.
We accumulate information while knowing which part of the information will be useful in the future as it should be relevant to do so if the forward direction is the preferred direction of the data.
The backward and forward hidden states and are obtained according to these equations:
\begin{equation} \begin{aligned} &h_{t-1}^{b}=R N N\left(h_{t}^{b}, e_{t}, W_{b}\right) \ &h_{t+1}^{f}=R N N\left(h_{t}^{f},\left[e_{t}, h_{t}^{b}\right], W_{f}\right) \end{aligned} \end{equation}
where and are learnable weights that are shared among pedestrians, and denotes concatenation. The last hidden state is then used as the encoding of the sequence.