What is: Temporal Distribution Matching?
Source | AdaRNN: Adaptive Learning and Forecasting of Time Series |
Year | 2000 |
Data Source | CC BY-SA - https://paperswithcode.com |
Temporal Distribution Matching, or TDM, is a module used in the AdaRNN architecture to match the distributions of the discovered periods to build a time series prediction model Given the learned time periods, the TDM module is designed to learn the common knowledge shared by different periods via matching their distributions. Thus, the learned model is expected to generalize well on unseen test data compared with the methods which only rely on local or statistical information.
Within the context of AdaRNN, Temporal Distribution Matching aims to adaptively match the distributions between the RNN cells of two periods while capturing the temporal dependencies. TDM introduces the importance vector to learn the relative importance of hidden states inside the RNN, where all the hidden states are weighted with a normalized . Note that for each pair of periods, there is an , and we omit the subscript if there is no confusion. In this way, we can dynamically reduce the distribution divergence of cross-periods.
Given a period-pair , the loss of temporal distribution matching is formulated as:
where denotes the distribution importance between the periods and at state .
All the hidden states of the RNN can be easily computed by following the standard RNN computation. Denote by the computation of a next hidden state based on a previous state. The state computation can be formulated as
The final objective of temporal distribution matching (one RNN layer) is:
where is a trade-off hyper-parameter. Note that in the second term, we compute the average of the distribution distances of all pairwise periods. For computation, we take a mini-batch of and to perform forward operation in RNN layers and concatenate all hidden features. Then, we can perform TDM using the above equation.