TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Methods/Temporal Distribution Characterization

Temporal Distribution Characterization

SequentialIntroduced 20001 papers
Source Paper

Description

Temporal Distribution Characterization, or TDC, is a module used in the AdaRNN architecture to characterize the distributional information in a time series.

Based on the principle of maximum entropy, maximizing the utilization of shared knowledge underlying a times series under temporal covariate shift can be done by finding periods which are most dissimilar to each other, which is also considered as the worst case of temporal covariate shift since the cross-period distributions are the most diverse. TDC achieves this goal for splitting the time-series by solving an optimization problem whose objective can be formulated as:

max⁡_0<K≤K_0max⁡_n_1,⋯ ,n_K1K∑1≤i≠j≤Kd(D_i,D_j)\max \_{0<K \leq K\_{0}} \max \_{n\_{1}, \cdots, n\_{K}} \frac{1}{K} \sum_{1 \leq i \neq j \leq K} d\left(\mathcal{D}\_{i}, \mathcal{D}\_{j}\right) max_0<K≤K_0max_n_1,⋯,n_KK1​1≤i=j≤K∑​d(D_i,D_j)  s.t. ∀i,Δ1<∣D_i∣<Δ2;∑i∣D_i∣=n\text { s.t. } \forall i, \Delta_{1}<\left|\mathcal{D}\_{i}\right|<\Delta_{2} ; \sum_{i}\left|\mathcal{D}\_{i}\right|=n s.t. ∀i,Δ1​<​D_i​<Δ2​;i∑​​D_i​=n

where ddd is a distance metric, Δ_1\Delta\_{1}Δ_1 and Δ_2\Delta\_{2}Δ_2 are predefined parameters to avoid trivial solutions (e.g., very small values or very large values may fail to capture the distribution information), and K_0K\_{0}K_0 is the hyperparameter to avoid over-splitting. The metric d(⋅,⋅)d(\cdot, \cdot)d(⋅,⋅) above can be any distance function, e.g., Euclidean or Editing distance, or some distribution-based distance / divergence, like MMD [14] and KL-divergence.

The learning goal of the optimization problem (1) is to maximize the averaged period-wise distribution distances by searching KKK and the corresponding periods so that the distributions of each period are as diverse as possible and the learned prediction model has better a more generalization ability.

Papers Using This Method

AdaRNN: Adaptive Learning and Forecasting of Time Series2021-08-10