pith. machine review for the scientific record. sign in

arxiv: 2510.24574 · v2 · submitted 2025-10-28 · 💻 cs.LG · cs.AI

Recognition: unknown

DistDF: Time-Series Forecasting Needs Joint-Distribution Wasserstein Alignment

Authors on Pith no claims yet
classification 💻 cs.LG cs.AI
keywords conditionaldiscrepancyforecastingtime-seriesdistdflabelachievesalignment
0
0 comments X
read the original abstract

Training time-series forecasting models requires aligning the conditional distribution of model forecasts with that of the label sequence. The standard direct forecast (DF) approach resorts to minimizing the conditional negative log-likelihood, typically estimated by the mean squared error. However, this estimation proves biased when the label sequence exhibits autocorrelation. In this paper, we propose DistDF, which achieves alignment by minimizing a distributional discrepancy between the conditional distributions of forecast and label sequences. Since such conditional discrepancies are difficult to estimate from finite time-series observations, we introduce a joint-distribution Wasserstein discrepancy for time-series forecasting, which provably upper bounds the conditional discrepancy of interest. The proposed discrepancy is tractable, differentiable, and readily compatible with gradient-based optimization. Extensive experiments show that DistDF improves diverse forecasting models and achieves leading performance. Code is available at https://anonymous.4open.science/r/DistDF-F66B.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 1 Pith paper

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. Optimal Transport for LLM Reward Modeling from Noisy Preference

    cs.LG 2026-05 unverdicted novelty 6.0

    SelectiveRM applies optimal transport with a joint consistency discrepancy and partial mass relaxation to produce reward models that optimize a tighter upper bound on clean risk while autonomously dropping noisy prefe...