MOMENT: A Family of Open Time-series Foundation Models
CoRR(2024)
摘要
We introduce MOMENT, a family of open-source foundation models for
general-purpose time-series analysis. Pre-training large models on time-series
data is challenging due to (1) the absence of a large and cohesive public
time-series repository, and (2) diverse time-series characteristics which make
multi-dataset training onerous. Additionally, (3) experimental benchmarks to
evaluate these models, especially in scenarios with limited resources, time,
and supervision, are still in their nascent stages. To address these
challenges, we compile a large and diverse collection of public time-series,
called the Time-series Pile, and systematically tackle time-series-specific
challenges to unlock large-scale multi-dataset pre-training. Finally, we build
on recent work to design a benchmark to evaluate time-series foundation models
on diverse tasks and datasets in limited supervision settings. Experiments on
this benchmark demonstrate the effectiveness of our pre-trained models with
minimal data and task-specific fine-tuning. Finally, we present several
interesting empirical observations about large pre-trained time-series models.
Our code is available anonymously at anonymous.4open.science/r/BETT-773F/.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要