STEB: In Search of the Best Evaluation Approach for Synthetic Time Series
STEB: In Search of the Best Evaluation Approach for Synthetic Time Series
The growing need for synthetic time series, due to data augmentation or privacy regulations, has led to numerous generative models, frameworks, and evaluation measures alike. Objectively comparing these measures on a large scale remains an open challenge. We propose the Synthetic Time series Evaluation Benchmark (STEB) -- the first benchmark framework that enables comprehensive and interpretable automated comparisons of synthetic time series evaluation measures. Using 10 diverse datasets, randomness injection, and 13 configurable data transformations, STEB computes indicators for measure reliability and score consistency. It tracks running time, test errors, and features sequential and parallel modes of operation. In our experiments, we determine a ranking of 41 measures from literature and confirm that the choice of upstream time series embedding heavily impacts the final score.
Michael Stenger、Robert Leppich、André Bauer、Samuel Kounev
自动化基础理论计算技术、计算机技术
Michael Stenger,Robert Leppich,André Bauer,Samuel Kounev.STEB: In Search of the Best Evaluation Approach for Synthetic Time Series[EB/OL].(2025-05-27)[2025-06-27].https://arxiv.org/abs/2505.21160.点此复制
评论