Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

ETSformer: Exponential Smoothing Transformers for Time-series Forecasting (2202.01381v2)

Published 3 Feb 2022 in cs.LG

Abstract: Transformers have been actively studied for time-series forecasting in recent years. While often showing promising results in various scenarios, traditional Transformers are not designed to fully exploit the characteristics of time-series data and thus suffer some fundamental limitations, e.g., they generally lack of decomposition capability and interpretability, and are neither effective nor efficient for long-term forecasting. In this paper, we propose ETSFormer, a novel time-series Transformer architecture, which exploits the principle of exponential smoothing in improving Transformers for time-series forecasting. In particular, inspired by the classical exponential smoothing methods in time-series forecasting, we propose the novel exponential smoothing attention (ESA) and frequency attention (FA) to replace the self-attention mechanism in vanilla Transformers, thus improving both accuracy and efficiency. Based on these, we redesign the Transformer architecture with modular decomposition blocks such that it can learn to decompose the time-series data into interpretable time-series components such as level, growth and seasonality. Extensive experiments on various time-series benchmarks validate the efficacy and advantages of the proposed method. Code is available at https://github.com/salesforce/ETSformer.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Gerald Woo (11 papers)
  2. Chenghao Liu (61 papers)
  3. Doyen Sahoo (47 papers)
  4. Akshat Kumar (29 papers)
  5. Steven Hoi (38 papers)
Citations (125)

Summary

Exponential Smoothing Transformers for Time-Series Forecasting

The paper introduces ETSformer, a novel Transformer-based architecture specifically designed for time-series forecasting tasks. Traditional Transformer models, known for their prowess in areas like NLP and CV, are not inherently optimized for time-series data, which often require different considerations such as decomposability, interpretability, and efficiency, particularly for long-term forecasting scenarios. ETSformer aims to bridge this gap by incorporating principles from exponential smoothing into a Transformer framework.

Key Contributions of ETSformer:

  1. Exponential Smoothing Attention (ESA) and Frequency Attention (FA):
    • ETSformer replaces the traditional self-attention mechanism found in vanilla Transformers with ESA and FA. ESA applies an exponential time decay to attention scoring, ensuring that more recent observations are weighted more heavily, mimicking classical exponential smoothing methods. This approach aligns more closely with the inherent nature of time-series data, where recent data points typically have greater relevance.
    • FA leverages the discrete Fourier transform to identify dominant frequencies in the data, capturing seasonal patterns effectively.
  2. Interpretable Modular Decomposition:
    • The architecture is modularly designed to decompose the input data into interpretable components such as level, growth, and seasonality. This decomposition is achieved through layer-wise blocks, enhancing both the explainability and robustness of the model. By doing so, ETSformer can generate forecasts that can be understood and validated by human experts.
  3. Efficiency and Effectiveness:
    • The proposed model achieves computational efficiency with an O(LlogL)O(L \log L) complexity for a time-series of length LL, thanks to the novel attention mechanisms. This performance allows it to handle long-term forecasting tasks more efficiently than traditional Transformers, often limited by quadratic complexity concerns.

Empirical Evaluation:

ETSformer has been validated through extensive experimentation across a variety of time-series benchmarks, achieving state-of-the-art performance in both multivariate and univariate settings. The results demonstrate its superior accuracy and efficacy, particularly when compared to other Transformer-based models and classical statistical approaches.

Implications and Speculative Future Developments:

The introduction of ETSformer marks a significant development in time-series forecasting methodologies. It brings together the strengths of deep learning architectures and the domain-specific insights of classical time-series methods. Practically, this model could find applications across industries reliant on accurate forecasting, such as finance, meteorology, and supply chain management.

Theoretically, ETSformer opens avenues for further research into hybrid models that incorporate domain-specific knowledge into generic deep learning frameworks, potentially extending beyond time-series to other structured data tasks. Future work could explore integrations with additional covariates, refinement of decomposition strategies, and enhancements to its learning dynamics to handle diverse time-series data characteristics.

In summary, ETSformer represents a substantial advancement in time-series forecasting, integrating the interpretability and structure of classical approaches with the flexibility and power of modern Transformer designs. Its development underscores an evolving trend in AI research focused on blending domain expertise with deep learning to tackle complex real-world challenges effectively.

Youtube Logo Streamline Icon: https://streamlinehq.com