Go to file
2024-04-19 09:38:06 +08:00
Classification Update TSLANet_classification.py 2024-04-19 09:37:50 +08:00
Forecasting Update TSLANet_Forecasting.py 2024-04-19 09:38:06 +08:00
misc Add files via upload 2024-04-18 15:55:35 +08:00
LICENSE Initial commit 2024-04-12 15:42:38 +08:00
README.md Update README.md 2024-04-18 15:56:04 +08:00

TSLANet: Rethinking Transformers for Time Series Representation Learning [Paper] [Cite]

by: Emadeldeen Eldele, Mohamed Ragab, Zhenghua Chen, Min Wu,and Xiaoli Li

Abstract

Time series data, characterized by its intrinsic long and short-range dependencies, poses a unique challenge across analytical applications. While Transformer-based models excel at capturing long-range dependencies, they face limitations in noise sensitivity, computational efficiency, and overfitting with smaller datasets. In response, we introduce a novel Time Series Lightweight Adaptive Network (TSLANet), as a universal convolutional model for diverse time series tasks. Specifically, we propose an Adaptive Spectral Block, harnessing Fourier analysis to enhance feature representation and to capture both long-term and short-term interactions while mitigating noise via adaptive thresholding. Additionally, we introduce an Interactive Convolution Block and leverage self-supervised learning to refine the capacity of TSLANet for decoding complex temporal patterns and improve its robustness on different datasets. Our comprehensive experiments demonstrate that TSLANet outperforms state-of-the-art models in various tasks spanning classification, forecasting, and anomaly detection, showcasing its resilience and adaptability across a spectrum of noise levels and data sizes.

Citation

If you found this work useful for you, please consider citing it.

@article{tslanet,
  title     = {TSLANet: Rethinking Transformers for Time Series Representation Learning},
  author    = {Eldele, Emadeldeen and Ragab, Mohamed and Chen, Zhenghua and Wu, Min and Li, Xiaoli},
  journal   = {arXiv preprint arXiv:2404.08472},
  year      = {2024},
}

Acknowledgements

The codes in this repository are inspired by the following: