aikunyi / time_series_frequency

Neural Time Series Analysis with Fourier Transform: A Survey

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Neural Time Series Analysis with Fourier Transform: A Survey

In this paper, we provide a comprehensive review of studies on neural time series analysis with Fourier transform. We aim to systematically investigate and summarize the latest research progress. Accordingly, we propose a novel taxonomy to categorize existing neural time series analysis methods from four perspectives, including characteristics, usage paradigms, network design, and applications. We also share some new research directions in this vibrant area.

Taxonomy

[taxonomy]

Characteristics

In this section, we analyze the advantaged characteristics of Fourier Transform, including decomposition, global view, sparse representation, and efficiency.

Decomposition

Global view

Sparse representation

Efficiency

[relationship]

Usage paradigms

In this section, we systematically summarize and discuss the research categorization and progress in terms of how to utilize the Fourier transform to enhance time series analysis, i.e., usage paradigms.

Feature engineering

ATFN [paper] proposes a frequency-domain block to capture dynamic and complicated periodic patterns of time series data, and integrates deep learning networks with frequency patterns.

TFAD [paper] utilizes a frequency domain analysis branch to detect complex pattern anomalies, e.g., periodic anomaly.

CoST [paper] learns the trend representations in the time domain, whereas the seasonal representations are learned by a Fourier layer in the frequency domain.

SFM [paper] explicitly decomposes trading patterns into various frequency components and each component models a particular frequency of latent trading pattern underlying the fluctuation of stock price.

mWDN [paper] proposes a wavelet-based neural network structure for building frequency-aware deep learning models for time series analysis.

RobustPeriod [paper] applies maximal overlap discrete wavelet transform to decouple time series into multiple levels of wavelet coefficients and then detect single periodicity at each level.

FEDformer [paper] combines Fourier analysis with the Transformer which helps Transformer better capture global properties of time series.

FFC [paper] harnesses the Fourier spectral theory and designs an operation unit to leverage frequency information for enlarging the receptive field of vanilla convolutions.

Compression

FiLM [paper] view time series forecasting from the sequence compression perspective and applies Fourier analysis to keep the part of the representation related to low-frequency Fourier components to remove the impact of noises.

FcaNet [paper] generalizes the compression of the channel attention mechanism in the DCT frequency domain and proposes a multi-spectral channel attention for frequency components selection.

Data Augmentation

CoST [paper] incorporates a novel frequency domain contrastive loss which encourages discriminative seasonal representations and side steps the issue of determining the period of seasonal patterns present in the time series data.
BTSF [paper] fuses the temporal and spectral features to enhance the discriminativity and expressiveness of the representations.
TF-C [paper] introduces frequency domain augmentations that it directly perturbs the frequency spectrum.

Fourier neural operator learning

FNO [paper] parameterize the integral kernel directly in the Fourier space, allowing for an expressive and efficient architecture for partial differential equations.
AFNO [paper] frames token mixing as operator learning and proposes an efficient token mixer that learns to mix in the Fourier domain.
FEDformer [paper] proposes Fourier enhanced blocks and Wavelet enhanced blocks to capture important structures in time series through frequency domain mapping.
EV-FGN [paper] reformulates the graph convolution operator in the frequency domain and efficiently computes graph convolutions over a supra-graph which represents non-static correlations between any two variables at any two timestamps.

Network design

complex-value

In fact, there are mainly two approaches for performing neural networks on complex value. One approach is to treat each part of complex value as a feature and then feed them to neural networks, respectively. The other one is to conduct complex multiplication in the frequency domain directly.

real-value

Except for capturing frequency patterns, in contrast to other network design, one main purpose of network design for frequency-based models is the frequency component selection to decide which component is discriminative or critical.

Applications

Forecasting

Stock Price Prediction via Discovering Multi-Frequency Trading Patterns, In KDD, 2017. [paper]
Spectral temporal graph neural network for multivariate time-series forecasting. In NeurIPS, 2020. [paper] [code]
Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. In NeurIPS, 2020. [paper] [code]
FEDformer: Frequency enhanced decomposed transformer for long-term series forecasting. In ICML, 2022. [paper] [code]
Cost: Contrastive learning of disentangled seasonal-trend representations for time series forecasting. In ICLR, 2022. [paper] [code]
Film: Frequency improved legendre memory model for long-term time series forecasting. In NIPS, 2022. [paper] [code]
Edge-Varying Fourier Graph Networks for Multivariate Time Series Forecasting. In arXiv, 2022. [paper]

Anomaly Detection

Time-series anomaly detection service at microsoft. In KDD, 2019. [paper]
Robusttad: Robust time series anomaly detection via decomposition and convolutional neural networks. In arXiv, 2020. [paper]
Fast and accurate partial fourier transform for time series data. In KDD, 2021. [paper] [code]
TFAD: A decomposition time series anomaly detection architecture with time-frequency analysis. In CIKM, 2022. [paper] [code]

Classification

Multilevel wavelet decomposition network for interpretable time series analysis. In KDD, 2018. [paper]
Learning filter widths of spectral decompositions with wavelets. In NeurIPS, 2018. [paper] [code]
Unsupervised time-series representation learning with iterative bilinear temporal-spectral fusion. In ICML, 2022. [paper]
Self-Supervised Contrastive Pre-Training For Time Series via Time-Frequency Consistency, in NIPS 2022, [paper] [code]

Discussion for Future Opportunities

Leveraging New Orthogonal Transform Technology

In future, it would be a promising direction to incorporate more new orthogonal transform technologies for deep learning in time series analysis, such as orthogonal polynomials, DCT, and FrFT.

Combination of Fourier Transform with Deep Learning

In future, combination of Fourier transform with deep learning deserves further investigating.

Combination of Learning in the Time and Frequency Domain

More time-frequency representation learning methods are required in the future.

About

Neural Time Series Analysis with Fourier Transform: A Survey