In this paper, we provide a comprehensive review of studies on neural time series analysis with Fourier transform. We aim to systematically investigate and summarize the latest research progress. Accordingly, we propose a novel taxonomy to categorize existing neural time series analysis methods from four perspectives, including characteristics, usage paradigms, network design, and applications. We also share some new research directions in this vibrant area.
In this section, we analyze the advantaged characteristics of Fourier Transform, including decomposition, global view, sparse representation, and efficiency.
In this section, we systematically summarize and discuss the research categorization and progress in terms of how to utilize the Fourier transform to enhance time series analysis, i.e., usage paradigms.
ATFN [paper] proposes a frequency-domain block to capture dynamic and complicated periodic patterns of time series data, and integrates deep learning networks with frequency patterns.
TFAD [paper] utilizes a frequency domain analysis branch to detect complex pattern anomalies, e.g., periodic anomaly.
CoST [paper] learns the trend representations in the time domain, whereas the seasonal representations are learned by a Fourier layer in the frequency domain.
SFM [paper] explicitly decomposes trading patterns into various frequency components and each component models a particular frequency of latent trading pattern underlying the fluctuation of stock price.
mWDN [paper] proposes a wavelet-based neural network structure for building frequency-aware deep learning models for time series analysis.
RobustPeriod [paper] applies maximal overlap discrete wavelet transform to decouple time series into multiple levels of wavelet coefficients and then detect single periodicity at each level.
FEDformer [paper] combines Fourier analysis with the Transformer which helps Transformer better capture global properties of time series.
FFC [paper] harnesses the Fourier spectral theory and designs an operation unit to leverage frequency information for enlarging the receptive field of vanilla convolutions.
FiLM [paper] view time series forecasting from the sequence compression perspective and applies Fourier analysis to keep the part of the representation related to low-frequency Fourier components to remove the impact of noises.
FcaNet [paper] generalizes the compression of the channel attention mechanism in the DCT frequency domain and proposes a multi-spectral channel attention for frequency components selection.
CoST [paper] incorporates a novel frequency domain contrastive loss which encourages discriminative seasonal representations and side steps the issue of determining the period of seasonal patterns present in the time series data.
BTSF [paper] fuses the temporal and spectral features to enhance the discriminativity and expressiveness of the representations.
TF-C [paper] introduces frequency domain augmentations that it directly perturbs the frequency spectrum.
FNO [paper] parameterize the integral kernel directly in the Fourier space, allowing for an expressive and efficient architecture for partial differential equations.
AFNO [paper] frames token mixing as operator learning and proposes an efficient token mixer that learns to mix in the Fourier domain.
FEDformer [paper] proposes Fourier enhanced blocks and Wavelet enhanced blocks to capture important structures in time series through frequency domain mapping.
EV-FGN [paper] reformulates the graph convolution operator in the frequency domain and efficiently computes graph convolutions over a supra-graph which represents non-static correlations between any two variables at any two timestamps.
In fact, there are mainly two approaches for performing neural networks on complex value. One approach is to treat each part of complex value as a feature and then feed them to neural networks, respectively. The other one is to conduct complex multiplication in the frequency domain directly.
Except for capturing frequency patterns, in contrast to other network design, one main purpose of network design for frequency-based models is the frequency component selection to decide which component is discriminative or critical.
Stock Price Prediction via Discovering Multi-Frequency Trading Patterns, In KDD, 2017. [paper]
Spectral temporal graph neural network for multivariate time-series forecasting. In NeurIPS, 2020. [paper] [code]
Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. In NeurIPS, 2020. [paper] [code]
FEDformer: Frequency enhanced decomposed transformer for long-term series forecasting. In ICML, 2022. [paper] [code]
Cost: Contrastive learning of disentangled seasonal-trend representations for time series forecasting. In ICLR, 2022. [paper] [code]
Film: Frequency improved legendre memory model for long-term time series forecasting. In NIPS, 2022. [paper] [code]
Edge-Varying Fourier Graph Networks for Multivariate Time Series Forecasting. In arXiv, 2022. [paper]
Time-series anomaly detection service at microsoft. In KDD, 2019. [paper]
Robusttad: Robust time series anomaly detection via decomposition and convolutional neural networks. In arXiv, 2020. [paper]
Fast and accurate partial fourier transform for time series data. In KDD, 2021. [paper] [code]
TFAD: A decomposition time series anomaly detection architecture with time-frequency analysis. In CIKM, 2022. [paper] [code]
Multilevel wavelet decomposition network for interpretable time series analysis. In KDD, 2018. [paper]
Learning filter widths of spectral decompositions with wavelets. In NeurIPS, 2018. [paper] [code]
Unsupervised time-series representation learning with iterative bilinear temporal-spectral fusion. In ICML, 2022. [paper]
Self-Supervised Contrastive Pre-Training For Time Series via Time-Frequency Consistency, in NIPS 2022, [paper] [code]
In future, it would be a promising direction to incorporate more new orthogonal transform technologies for deep learning in time series analysis, such as orthogonal polynomials, DCT, and FrFT.
In future, combination of Fourier transform with deep learning deserves further investigating.
More time-frequency representation learning methods are required in the future.