SlideShare a Scribd company logo
Quang-Huy Tran
Network Science Lab
Dept. of Artificial Intelligence
The Catholic University of Korea
E-mail: huytran1126@gmail.com
2024-07-22
WaveForM: Graph Enhanced Wavelet
Learning for Long Sequence Forecasting
of Multivariate Time Series
Fuhao Yang et al.
AAAI-2023: The Thirty-Seventh AAAI Conference on Artificial Intelligence
2
OUTLINE
• MOTIVATION
• METHODOLOGY
• EXPERIMENT & RESULT
• CONCLUSION
3
MOTIVATION
• Multiple interconnected streams of data, or Multivariate time series (MTS), have
pervasive presence in real world applications:
o Recorded traffic flows from sensors, weather observations from weather stations, etc.
o Forecasting based on historical MTS observations for making meaningful and accurate application-
wide predictions.
Overview and Limitation
• Challenges:
o Existing work still overlooks long sequence forecasting (LSF) of MTS: Uses a given length of MTS to
predict longer future sequence.
 Long-term MTS are often composed of more entangled temporal patterns than short-term. Overlooking
leads unreliable discoveries of temporal dependencies.
o Transformer-based have proven effectiveness in modeling sequential data.
 Suffer from high computational cost in LSF.
4
INTRODUCTION
• Propose a global graph constructor to extract global information on the interrelationship
among variables in the wavelet domain:
o Preventing training from overfitting.
Contribution
• Propose a DWT-based end-to-end framework:
o Transforms MTS into a wavelet domain for MTS long sequence prediction tasks.
o Capable of fully exploiting the inherent features of MTS in both frequency and time domains.
5
METHODOLOGY
Problem Definition
• Given an MTS
o : N -variate time series.
o : time series of the i-th variable, which consists of sequential recordings at T timestamps.
• Problem: given
o An observation window H for historical time series and a forecasting window P for prediction.
where is learnable parameter set.
o Long sequence MTS forecasting: .
o Goal is to learn a mapping function : predicting the next P timesteps from H
6
METHODOLOGY
Main Architecture
7
METHODOLOGY
Discrete Wavelet Transform Module and Its Inverse Version
• DWT module:
o Transforms an input MTS into its corresponding multi-scale frequency representations with DWT.
o Used to decompose input signals into a set of wavelets: capture frequency and time features of original
signals.
o Each DWT uses a high-pass filter h and a lowpass filter g to decompose a time series signal x into
different resolutions: , is the -th decomposition and .
where represents the length of after decomposing (l − 1) times and s represents the scale
8
METHODOLOGY
Discrete Wavelet Transform Module and Its Inverse Version
• DWT module:
o After l layers of decomposition, for each , outputs a set of coefficients .
o Let represent the layered wavelet coefficients:
where H is length of the input MTS and N is
number of variables.
o After following graph-enhanced modules, output the i-th variable’s coefficients for future P time steps
apply Inverse Discrete Wavelet Transform (IDWT) to reconstruct corresponding sequence in the time
domain
where are the synthesis version of h and g.
9
METHODOLOGY
Global Graph Constructor (GGC)
• After obtaining the wavelet coefficients at different scales:
o Model intends to forecast the coefficient changes overtime in the wavelet domain.
o assume the variables share the same basic interaction structure at different resolutions.
o Using a global graph rather than learning graphs in each GP module also avoids overfitting and saves
memory.
o Apply graph structure learning to learn two embedding representations for each node after assigning
each node/variable as an integer scalar.
 Variable representations obtained from 2 different layers: ,
 Adjacency matrix:
where and is the hyper-parameter for activation function.
10
METHODOLOGY
Graph-Enhanced Prediction Modules
• Given the learnable adjacency matrix:
o Build Graph-enhanced Prediction (GP) modules to exploit the graphical information for predictions.
where is dilated convolution operator and is sigmoid function.
• Dilated Convolution Component:
o Input through stacked 1D dilated convolution: filter wavelet coefficients to incorporate wavelet
information.
o Utilize multiple dilated convolution filters with different kernel sizes to capture respective features for
wavelet coefficients at each level of resolutions.
11
METHODOLOGY
Graph-Enhanced Prediction Modules
o Given the dilated convolution component’s output Z, the process of graph convolution component:
where is representations output from the previous layer and is a hyperparameter controls
the proportion of information maintained from the previous representation.
• Graph Convolution Component:
o aggregate node information with its neighbors’ information to capture global dependencies among
different variable.
o To mitigate over-smoothing of GCN, utilize MixHop layer to capture complex relationshipsof neighbors at
various hops. The process of K-layer MixHop:
12
METHODOLOGY
Graph-Enhanced Prediction Modules
• Skip Connection and Output:
o Improve representational capability by preserving original information.
o Given wavelet coefficients , initialize 2 factors:
where is a 1×1 convolution kernel in GP and is a 1×L convolution
kernel for a skip connection layer.
o Input to pass through a K-layer stacked GP modules:
o The skip-output and other output factor representations of current GP module:
where is a hyperparameter to control the balance.
13
EXPERIMENT AND RESULT
EXPERIMENT SETTINGs
• Dataset:
o Electricity, Traffic, Weather and Solar-Energy.
• Baselines:
o Deep Learning: LSTM, Transformer, Informer [1], and Autoformer [2].
o STGNN: GraphWaveNet [3] and MTGNN [4] .
[1] Zhou, H., Zhang, S., Peng, J., Zhang, S., Li, J., Xiong, H., & Zhang, W. (2021, May). Informer: Beyond efficient transformer for long sequence time-series forecasting. In Proceedings of the AAAI conference on artificial intelligence (Vol. 35, No. 12, pp. 11106-11115).
[2] Wu, H., Xu, J., Wang, J., & Long, M. (2021). Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. Advances in neural information processing systems, 34, 22419-22430.
[3] Wu, Z., Pan, S., Long, G., Jiang, J., & Zhang, C. (2019). Graph wavenet for deep spatial-temporal graph modeling. arXiv preprint arXiv:1906.00121.
[4] Wu, Z., Pan, S., Long, G., Jiang, J., Chang, X., & Zhang, C. (2020, August). Connecting the dots: Multivariate time series forecasting with graph neural networks. In Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery & data mining (pp. 753-763).
• Measurement:
o Mean absolute error (MAE) and mean squared error (MSE).
14
EXPERIMENT AND RESULT
RESULT – Overall Performance
15
EXPERIMENT AND RESULT
RESULT – Ablation Study
16
CONCLUSION
• Proposed WaveForM, a novel framework for long sequence multivariate time series
forecasting
o Use DWT to transform the time-domain series into wavelet-domain coefficients at multiple
resolutions.
o Apply a graph convolution module to model the relationships between multivariate.
Summarization
• The transformed coefficients in the wavelet domain are more capable of describing the
input series from multiple resolutions
o allowing the model to learn fine-grained complex patterns.
[20240722_LabSeminar_Huy]WaveForM: Graph Enhanced Wavelet Learning for Long Sequence Forecasting of Multivariate Time Series.pptx
[20240722_LabSeminar_Huy]WaveForM: Graph Enhanced Wavelet Learning for Long Sequence Forecasting of Multivariate Time Series.pptx

More Related Content

PPTX
[20240923_LabSeminar_Huy]MSGNet: Learning Multi-Scale Inter-Series Correlatio...
PPTX
[20240628_LabSeminar_Huy]ScalableSTGNN.pptx
PPTX
[20240710_LabSeminar_Huy]PDFormer: Propagation Delay-Aware Dynamic Long-Range...
PPTX
[20240729_LabSeminar_Huy]Spatio-Temporal Self-Supervised Learning for Traffic...
PPTX
[20240429_LabSeminar_Huy]Spatio-Temporal Graph Neural Point Process for Traff...
PPTX
[20240819_LabSeminar_Huy]Learning Decomposed Spatial Relations for Multi-Vari...
PPTX
[20240703_LabSeminar_Huy]MakeGNNGreatAgain.pptx
PPTX
[20240408_LabSeminar_Huy]PivotalSTGNN.pptx
[20240923_LabSeminar_Huy]MSGNet: Learning Multi-Scale Inter-Series Correlatio...
[20240628_LabSeminar_Huy]ScalableSTGNN.pptx
[20240710_LabSeminar_Huy]PDFormer: Propagation Delay-Aware Dynamic Long-Range...
[20240729_LabSeminar_Huy]Spatio-Temporal Self-Supervised Learning for Traffic...
[20240429_LabSeminar_Huy]Spatio-Temporal Graph Neural Point Process for Traff...
[20240819_LabSeminar_Huy]Learning Decomposed Spatial Relations for Multi-Vari...
[20240703_LabSeminar_Huy]MakeGNNGreatAgain.pptx
[20240408_LabSeminar_Huy]PivotalSTGNN.pptx

Similar to [20240722_LabSeminar_Huy]WaveForM: Graph Enhanced Wavelet Learning for Long Sequence Forecasting of Multivariate Time Series.pptx (20)

PPTX
[20240712_LabSeminar_Huy]Spatio-Temporal Neural Structural Causal Models for ...
PPTX
[20240930_LabSeminar_Huy]GinAR: An End-To-End Multivariate Time Series Foreca...
PPTX
[20240617_LabSeminar_Huy]Long-term Spatio-Temporal Forecasting via Dynamic Mu...
PPTX
[20240422_LabSeminar_Huy]Taming_Effect.pptx
PDF
The Application of Wavelet Neural Network in the Settlement Monitoring of Subway
PPTX
[20240520_LabSeminar_Huy]DSTAGNN: Dynamic Spatial-Temporal Aware Graph Neural...
PPTX
[20240701_LabSeminar_Huy]TelTrans: Applying Multi-Type Telecom Data to Transp...
PDF
Paper id 26201482
PPTX
[20240812_LabSeminar_Huy]Spatio-Temporal Fusion for Human Action Recognition ...
PPTX
[20240325_LabSeminar_Huy]Spatial-Temporal Fusion Graph Neural Networks for Tr...
PPTX
[20240624_LabSeminar_Huy]Towards Dynamic Spatial-Temporal Graph Learning: A D...
PPTX
[20240415_LabSeminar_Huy]Deciphering Spatio-Temporal Graph Forecasting: A Cau...
PDF
CHANNEL ESTIMATION AND MULTIUSER DETECTION IN ASYNCHRONOUS SATELLITE COMMUNIC...
PDF
Digital Wave Simulation of Lossy Lines for Multi-Gigabit Applications
PPT
Where Next
PPTX
[20240902_LabSeminar_Huy]Dynamic Semantic-Based Spatial Graph Convolution Net...
PDF
DIGITAL WAVE SIMULATION OF LOSSY LINES FOR MULTI-GIGABIT APPLICATION
PPTX
[20240805_LabSeminar_Huy]GPT-ST: Generative Pre-Training of Spatio-Temporal G...
PDF
9517cnc03
PDF
Frequency domain behavior of S-parameters piecewise-linear fitting in a digit...
[20240712_LabSeminar_Huy]Spatio-Temporal Neural Structural Causal Models for ...
[20240930_LabSeminar_Huy]GinAR: An End-To-End Multivariate Time Series Foreca...
[20240617_LabSeminar_Huy]Long-term Spatio-Temporal Forecasting via Dynamic Mu...
[20240422_LabSeminar_Huy]Taming_Effect.pptx
The Application of Wavelet Neural Network in the Settlement Monitoring of Subway
[20240520_LabSeminar_Huy]DSTAGNN: Dynamic Spatial-Temporal Aware Graph Neural...
[20240701_LabSeminar_Huy]TelTrans: Applying Multi-Type Telecom Data to Transp...
Paper id 26201482
[20240812_LabSeminar_Huy]Spatio-Temporal Fusion for Human Action Recognition ...
[20240325_LabSeminar_Huy]Spatial-Temporal Fusion Graph Neural Networks for Tr...
[20240624_LabSeminar_Huy]Towards Dynamic Spatial-Temporal Graph Learning: A D...
[20240415_LabSeminar_Huy]Deciphering Spatio-Temporal Graph Forecasting: A Cau...
CHANNEL ESTIMATION AND MULTIUSER DETECTION IN ASYNCHRONOUS SATELLITE COMMUNIC...
Digital Wave Simulation of Lossy Lines for Multi-Gigabit Applications
Where Next
[20240902_LabSeminar_Huy]Dynamic Semantic-Based Spatial Graph Convolution Net...
DIGITAL WAVE SIMULATION OF LOSSY LINES FOR MULTI-GIGABIT APPLICATION
[20240805_LabSeminar_Huy]GPT-ST: Generative Pre-Training of Spatio-Temporal G...
9517cnc03
Frequency domain behavior of S-parameters piecewise-linear fitting in a digit...
Ad

More from thanhdowork (20)

PPTX
[NS][Lab_Seminar_250811]Imagine and Seek: Improving Composed Image Retrieval ...
PPTX
250811_HW_LabSeminar[Self-Supervised Graph Information Bottleneck for Multivi...
PPTX
250811_Thien_Labseminar[Cluster-GCN].pptx
PPTX
250811_Thuy_Labseminar[BioBRIDGE: BRIDGING BIOMEDICAL FOUNDATION MODELS VIA K...
PPTX
[NS][Lab_Seminar_250728]On the Trade-off between Over-smoothing and Over-squa...
PPTX
250804_HW_LabSeminar[Discrete Curvature Graph Information Bottleneck].pptx
PPTX
250728_Thuy_Labseminar[Knowledge Enhanced Representation Learning for Drug Di...
PPTX
250728_Thuy_Labseminar[Predictive Chemistry Augmented with Text Retrieval].pptx
PPTX
[NS][Lab_Seminar_250728]NeuralWalker.pptx
PPTX
A Novel Shape-Aware Topological Representation for GPR Data with DNN Integrat...
PPTX
250721_Thien_Labseminar[Variational Graph Auto-Encoders].pptx
PPTX
250721_HW_LabSeminar[RingFormer: A Ring-Enhanced Graph Transformer for Organi...
PPTX
250721_Thuy_Labseminar[Thought Propagation: An Analogical Approach to Complex...
PPTX
[NS][Lab_Seminar_250721]On Measuring Long-Range Interactions in Graph Neural ...
PPTX
250714_HW_LabSeminar[Structural Reasoning Improves Molecular Understanding of...
PPTX
[NS][Lab_Seminar_250714]Candidate Set Re-ranking for Composed Image Retrieval...
PPTX
250714_Thuy_Labseminar[BioT5: Enriching Cross-modal Integration in Biology wi...
PPTX
250707_HW_LabSeminar[CHEMICAL-REACTION-AWARE MOLECULE REPRESENTATION LEARNING...
PPTX
[NS][Lab_Seminar_250707]Learning with Noisy Triplet Correspondence for Compos...
PPTX
250707_JW_labseminar[CBAM: Convolutional Block Attention Module].pptx
[NS][Lab_Seminar_250811]Imagine and Seek: Improving Composed Image Retrieval ...
250811_HW_LabSeminar[Self-Supervised Graph Information Bottleneck for Multivi...
250811_Thien_Labseminar[Cluster-GCN].pptx
250811_Thuy_Labseminar[BioBRIDGE: BRIDGING BIOMEDICAL FOUNDATION MODELS VIA K...
[NS][Lab_Seminar_250728]On the Trade-off between Over-smoothing and Over-squa...
250804_HW_LabSeminar[Discrete Curvature Graph Information Bottleneck].pptx
250728_Thuy_Labseminar[Knowledge Enhanced Representation Learning for Drug Di...
250728_Thuy_Labseminar[Predictive Chemistry Augmented with Text Retrieval].pptx
[NS][Lab_Seminar_250728]NeuralWalker.pptx
A Novel Shape-Aware Topological Representation for GPR Data with DNN Integrat...
250721_Thien_Labseminar[Variational Graph Auto-Encoders].pptx
250721_HW_LabSeminar[RingFormer: A Ring-Enhanced Graph Transformer for Organi...
250721_Thuy_Labseminar[Thought Propagation: An Analogical Approach to Complex...
[NS][Lab_Seminar_250721]On Measuring Long-Range Interactions in Graph Neural ...
250714_HW_LabSeminar[Structural Reasoning Improves Molecular Understanding of...
[NS][Lab_Seminar_250714]Candidate Set Re-ranking for Composed Image Retrieval...
250714_Thuy_Labseminar[BioT5: Enriching Cross-modal Integration in Biology wi...
250707_HW_LabSeminar[CHEMICAL-REACTION-AWARE MOLECULE REPRESENTATION LEARNING...
[NS][Lab_Seminar_250707]Learning with Noisy Triplet Correspondence for Compos...
250707_JW_labseminar[CBAM: Convolutional Block Attention Module].pptx
Ad

Recently uploaded (20)

PDF
OBE - B.A.(HON'S) IN INTERIOR ARCHITECTURE -Ar.MOHIUDDIN.pdf
PDF
Practical Manual AGRO-233 Principles and Practices of Natural Farming
PPTX
Cell Structure & Organelles in detailed.
PDF
Paper A Mock Exam 9_ Attempt review.pdf.
PDF
A GUIDE TO GENETICS FOR UNDERGRADUATE MEDICAL STUDENTS
PDF
Module 4: Burden of Disease Tutorial Slides S2 2025
PPTX
Orientation - ARALprogram of Deped to the Parents.pptx
PDF
Microbial disease of the cardiovascular and lymphatic systems
PDF
What if we spent less time fighting change, and more time building what’s rig...
PDF
Computing-Curriculum for Schools in Ghana
PPTX
Radiologic_Anatomy_of_the_Brachial_plexus [final].pptx
PPTX
1st Inaugural Professorial Lecture held on 19th February 2020 (Governance and...
PPTX
Introduction-to-Literarature-and-Literary-Studies-week-Prelim-coverage.pptx
PDF
RTP_AR_KS1_Tutor's Guide_English [FOR REPRODUCTION].pdf
PDF
Black Hat USA 2025 - Micro ICS Summit - ICS/OT Threat Landscape
PDF
Classroom Observation Tools for Teachers
PPTX
school management -TNTEU- B.Ed., Semester II Unit 1.pptx
PDF
Yogi Goddess Pres Conference Studio Updates
PDF
01-Introduction-to-Information-Management.pdf
PDF
grade 11-chemistry_fetena_net_5883.pdf teacher guide for all student
OBE - B.A.(HON'S) IN INTERIOR ARCHITECTURE -Ar.MOHIUDDIN.pdf
Practical Manual AGRO-233 Principles and Practices of Natural Farming
Cell Structure & Organelles in detailed.
Paper A Mock Exam 9_ Attempt review.pdf.
A GUIDE TO GENETICS FOR UNDERGRADUATE MEDICAL STUDENTS
Module 4: Burden of Disease Tutorial Slides S2 2025
Orientation - ARALprogram of Deped to the Parents.pptx
Microbial disease of the cardiovascular and lymphatic systems
What if we spent less time fighting change, and more time building what’s rig...
Computing-Curriculum for Schools in Ghana
Radiologic_Anatomy_of_the_Brachial_plexus [final].pptx
1st Inaugural Professorial Lecture held on 19th February 2020 (Governance and...
Introduction-to-Literarature-and-Literary-Studies-week-Prelim-coverage.pptx
RTP_AR_KS1_Tutor's Guide_English [FOR REPRODUCTION].pdf
Black Hat USA 2025 - Micro ICS Summit - ICS/OT Threat Landscape
Classroom Observation Tools for Teachers
school management -TNTEU- B.Ed., Semester II Unit 1.pptx
Yogi Goddess Pres Conference Studio Updates
01-Introduction-to-Information-Management.pdf
grade 11-chemistry_fetena_net_5883.pdf teacher guide for all student

[20240722_LabSeminar_Huy]WaveForM: Graph Enhanced Wavelet Learning for Long Sequence Forecasting of Multivariate Time Series.pptx

  • 1. Quang-Huy Tran Network Science Lab Dept. of Artificial Intelligence The Catholic University of Korea E-mail: [email protected] 2024-07-22 WaveForM: Graph Enhanced Wavelet Learning for Long Sequence Forecasting of Multivariate Time Series Fuhao Yang et al. AAAI-2023: The Thirty-Seventh AAAI Conference on Artificial Intelligence
  • 2. 2 OUTLINE • MOTIVATION • METHODOLOGY • EXPERIMENT & RESULT • CONCLUSION
  • 3. 3 MOTIVATION • Multiple interconnected streams of data, or Multivariate time series (MTS), have pervasive presence in real world applications: o Recorded traffic flows from sensors, weather observations from weather stations, etc. o Forecasting based on historical MTS observations for making meaningful and accurate application- wide predictions. Overview and Limitation • Challenges: o Existing work still overlooks long sequence forecasting (LSF) of MTS: Uses a given length of MTS to predict longer future sequence.  Long-term MTS are often composed of more entangled temporal patterns than short-term. Overlooking leads unreliable discoveries of temporal dependencies. o Transformer-based have proven effectiveness in modeling sequential data.  Suffer from high computational cost in LSF.
  • 4. 4 INTRODUCTION • Propose a global graph constructor to extract global information on the interrelationship among variables in the wavelet domain: o Preventing training from overfitting. Contribution • Propose a DWT-based end-to-end framework: o Transforms MTS into a wavelet domain for MTS long sequence prediction tasks. o Capable of fully exploiting the inherent features of MTS in both frequency and time domains.
  • 5. 5 METHODOLOGY Problem Definition • Given an MTS o : N -variate time series. o : time series of the i-th variable, which consists of sequential recordings at T timestamps. • Problem: given o An observation window H for historical time series and a forecasting window P for prediction. where is learnable parameter set. o Long sequence MTS forecasting: . o Goal is to learn a mapping function : predicting the next P timesteps from H
  • 7. 7 METHODOLOGY Discrete Wavelet Transform Module and Its Inverse Version • DWT module: o Transforms an input MTS into its corresponding multi-scale frequency representations with DWT. o Used to decompose input signals into a set of wavelets: capture frequency and time features of original signals. o Each DWT uses a high-pass filter h and a lowpass filter g to decompose a time series signal x into different resolutions: , is the -th decomposition and . where represents the length of after decomposing (l − 1) times and s represents the scale
  • 8. 8 METHODOLOGY Discrete Wavelet Transform Module and Its Inverse Version • DWT module: o After l layers of decomposition, for each , outputs a set of coefficients . o Let represent the layered wavelet coefficients: where H is length of the input MTS and N is number of variables. o After following graph-enhanced modules, output the i-th variable’s coefficients for future P time steps apply Inverse Discrete Wavelet Transform (IDWT) to reconstruct corresponding sequence in the time domain where are the synthesis version of h and g.
  • 9. 9 METHODOLOGY Global Graph Constructor (GGC) • After obtaining the wavelet coefficients at different scales: o Model intends to forecast the coefficient changes overtime in the wavelet domain. o assume the variables share the same basic interaction structure at different resolutions. o Using a global graph rather than learning graphs in each GP module also avoids overfitting and saves memory. o Apply graph structure learning to learn two embedding representations for each node after assigning each node/variable as an integer scalar.  Variable representations obtained from 2 different layers: ,  Adjacency matrix: where and is the hyper-parameter for activation function.
  • 10. 10 METHODOLOGY Graph-Enhanced Prediction Modules • Given the learnable adjacency matrix: o Build Graph-enhanced Prediction (GP) modules to exploit the graphical information for predictions. where is dilated convolution operator and is sigmoid function. • Dilated Convolution Component: o Input through stacked 1D dilated convolution: filter wavelet coefficients to incorporate wavelet information. o Utilize multiple dilated convolution filters with different kernel sizes to capture respective features for wavelet coefficients at each level of resolutions.
  • 11. 11 METHODOLOGY Graph-Enhanced Prediction Modules o Given the dilated convolution component’s output Z, the process of graph convolution component: where is representations output from the previous layer and is a hyperparameter controls the proportion of information maintained from the previous representation. • Graph Convolution Component: o aggregate node information with its neighbors’ information to capture global dependencies among different variable. o To mitigate over-smoothing of GCN, utilize MixHop layer to capture complex relationshipsof neighbors at various hops. The process of K-layer MixHop:
  • 12. 12 METHODOLOGY Graph-Enhanced Prediction Modules • Skip Connection and Output: o Improve representational capability by preserving original information. o Given wavelet coefficients , initialize 2 factors: where is a 1×1 convolution kernel in GP and is a 1×L convolution kernel for a skip connection layer. o Input to pass through a K-layer stacked GP modules: o The skip-output and other output factor representations of current GP module: where is a hyperparameter to control the balance.
  • 13. 13 EXPERIMENT AND RESULT EXPERIMENT SETTINGs • Dataset: o Electricity, Traffic, Weather and Solar-Energy. • Baselines: o Deep Learning: LSTM, Transformer, Informer [1], and Autoformer [2]. o STGNN: GraphWaveNet [3] and MTGNN [4] . [1] Zhou, H., Zhang, S., Peng, J., Zhang, S., Li, J., Xiong, H., & Zhang, W. (2021, May). Informer: Beyond efficient transformer for long sequence time-series forecasting. In Proceedings of the AAAI conference on artificial intelligence (Vol. 35, No. 12, pp. 11106-11115). [2] Wu, H., Xu, J., Wang, J., & Long, M. (2021). Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. Advances in neural information processing systems, 34, 22419-22430. [3] Wu, Z., Pan, S., Long, G., Jiang, J., & Zhang, C. (2019). Graph wavenet for deep spatial-temporal graph modeling. arXiv preprint arXiv:1906.00121. [4] Wu, Z., Pan, S., Long, G., Jiang, J., Chang, X., & Zhang, C. (2020, August). Connecting the dots: Multivariate time series forecasting with graph neural networks. In Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery & data mining (pp. 753-763). • Measurement: o Mean absolute error (MAE) and mean squared error (MSE).
  • 14. 14 EXPERIMENT AND RESULT RESULT – Overall Performance
  • 15. 15 EXPERIMENT AND RESULT RESULT – Ablation Study
  • 16. 16 CONCLUSION • Proposed WaveForM, a novel framework for long sequence multivariate time series forecasting o Use DWT to transform the time-domain series into wavelet-domain coefficients at multiple resolutions. o Apply a graph convolution module to model the relationships between multivariate. Summarization • The transformed coefficients in the wavelet domain are more capable of describing the input series from multiple resolutions o allowing the model to learn fine-grained complex patterns.

Editor's Notes

  • #6: WAVEFORM is a multi-resolution analysis (MRA) model based on discrete wavelet transforms, and it forecasts MTS in the wavelet domain. 3 main components: discrete wavelet transform (DWT) module, global graph constructor (GGC), and graph-enhanced prediction (GP) modules