Skip to content

rancharles/clvsa

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

17 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

CLVSA

We aim to reproduce the core findings of CLVSA: A Convolutional LSTM-Based Variational Sequence-to-Sequence Model with Attention for Predicting Trends of Financial Markets (Wang et al., 2019).

This repository provides an open-source PyTorch implementation of our interpretation of the model.

Getting Started

Install packages with

pip install -r requirements.txt

Note that you may have to change the PyTorch installation if you want to use CUDA.

Set your HF_TOKEN in a .env file in the root directory to speed up the dataset download, which could take a few minutes. Note that it will be cached locally after the first download.

To recieve a summary of the dataset, run python train.py --dataset_only.

To train the model, run python train.py. You can run it with optional flags to tweak training parameters. Run python train.py --help for details.

Differences from the paper

  • We used scaled attention while the paper did not scale attention.
  • We used log-relative OHLCV feature transforms to avoid regime drifting of validation and testing sets.
  • We opt for AdamW as the optimizer instead of Adam
  • Our model adds feature flags for the seq2seq inter-attention and self-attention mechanisms. We also implement the option to use seperate row-specific kernels for the convLSTM rather than row-shared kernels.

References

Jia Wang, Tong Sun, Benyuan Liu, Yu Cao, and Hongwei Zhu. CLVSA: A Convolutional LSTM-Based Variational Sequence-to-Sequence Model with Attention for Predicting Trends of Financial Markets. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, pages 3705–3711, 2019.

About

Variational convLSTM seq2seq model with attention

Resources

Stars

Watchers

Forks

Contributors

Languages