Affiliation:
1. Université Paris-Saclay, CentraleSupélec, Laboratoire MICS, 91190 Gif-sur-Yvette, France
Abstract
We extend recurrent neural networks to include several flexible timescales for each dimension of their output, which mechanically improves their abilities to account for processes with long memory or highly disparate timescales. We compare the ability of vanilla and extended long short-term memory networks (LSTMs) to predict the intraday volatility of a collection of equity indices known to have a long memory. Generally, the number of epochs needed to train the extended LSTMs is divided by about two, while the variation in validation and test losses among models with the same hyperparameters is much smaller. We also show that the single model with the smallest validation loss systemically outperforms rough volatility predictions for the average intraday volatility of equity indices by about 20% when trained and tested on a dataset with multiple time series.
Funder
the Association Nationale de la Recherche et de la Technologie
Reference28 articles.
1. Designing realized kernels to measure the ex post variation of equity prices in the presence of noise;Hansen;Econometrica,2008
2. Optimal approximations of power laws with exponentials: Application to volatility models with long memory;Bochud;Quantitative Finance,2007
3. Analyzing and modeling 1 + 1d markets;Challet;Physica A: Statistical Mechanics and Its Applications,2001
4. Cho, Kyunghyun, Merrienboer, Bart van, Gulcehre, Caglar, Bahdanau, Dzmitry, Bougares, Fethi, Schwenk, Holger, and Bengio, Yoshua (, January October). Learning phrase representations using rnn encoder–decoder for statistical machine translation. Paper Presented at the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar.
5. Empirical properties of asset returns: Stylized facts and statistical issues;Cont;Quantitative Finance,2001