Submitted: 25 March, 2021
Access to the
video
and to the
SSRN paper
Authors:
Eric Benhamou,
David Saltiel,
Serge Tabachnik,
Sui Kai Wong
and
François Chareyron
Abstract:
Model Free Reinforcement Learning has achieved great results in stable
environments but has not been able sofar to generalize well in regime changing environments like financial markets.
In contrast, model based RL are able to capture some fundamental and dynamical concepts of the environment but suffer
from cognitive bias.
In this work, we propose to combine the best of the two approaches by selecting thanks to Model
free Deep Reinforcement Learning various model based approaches. Using not only past performance and volatility, we
include additional contextual information to account for implicit regime changes like macro and risk appetite signals .
We also adapt traditional RL methods to take into account that in real life training takes always place in the past.
Hence we cannot use future information in our training data set as implied by K-fold cross validation. Building on traditional
statistical methods, we introduce "walk-forward analysis", which is defined by successive training and testing based on expanding
periods, to assert the robustness of the resulting agent. Last but not least, we present the concept of statistical difference
significance based on a two-tailed T-test, to highlight the ways in which our models differ from more traditional ones.
Our experimental results show that our approach outperforms traditional financial baselines portfolio models like Markowitz in
almost all evaluation metrics commonly used in financial mathematics, namely net performance, Sharpe ratio, Sortino, maximum drawdown,
maximum drawdown over volatility.
Read more