Difference between revisions of "Applications of Reinforcement Learning in Finance"

From EU COST Fin-AI
Jump to navigation Jump to search
(Created page with "== Details == * Authors: Frensi Zejnullahu, Maurice Moser * Title: Applications of Reinforcement Learning in Finance - Trading with a double Deep Q-Network. * Supervisior: Pr...")
 
 
(10 intermediate revisions by the same user not shown)
Line 1: Line 1:
 
== Details ==  
 
== Details ==  
 
* Authors: Frensi Zejnullahu, Maurice Moser
 
* Authors: Frensi Zejnullahu, Maurice Moser
* Title: Applications of Reinforcement Learning in Finance - Trading with a double Deep Q-Network.
+
* Title: Applications of Reinforcement Learning in Finance - Trading with a Double Deep Q-Network
 
* Supervisior: Prof. Dr. Jörg Osterrieder
 
* Supervisior: Prof. Dr. Jörg Osterrieder
 
* Degree: Bachelor of Science
 
* Degree: Bachelor of Science
Line 10: Line 10:
 
== Summary ==  
 
== Summary ==  
  
Examining if additional input features in a feedforward neural network improve the prediction accuracy for daily Bitcoin price movements and if the Bitcoin market is efficient as per the weak form of the efficient market hypothesis.
+
Analyzing the performance and behavior of an RL agent with respect to additional features and environmental conditions. These include increasing features with other assets, trading costs, and pre- and post-crisis testing.
  
 
== Abstract ==  
 
== Abstract ==  
Line 16: Line 16:
  
 
== Important links ==  
 
== Important links ==  
* [https://www.overleaf.com/read/pcpfrwsvyhqn Overleaf]  
+
* [https://arxiv.org/abs/2206.14267 Arxiv]
* [https://github.zhaw.ch/zejnufre/RL_DDQN.git GitHub]
+
* [https://github.com/frensizej/RL_DDQN.git GitHub]
* [https://drive.google.com/drive/folders/1E2sGajWczI4jgBBEKfqNORDO1oWQv3M2 Google Drive with paper and links]
+
* [https://www.overleaf.com/read/pcpfrwsvyhqn Overleaf]
 +
* [https://www.overleaf.com/read/sbydjmfrkksf Presentation Overleaf]
 +
* [https://drive.google.com/drive/folders/1bksxRmJpy037pzALmeHiKWD7e8eBWBbh?usp=sharing Google Drive with paper and links]
  
 
== Data ==
 
== Data ==
* Bitcoin (BTC-USD)
+
* S&P500 Future (ESc1)
* S&P 500 (ˆGSPC)
+
* Russell 2000 ETF (IWM)
* Russell 2000 (ˆRUT)
+
* Gold Future (GCc1)
* EUR/USD (EURUSD=X)
+
* WTI Crude Oil Future (CLc1)
* 10 Year Treasury Note Yield (^TNX)
 
* PHLX Gold/Silver Sector Index (^XAU)
 
  
Data source: [https://finance.yahoo.com/ Yahoo Finance]
+
Data source: [https://www.refinitiv.com/ Refinitiv]
  
 
== Contact ==  
 
== Contact ==  
 
* [https://www.linkedin.com/in/joergosterrieder/ Prof. Dr. Jörg Osterrieder]
 
* [https://www.linkedin.com/in/joergosterrieder/ Prof. Dr. Jörg Osterrieder]
* [mailto:kraehmik@students.zhaw.ch Mike Krähenbühl]
+
* [mailto:zejnufre@students.zhaw.ch Frensi Zejnullahu]
 +
* [mailto:mosermau@students.zhaw.ch Maurice Moser]

Latest revision as of 07:22, 7 July 2022

Details

  • Authors: Frensi Zejnullahu, Maurice Moser
  • Title: Applications of Reinforcement Learning in Finance - Trading with a Double Deep Q-Network
  • Supervisior: Prof. Dr. Jörg Osterrieder
  • Degree: Bachelor of Science
  • University: Zurich University of Applied Sciences
  • Year: 2022
  • Status: Working Paper

Summary

Analyzing the performance and behavior of an RL agent with respect to additional features and environmental conditions. These include increasing features with other assets, trading costs, and pre- and post-crisis testing.

Abstract

This paper presents a Double Deep Q-Network algorithm for trading single assets, namely the E-mini S&P 500 continuous futures contract. We use a proven setup as the foundation for our environment with multiple extensions. The features of our trading agent are constantly being expanded to include additional assets such as commodities, resulting in four models. We also respond to environmental conditions, including costs and crises. Our trading agent is first trained for a specific time period and tested on new data and compared with the long-and-hold strategy as a benchmark (market). We analyze the differences between the various models and the in-sample/out-of-sample performance with respect to the environment. The experimental results show that the trading agent follows an appropriate behavior. It can adjust its policy to different circumstances, such as more extensive use of the neutral position when trading costs are present. Furthermore, the net asset value exceeded that of the benchmark, and the agent outperformed the market in the test set. We provide initial insights into the behavior of an agent in a financial domain using a DDQN algorithm. The results of this study can be used for further development.

Important links

Data

  • S&P500 Future (ESc1)
  • Russell 2000 ETF (IWM)
  • Gold Future (GCc1)
  • WTI Crude Oil Future (CLc1)

Data source: Refinitiv

Contact