Deep Reinforcement Learning for Market Making in Corporate Bonds: Beating the Curse of Dimensionality
Autor: | Olivier Guéant, Iuliia Manziuk |
---|---|
Přispěvatelé: | Centre d'économie de la Sorbonne (CES), Université Paris 1 Panthéon-Sorbonne (UP1)-Centre National de la Recherche Scientifique (CNRS) |
Jazyk: | angličtina |
Rok vydání: | 2019 |
Předmět: |
Stochastic control
050208 finance Financial economics Applied Mathematics Bond 05 social sciences TheoryofComputation_GENERAL Actor-critic algorithms [SHS.ECO]Humanities and Social Sciences/Economics and Finance 01 natural sciences Market maker Corporate bond 010104 statistics & probability Stochastic optimal control ComputerApplications_MISCELLANEOUS 0502 economics and business Reinforcement learning Economics Asset (economics) 0101 mathematics Bid price Finance Curse of dimensionality Market making |
Zdroj: | Applied Mathematical Finance Applied Mathematical Finance, Taylor & Francis (Routledge): SSH Titles, 2019, 26 (5), pp.387-452. ⟨10.1080/1350486X.2020.1714455⟩ |
ISSN: | 1350-486X 1466-4313 |
DOI: | 10.1080/1350486X.2020.1714455⟩ |
Popis: | International audience; In corporate bond markets, which are mainly OTC markets, market makers play a central role by providing bid and ask prices for bonds to asset managers. Determining the optimal bid and ask quotes that a market maker should set for a given universe of bonds is a complex task. The existing models, mostly inspired by the Avellaneda-Stoikov model, describe the complex optimization problem faced by market makers: proposing bid and ask prices for making money out of the difference between them while mitigating the market risk associated with holding inventory. While most of the models only tackle one-asset market making, they can often be generalized to a multi-asset framework. However, the problem of solving the equations characterizing the optimal bid and ask quotes numerically is seldom tackled in the literature, especially in high dimension. In this paper, we propose a numerical method for approximating the optimal bid and ask quotes over a large universe of bonds in a model à la Avellaneda–Stoikov. As classical finite difference methods cannot be used in high dimension, we present a discrete-time method inspired by reinforcement learning techniques, namely, a model-based deep actor-critic algorithm. |
Databáze: | OpenAIRE |
Externí odkaz: |