Advancing Investment Frontiers: Industry-grade Deep Reinforcement Learning for Portfolio Optimization (2024)

Philip Ndikum
&Serge Ndikum
Disclaimer: This research paper, authored by Philip Ndikum & Serge Ndikum, is for informational and academic purposes only. The authors disclaim any representation or warranty for its accuracy or completeness. It is not intended as investment advice or an endorsem*nt of any specific investment or strategy. Copyright ©  2024, Philip Ndikum & Serge Ndikum. All rights reserved.

(February 2024)

Abstract

This research paper delves into the application of Deep Reinforcement Learning (DRL) in asset-class agnostic portfolio optimization, integrating industry-grade methodologies with quantitative finance. At the heart of this integration is our robust framework that not only merges advanced DRL algorithms with modern computational techniques but also emphasizes stringent statistical analysis, software engineering and regulatory compliance. To the best of our knowledge, this is the first study integrating financial Reinforcement Learning with sim-to-real methodologies from robotics and mathematical physics, thus enriching our frameworks and arguments with this unique perspective. Our research culminates with the introduction of AlphaOptimizerNet, a proprietary Reinforcement Learning agent (and corresponding library). Developed from a synthesis of state-of-the-art (SOTA) literature and our unique interdisciplinary methodology, AlphaOptimizerNet demonstrates encouraging risk-return optimization across various asset classes with realistic constraints. These preliminary results underscore the practical efficacy of our frameworks. As the finance sector increasingly gravitates towards advanced algorithmic solutions, our study bridges theoretical advancements with real-world applicability, offering a template for ensuring safety and robust standards in this technologically driven future.

Keywords Portfolio Optimization, Artificial Intelligence, Mathematical Finance, Deep Reinforcement Learning, Robotics, Sim-to-Real Transfer, Risk Management, Regulation, Software Engineering System Design.

1 Introduction and Broader Impact

The integration of Reinforcement Learning (RL) in asset-class agnostic portfolio optimization marks the beginning of a transformative era in quantitative finance, demanding interdisciplinary expertise and novel approaches. This study exemplifies this transformative shift, introducing an industry-grade RL framework integrated with state-of-the-art (SOTA) computational techniques, all tailored for real-world financial applications. To our knowledge, this is the first research paper to adapt simulation-to-real (sim-to-real) transfer perspectives from robotics and mathematical physics to finance. Our frameworks are designed to navigate high-stakes, non-stationary market dynamics within stringent regulatory frameworks. Arguably, this multidisciplinary approach enhances the practicality and robustness of financial models, paving the way for future research that rigorously accounts for historical risks [1, 2, 3, 4, 5, 6, 7, 8, 9, 10]. The domain of Deep Reinforcement Learning (DRL) has seen significant breakthroughs across various complex fields, underlining its potential to tackle previously insurmountable challenges. Among these, the success of DeepMind’s AlphaGo Zero in mastering the ancient game of Go stands as a testament to RL’s capability in navigating intricate strategic landscapes. Reinforcement Learning’s effectiveness in complex games, particularly in non-stationary environments, with partial information and multi-agent dynamics, closely mirrors the intricate challenges faced in financial markets. Recent advancements in RL, especially in diplomatic simulation games, underscore its potential in navigating stochastic decision-making processes under uncertainty and incomplete information. Such developments are emblematic of RL’s adaptability and skill, mirroring the complexities of financial portfolio management and signifying the technology’s applicability in this sophisticated domain [11, 12, 13, 14, 15, 16]. In robotics, leading companies like Boston Dynamics, DiDi, Tesla and others are leveraging Reinforcement Learning to develop advanced autonomous robots and vehicles. This highlights RL’s versatility and effectiveness in various high-stakes scenarios [17, 18, 19, 20]. In a parallel development, Large Language Models (LLMs) have undergone significant evolution through Reinforcement Learning with Human Feedback (RLHF). These models have not only achieved progress in natural language processing but also show promise in specialized domains like finance. Preliminary research suggests that financial LLMs may eventually outperform traditional human analysts in analyzing and interpreting complex financial data, indicating a potential shift in how financial information is processed and utilized [21, 22, 23, 24].

Our research endeavors to bridge the gap between theoretical concepts in academia and the dynamic, often unpredictable realities of real-world financial markets. Despite commendable and ambitious goals, many financial RL papers and open-source software artifacts, such as FinRL and its derivatives, often do not fully meet the stringent requirements of robust software engineering, regulatory compliance, and practical market constraints [25, 26, 27, 28]. We hope that this paper offers insights to enhance industry-grade applications of Reinforcement Learning in portfolio optimization and Financial RL more broadly. Our collaboration was born from our observation that many available open-source tools lack consistent requirements, robust documentation, and often exhibit coding errors stemming from inadequate finance domain knowledge. It is worth noting, that the arguments presented also align with industrial trends in the Artificial Intelligence (AI) industry, with many practitioners advocating for improved engineering and software design practices [29, 30, 31, 32, 33, 34]. In economics and AI research, replication challenges and incomplete findings, have been addressed in many excellent papers by industry practitioners but few have provided frameworks to concretely address these challenges in modern finance [35, 36, 37, 38, 39, 40, 41, 42, 43]. Given the high stakes of financial management, as evidenced by past financial crises and the fluctuating performance of investors, it is crucial that RL algorithms deployed in finance undergo rigorous stress-testing, statistical experiment design and conform to high engineering standards [44, 45, 46, 47, 48]. The proprietary library and RL agent, AlphaOptimizerNet, were meticulously crafted to ensure robustness in experiment design and performance analysis111It is important to note that many developments in the fields of finance and investment are proprietary and thus not publicly disclosed. As a result, this paper solely references publicly available academic literature and does not speculate nor comment on the advancements or strategies employed within private institutions. The development of our RL agent, AlphaOptimizerNet, and the accompanying proprietary library, is an example of such non-public, specialized work in the field.. Embracing an iterative and system-driven design philosophy, the development acknowledges the nuanced complexities inherent in contemporary AI system design. This research underscores the potential significance of sim-to-real considerations, a central theme throughout this paper. Such an approach may be crucial for organizations operating in an ecosystem increasingly shaped by advanced algorithms. Our objective is to argue for the adoption of elevated standards in both academic and industrial realms, contributing to the future landscape of financial management. The insights presented in this research might be particularly relevant to sovereign wealth funds, family offices, pension funds and public policy makers. In today’s financial environment, these institutions play a pivotal role as guardians of capital for the benefit of families, citizens, and nations. We aim to enhance individual well-being and positively influence the broader trajectory towards global prosperity.

2 Transcending Traditional Portfolio Optimization: A Reinforcement Learning Approach

2.1 Modern Portfolio Theory: Limitations and RL’s Role in Future Directions

Portfolio optimization is a fundamental component of quantitative investing, tasked with the strategic allocation of assets to maximize returns while minimizing risk. It encompasses a broad spectrum of asset classes and investment timelines, reflecting the multifaceted nature of financial markets. In our research, we consider portfolio optimization as a prime example to demonstrate the application of our Deep Reinforcement Learning (RL) methodologies. Building on the work of Ndikum (2020) [1], we highlight the critical role of financial, econometric and regulatory domain knowledge for the successful deployment of advanced AI algorithms in finance. As we delve into Deep RL techniques for portfolio optimization, we consciously choose to employ simplified mathematical formulations. This approach, aimed at making our work accessible to a diverse audience from various fields, reflects our commitment to balance technical detail with general comprehensibility. Through this strategy, we aim not only to engage a broader readership but also to foster meaningful discourse across different disciplines, maintaining the essence and integrity of our arguments throughout 222For a comprehensive textbook on Reinforcement Learning, we recommend Reinforcement Learning and Stochastic Optimization: A unified framework for sequential decisions by Professor Warren Powell, of Princeton University, 2022. [50].. Our objectives in this section are twofold: Firstly, we aim to elucidate the fundamental concept of diversification and the mechanics underlying Modern Portfolio Theory (MPT). Secondly, we delve into the inherent limitations of MPT, particularly in the context of today’s dynamic financial markets. A thorough grasp of these foundational concepts is indispensable, as it lays the groundwork for a deeper appreciation of the sim-to-real framework presented later in our paper. In the realm of portfolio management, diversification is not merely a strategy for risk mitigation; instead it fundamentally forms the philosophical underpinning of Modern Portfolio Theory (MPT). At the time of its inception, MPT represented a paradigm shift in financial portfolio management, systematically harnessing the principles of diversification to mathematically balance risk and return. This innovative approach revolutionized the way portfolios were constructed, moving beyond intuitive strategies to a quantifiable and objective models for optimization. Diversification, a central tenet of Modern Portfolio Theory (MPT), can be quantitatively expressed by considering the expected risk, represented as the variance of returns σp2superscriptsubscript𝜎𝑝2\sigma_{p}^{2}italic_σ start_POSTSUBSCRIPT italic_p end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT, in a diversified portfolio of n𝑛nitalic_n distinct assets. This approach to risk quantification considers the expected fluctuations in portfolio returns. We follow a mathematical formulation similar to Bodie et al. (2020) [51]:

σp2=i=1nj=1nwiwjCov(ri,rj).superscriptsubscript𝜎𝑝2superscriptsubscript𝑖1𝑛superscriptsubscript𝑗1𝑛subscript𝑤𝑖subscript𝑤𝑗Covsubscript𝑟𝑖subscript𝑟𝑗\sigma_{p}^{2}=\sum_{i=1}^{n}\sum_{j=1}^{n}w_{i}w_{j}\text{Cov}(r_{i},r_{j}).italic_σ start_POSTSUBSCRIPT italic_p end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT = ∑ start_POSTSUBSCRIPT italic_i = 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_n end_POSTSUPERSCRIPT ∑ start_POSTSUBSCRIPT italic_j = 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_n end_POSTSUPERSCRIPT italic_w start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT italic_w start_POSTSUBSCRIPT italic_j end_POSTSUBSCRIPT Cov ( italic_r start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT , italic_r start_POSTSUBSCRIPT italic_j end_POSTSUBSCRIPT ) .(1)

Equation 1 calculates the expected variance of portfolio returns, σp2superscriptsubscript𝜎𝑝2\sigma_{p}^{2}italic_σ start_POSTSUBSCRIPT italic_p end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT, illustrating how diversification impacts this risk333In this context, the notation σp2superscriptsubscript𝜎𝑝2\sigma_{p}^{2}italic_σ start_POSTSUBSCRIPT italic_p end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT represents the variance of the portfolio’s returns, and ρijsubscript𝜌𝑖𝑗\rho_{ij}italic_ρ start_POSTSUBSCRIPT italic_i italic_j end_POSTSUBSCRIPT (Greek Rho) refers to the correlation coefficient between the returns of assets i𝑖iitalic_i and j𝑗jitalic_j.. Here, wisubscript𝑤𝑖w_{i}italic_w start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT and wjsubscript𝑤𝑗w_{j}italic_w start_POSTSUBSCRIPT italic_j end_POSTSUBSCRIPT represent the portfolio weights allocated to assets i𝑖iitalic_i and j𝑗jitalic_j respectively, and Cov(ri,rj)Covsubscript𝑟𝑖subscript𝑟𝑗\text{Cov}(r_{i},r_{j})Cov ( italic_r start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT , italic_r start_POSTSUBSCRIPT italic_j end_POSTSUBSCRIPT ) is the covariance between their returns, measuring how their returns move in relation to each other. The essence of diversification in risk management lies in combining assets with varying degrees of correlation, thereby reducing the portfolio’s overall expected risk. The influence of this diversification becomes more significant as the number of assets increases and their correlations diversify. This dynamic is captured in the following equation:

limn,ρij0σp2=limn,ρij0i=1nj=1nwiwjCov(ri,rj)=C=infi𝒩σi2,\lim_{{n\to\infty,\atop\rho_{ij}\to 0}}\sigma_{p}^{2}=\lim_{{n\to\infty,\atop%\rho_{ij}\to 0}}\sum_{i=1}^{n}\sum_{j=1}^{n}w_{i}w_{j}\text{Cov}(r_{i},r_{j})=%C=\inf_{i\in\mathcal{N}}\sigma_{i}^{2},roman_lim start_POSTSUBSCRIPT FRACOP start_ARG italic_n → ∞ , end_ARG start_ARG italic_ρ start_POSTSUBSCRIPT italic_i italic_j end_POSTSUBSCRIPT → 0 end_ARG end_POSTSUBSCRIPT italic_σ start_POSTSUBSCRIPT italic_p end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT = roman_lim start_POSTSUBSCRIPT FRACOP start_ARG italic_n → ∞ , end_ARG start_ARG italic_ρ start_POSTSUBSCRIPT italic_i italic_j end_POSTSUBSCRIPT → 0 end_ARG end_POSTSUBSCRIPT ∑ start_POSTSUBSCRIPT italic_i = 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_n end_POSTSUPERSCRIPT ∑ start_POSTSUBSCRIPT italic_j = 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_n end_POSTSUPERSCRIPT italic_w start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT italic_w start_POSTSUBSCRIPT italic_j end_POSTSUBSCRIPT Cov ( italic_r start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT , italic_r start_POSTSUBSCRIPT italic_j end_POSTSUBSCRIPT ) = italic_C = roman_inf start_POSTSUBSCRIPT italic_i ∈ caligraphic_N end_POSTSUBSCRIPT italic_σ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT ,(2)

Equation 2 captures the theoretical impact of diversification on the expected risk of a portfolio, measured using the portfolio variance σp2superscriptsubscript𝜎𝑝2\sigma_{p}^{2}italic_σ start_POSTSUBSCRIPT italic_p end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT. As the number of assets n𝑛nitalic_n in the portfolio grows indefinitely and their pairwise correlations ρijsubscript𝜌𝑖𝑗\rho_{ij}italic_ρ start_POSTSUBSCRIPT italic_i italic_j end_POSTSUBSCRIPT converge to zero, the portfolio’s expected risk trends towards a theoretical minimum limit, denoted as C𝐶Citalic_C. This limit, C𝐶Citalic_C, is specifically defined as the infimum (infinfimum\infroman_inf) of the variances (σi2superscriptsubscript𝜎𝑖2\sigma_{i}^{2}italic_σ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT) of the individual assets within the portfolio set 𝒩𝒩\mathcal{N}caligraphic_N. In essence, C𝐶Citalic_C represents the lowest level of risk (variance) that is theoretically achievable under the ideal scenario of infinite diversification with completely uncorrelated assets. This concept is central to portfolio theory, emphasizing the advantage of having a diverse mix of uncorrelated or negatively correlated assets in a portfolio. Such diversification is a key strategy in risk management, aiming to reduce the impact of market volatility and improve the overall stability of an investment portfolio. In practical terms, it means spreading investments across a variety of asset classes to buffer against unexpected market movements, thereby enhancing the resilience of one’s financial portfolio. The use of infimum notation in this equation highlights the inherently positive nature of this theoretical minimum risk. It underscores that even with ideal diversification, portfolio risk doesn’t reduce to zero but converges towards the lowest variance found among the individual assets.

This concept of minimizing expected risk through diversification leads us to the core principles of Modern Portfolio Theory (MPT), as formulated by Harry Markowitz. In recognition of their pivotal contribution to the field, Markowitz, along with Merton Miller and William Sharpe, were collectively awarded the Nobel Prize in Economics in 1990 [52, 53, 54, 55]. MPT pragmatically applies the principle of diversification in the context of portfolio optimization. MPT algorithms are often categorized under Quadratic Programming (QP) or Mixed Integer Programming (MIP) problems within the fields of numerical analysis and applied mathematics. Definition 3 provides a mathematical formulation of a single-period MPT, drawing upon the formulations by Chang et al. (2002) and Cesarone et al. (2013) [56, 57, 58, 59, 60]. In this definition, n𝑛nitalic_n represents the total number of financial assets, wisubscript𝑤𝑖w_{i}italic_w start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT represents the portfolio allocation to asset i𝑖iitalic_i in the range [0,1]01[0,1][ 0 , 1 ]. The expected return and covariance for each asset are denoted as μisubscript𝜇𝑖\mu_{i}italic_μ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT and σijsubscript𝜎𝑖𝑗\sigma_{ij}italic_σ start_POSTSUBSCRIPT italic_i italic_j end_POSTSUBSCRIPT, respectively. In simple terms, we aim to optimize portfolio weights to minimize risk while targeting a specific expected return 𝔼[R]=R𝔼𝑅𝑅\operatorname{\mathbb{E}}[R]=Rblackboard_E [ italic_R ] = italic_R. This unconstrained formulation can be solved efficiently with modern computational tools, allowing for the calculation of optimal portfolio weights, designated as ϕ(𝔼[R])italic-ϕ𝔼𝑅\phi(\operatorname{\mathbb{E}}[R])italic_ϕ ( blackboard_E [ italic_R ] ). The resultant efficient frontier identifies portfolios that optimize expected return for various levels of risk within the interval [Rmin,Rmax]subscript𝑅minsubscript𝑅max[R_{\text{min}},R_{\text{max}}][ italic_R start_POSTSUBSCRIPT min end_POSTSUBSCRIPT , italic_R start_POSTSUBSCRIPT max end_POSTSUBSCRIPT ]. In real-world investing scenarios, this deceptively simple formulation of Modern Portfolio Theory (MPT) is significantly complicated by the introduction of various real-world constraints.

Definition 2.1 (Markowitz Single Period Unconstrained Portfolio Optimization[56, 57]).
min\displaystyle\min\quadroman_mini=1nj=1nσijwiwjsuperscriptsubscript𝑖1𝑛superscriptsubscript𝑗1𝑛subscript𝜎𝑖𝑗subscript𝑤𝑖subscript𝑤𝑗\displaystyle\sum_{i=1}^{n}\sum_{j=1}^{n}\sigma_{ij}w_{i}w_{j}∑ start_POSTSUBSCRIPT italic_i = 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_n end_POSTSUPERSCRIPT ∑ start_POSTSUBSCRIPT italic_j = 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_n end_POSTSUPERSCRIPT italic_σ start_POSTSUBSCRIPT italic_i italic_j end_POSTSUBSCRIPT italic_w start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT italic_w start_POSTSUBSCRIPT italic_j end_POSTSUBSCRIPT(3)
s.t.i=1nwiμi=𝔼[R]superscriptsubscript𝑖1𝑛subscript𝑤𝑖subscript𝜇𝑖𝔼𝑅\displaystyle\sum_{i=1}^{n}w_{i}\mu_{i}=\operatorname{\mathbb{E}}[R]∑ start_POSTSUBSCRIPT italic_i = 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_n end_POSTSUPERSCRIPT italic_w start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT italic_μ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT = blackboard_E [ italic_R ]
i=1nwi=1superscriptsubscript𝑖1𝑛subscript𝑤𝑖1\displaystyle\sum_{i=1}^{n}w_{i}=1∑ start_POSTSUBSCRIPT italic_i = 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_n end_POSTSUPERSCRIPT italic_w start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT = 1
wi[0,1],fori{1,,n}.formulae-sequencesubscript𝑤𝑖01for𝑖1𝑛\displaystyle w_{i}\in[0,1],\text{ for }i\in\{1,\dots,n\}.italic_w start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ∈ [ 0 , 1 ] , for italic_i ∈ { 1 , … , italic_n } .

The Financial literature published in the last few decades has thoroughly explored and addressed these constraints, adding complexity to MPT models. The intricate and high-stakes nature of finance necessitates a deep dive into this rich history. While contemporary and future technological advancements in Reinforcement Learning, present innovative possibilities, their effective application demands a thorough grounding in the enduring realities of financial markets and regulatory environments. Neglecting the depth of historical financial knowledge can oversimplify intricate market dynamics, potentially steering strategies towards catastrophic outcomes for financial institutions and investors in both the private and public sectors. Below, we enumerate some constraints and place them in context with real-world problems:

  1. 1.

    Cardinality Constraint:This constraint limits the number of assets in a portfolio to a maximum of n𝑛nitalic_n:

    |supp(w)|0n,wheresupp(w):={i:wi>0}.formulae-sequencesubscriptsupp𝑤0𝑛assignwheresupp𝑤conditional-set𝑖subscript𝑤𝑖0|\mathrm{supp}(w)|_{0}\leq n,\quad\text{where }\mathrm{supp}(w):=\{i:w_{i}>0\}.| roman_supp ( italic_w ) | start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT ≤ italic_n , where roman_supp ( italic_w ) := { italic_i : italic_w start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT > 0 } .(4)

    Here, supp(w)supp𝑤\mathrm{supp}(w)roman_supp ( italic_w ) denotes the set of indices i𝑖iitalic_i for which the investment in asset i𝑖iitalic_i (wisubscript𝑤𝑖w_{i}italic_w start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT) is greater than zero, implying active inclusion of the asset in the portfolio. The expression |supp(w)|0subscriptsupp𝑤0|\mathrm{supp}(w)|_{0}| roman_supp ( italic_w ) | start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT counts the total number of distinct assets with non-zero investments in the portfolio. The constraint |supp(w)|0nsubscriptsupp𝑤0𝑛|\mathrm{supp}(w)|_{0}\leq n| roman_supp ( italic_w ) | start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT ≤ italic_n ensures that this number does not exceed the predetermined maximum n𝑛nitalic_n, thereby limiting the total number of different assets included in the portfolio. In asset management, this constraint is relevant for ensuring portfolio transparency and manageability.

  2. 2.

    Volume Constraint:The volume constraint restricts the allocation to each asset in the portfolio within a specified range, denoted by

    iwiuifor each asseti{1,,n},subscript𝑖subscript𝑤𝑖subscript𝑢𝑖for each asset𝑖1𝑛\ell_{i}\leq w_{i}\leq u_{i}\;\text{for each asset }i\in\{1,\dots,n\},roman_ℓ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ≤ italic_w start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ≤ italic_u start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT for each asset italic_i ∈ { 1 , … , italic_n } ,(5)

    where isubscript𝑖\ell_{i}roman_ℓ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT and uisubscript𝑢𝑖u_{i}italic_u start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT define the lower and upper bounds, respectively, for the investment in asset i𝑖iitalic_i. This constraint is critical in preventing over-concentration in a single asset, thereby reducing the risk of significant loss from adverse movements in that asset’s value. In high frequency trading (HFT), the volume constraint indirectly influences turnover and liquidity, both of which are critical factors affecting the overall effectiveness of trading strategies.

  3. 3.

    Regulatory Constraints and Transaction Costs:Legal and regulatory constraints may dictate investment limitations in certain sectors or geographies. For example, contemporary investment strategies may be influenced by client or government mandates that encourage investments in certain sectors, such as renewable energy, reflecting broader objectives that may not be captured mathematically by classical models. These external directives can significantly shape portfolio composition. Beyond regulatory considerations, transaction costs, including taxes and brokerage fees, play a crucial role. The incorporation of such constraints can dramatically alter the performance of back-tested simulations in both Modern Portfolio Theory and newer methodologies such as Reinforcement Learning.

As we can observe, the adoption of Modern Portfolio Theory (MPT) in dynamic financial scenarios necessitates confronting a myriad of complexities. Additionally, we note that these classical solutions operate under static market assumptions and are often geared towards single-period frameworks. This inherent limitation impedes their utility in environments characterized by asymmetric risk profiles444Post-Modern Portfolio Theory (PMPT) seeks to mitigate challenges associated with asymmetric risk. Nevertheless, it predominantly utilizes conventional optimization methods, which are subject to analogous computational challenges [61, 62]., human irrationality, and complex game-theoretic dynamics. It is within this context that the concept of multi-period optimization solutions have gained recent prominence:

Definition 2.2 (Multi-period Optimization).

Consider a universe of n𝑛nitalic_n assets over a time horizon of K𝐾Kitalic_K periods. We provide a similar formulation to Lezmi et al. (2022) [63] and transition to a probabilistic framework to model our multi-period optimization as follows:

W*=argmaxWΩ𝔼[U(W)|t]superscript𝑊subscriptargmax𝑊Ω𝔼delimited-[]conditional𝑈𝑊subscript𝑡W^{*}=\operatorname*{arg\,max}_{W\in\Omega}\mathbb{E}\left[U(W)\,|\,\mathcal{F%}_{t}\right]italic_W start_POSTSUPERSCRIPT * end_POSTSUPERSCRIPT = start_OPERATOR roman_arg roman_max end_OPERATOR start_POSTSUBSCRIPT italic_W ∈ roman_Ω end_POSTSUBSCRIPT blackboard_E [ italic_U ( italic_W ) | caligraphic_F start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ](6)

where

W=(w1,t+1w1,t+2w1,t+Kw2,t+1w2,t+2w2,t+Kwn,t+1wn,t+2wn,t+K),𝑊matrixsubscript𝑤1𝑡1subscript𝑤1𝑡2subscript𝑤1𝑡𝐾subscript𝑤2𝑡1subscript𝑤2𝑡2subscript𝑤2𝑡𝐾subscript𝑤𝑛𝑡1subscript𝑤𝑛𝑡2subscript𝑤𝑛𝑡𝐾W=\begin{pmatrix}w_{1,t+1}&w_{1,t+2}&\cdots&w_{1,t+K}\\w_{2,t+1}&w_{2,t+2}&\cdots&w_{2,t+K}\\\vdots&\vdots&\ddots&\vdots\\w_{n,t+1}&w_{n,t+2}&\cdots&w_{n,t+K}\end{pmatrix},italic_W = ( start_ARG start_ROW start_CELL italic_w start_POSTSUBSCRIPT 1 , italic_t + 1 end_POSTSUBSCRIPT end_CELL start_CELL italic_w start_POSTSUBSCRIPT 1 , italic_t + 2 end_POSTSUBSCRIPT end_CELL start_CELL ⋯ end_CELL start_CELL italic_w start_POSTSUBSCRIPT 1 , italic_t + italic_K end_POSTSUBSCRIPT end_CELL end_ROW start_ROW start_CELL italic_w start_POSTSUBSCRIPT 2 , italic_t + 1 end_POSTSUBSCRIPT end_CELL start_CELL italic_w start_POSTSUBSCRIPT 2 , italic_t + 2 end_POSTSUBSCRIPT end_CELL start_CELL ⋯ end_CELL start_CELL italic_w start_POSTSUBSCRIPT 2 , italic_t + italic_K end_POSTSUBSCRIPT end_CELL end_ROW start_ROW start_CELL ⋮ end_CELL start_CELL ⋮ end_CELL start_CELL ⋱ end_CELL start_CELL ⋮ end_CELL end_ROW start_ROW start_CELL italic_w start_POSTSUBSCRIPT italic_n , italic_t + 1 end_POSTSUBSCRIPT end_CELL start_CELL italic_w start_POSTSUBSCRIPT italic_n , italic_t + 2 end_POSTSUBSCRIPT end_CELL start_CELL ⋯ end_CELL start_CELL italic_w start_POSTSUBSCRIPT italic_n , italic_t + italic_K end_POSTSUBSCRIPT end_CELL end_ROW end_ARG ) ,

denotes the matrix of asset allocations across K𝐾Kitalic_K future time periods, with each wi,t+jsubscript𝑤𝑖𝑡𝑗w_{i,t+j}italic_w start_POSTSUBSCRIPT italic_i , italic_t + italic_j end_POSTSUBSCRIPT being the column vector of portfolio weights allocated to asset i𝑖iitalic_i at time period t+j𝑡𝑗t+jitalic_t + italic_j. Here, W*superscript𝑊W^{*}italic_W start_POSTSUPERSCRIPT * end_POSTSUPERSCRIPT represents the optimal series of allocations. The function U(W)𝑈𝑊U(W)italic_U ( italic_W ) represents the inter-temporal utility function, expressing the investor’s preference over varied portfolio allocations. The term tsubscript𝑡\mathcal{F}_{t}caligraphic_F start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT is the filtration, representing the cumulative information available up to time t𝑡titalic_t, including past market data and events. The optimization is subject to the condition WΩ𝑊normal-ΩW\in\Omegaitalic_W ∈ roman_Ω, where Ωnormal-Ω\Omegaroman_Ω denotes the sample space of all feasible sequences of portfolio allocations, subject to a set of linear and non-linear constraints. This framework allows for the ranking of allocations based on desirability, with superior utility values indicating more favored allocations, factoring in considerations such as returns, risks, and other investor inclinations.

The multi-period optimization shown in Definition 2.2 aims to forecast and optimize portfolio weights across several forward-looking timeframes. Despite its theoretical robustness, real-world application of multi-period optimization presents substantial challenges adequately addressed and summarized by Kolm et al. (2014) in their analysis of portfolio optimization’s evolution and challenges:

“In practice, multi-period models are seldom used. There are several practical reasons for that. First, it is often very difficult to accurately estimate return and risk for multiple periods, let alone for a single period. Second, multi-period models are in general computationally intensive, especially if the universe of assets considered is large. Third, the most common existing multi-period models do not handle real-world constraints …For these reasons, practitioners typically use single-period models to rebalance the portfolio from one period to another” - 60 Years of portfolio optimization: Practical challenges and current trends, Kolm et al. (2014) [64].

This complexity is compounded in practical numerical solutions to MPT, such as Mean-Variance Optimization (MVO). MVO applies the principles of MPT in a quantitative framework to identify the optimal asset allocation for a given risk level. A key requirement in MVO is for the covariance matrix, which represents the relationships between the returns of different assets, to be positive semi-definite. This ensures that the calculated variance of the portfolio, a key risk measure in MPT, is always non-negative. Additionally, the inherent variability or volatility in financial markets adds another layer of complexity. The assumption of stable volatility, a cornerstone of MVO, often does not hold true in real-world scenarios where market conditions can rapidly change. In contrast, Reinforcement Learning (RL) offers a promising alternative. RL thrives in environments characterized by uncertainty and the need for sequential decision-making, aligning well with the dynamic nature of financial markets. Unlike MVO, RL does not require the stringent assumptions about volatility and matrix properties. RL’s ability to handle complex, dynamic environments, coupled with its flexibility in integrating real-world constraints, makes it a formidable tool in addressing the challenges highlighted by Kolm et al. [65, 66, 67, 68]. State-of-the-art RL algorithms offer the potential to navigate the evolving landscape of risks and large datasets, holding the promise of superior and autonomous risk-adjusted returns. This shift in focus towards RL and similar advanced algorithms, however, must not eclipse the importance of grounding these technologies in the contemporary and historical realities of financial markets and regulatory environments [69, 70]. Given the stakes involved in managing client capital, it seems imperative for the financial industry, similar to other high-risk sectors where RL is currently being used in industry (e.g., robotics), to establish rigorous standards ensuring the responsible deployment of these algorithms. In this paper, we advocate for a paradigm where the effectiveness of techniques such as RL is gauged not merely by their theoretical prowess, but by their alignment with the dynamic - and highly regulated demands of contemporary international finance which we will explore in the subsequent sections.

2.2 Reinforcement Learning in Portfolio Optimization: Bridging Theory and Practice

The transition to Reinforcement Learning (RL) in financial portfolio management marks a pivotal shift from conventional strategies to advanced, algorithm-driven approaches. Deep Reinforcement Learning (DRL), an extension of RL, effectively employs deep neural networks to navigate complex, high-dimensional environments. This subject will be further explored in upcoming sections. RL, set apart from the single-period, deterministic models of Modern Portfolio Theory (MPT), adopts a dynamic, data-driven approach. It is adept at managing multiple, even infinite time horizons signaling a move towards creating computational equivalents of human traders and investment managers. Reinforcement Learning systems operate as an autonomous agents (advanced algorithm or models), learning and adapting continuously via interactions with a simulated environment that mirrors real-world market complexities. This learning process, entirely driven by data, enables the RL agent to identify patterns and make informed decisions based on extensive market data. This continual, dynamic learning process is essential for developing investment strategies that are responsive to the rapidly changing and unpredictable financial markets. This evolution underscores a departure from the traditional reliance on human emotional resilience in financial decision-making. With its capacity to process large data volumes, RL holds the potential to match or even exceed human capabilities in identifying and capitalizing on market inefficiencies. Free from human psychological biases and fatigue, RL systems introduce a major transformation in the financial sector. They combine human expertise with the precision of computational systems. In this new era, trust in financial decision-making extends beyond human judgment to the dependability and accuracy of these meticulously engineered computational systems. This transition represents a major shift in the landscape of financial portfolio management. To facilitate a deeper understanding and bridge the gap towards our novel sim-to-real frameworks, we employ analogies from robotics throughout our discussion. These analogies not only provide a familiar reference point but also align closely with the principles and challenges inherent in both fields:

  • State (stsubscript𝑠𝑡s_{t}italic_s start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT): In RL, the state at time t𝑡titalic_t, stsubscript𝑠𝑡s_{t}italic_s start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT, represents a snapshot of the current situation, analogous to the perception of an environment by an RL agent in robotics. In portfolio optimization, stsubscript𝑠𝑡s_{t}italic_s start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT encapsulates market conditions, including key financial indicators such as asset prices and trading volumes. Understanding the state’s formulation requires a nuanced grasp of the asset class and regulatory context, comparable to the detailed environmental awareness necessary for an RL agent in robotics. The strategic design of the state space, similar to sensor range and resolution considerations in robotics, is crucial for capturing detailed market representation while maintaining computational manageability. This process is central to optimizing the RL model’s efficiency in interpreting and reacting to market dynamics, paralleling the way an RL agent in robotics processes sensory data to understand its environment.

  • Action (atsubscript𝑎𝑡a_{t}italic_a start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT): The action atsubscript𝑎𝑡a_{t}italic_a start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT at time t𝑡titalic_t signifies the decision made by the RL agent, comparable to decision-making processes of RL agents in robotics. In portfolio optimization, this action primarily involves determining the asset allocations, represented as a vector. The constraint i=1nai=1superscriptsubscript𝑖1𝑛subscript𝑎𝑖1\sum_{i=1}^{n}a_{i}=1∑ start_POSTSUBSCRIPT italic_i = 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_n end_POSTSUPERSCRIPT italic_a start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT = 1 ensures that the total allocation across all assets equals one, reflecting a similar allocation of resources or efforts by RL agents in robotics. This constraint is critical in ensuring that the portfolio weights are proportionally distributed. Actions may also incorporate factors such as maintaining cash reserves or opting for periods of inactivity, dependent on the strategy. The action space design must align with specific investment strategies and comply with regulation and investment mandates, including decisions regarding the use of leverage.

  • Reward (rtsubscript𝑟𝑡r_{t}italic_r start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT): The reward rtsubscript𝑟𝑡r_{t}italic_r start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT at time t𝑡titalic_t functions as the guiding incentive for the RL agent, comparable to the objectives pursued in robotics where agents learn to navigate and interact with dynamic and complex physical environments. In portfolio optimization, the agent’s reward is typically aligned with the maximization of risk-adjusted returns, tailored to meet investor-specific goals and limitations. This mechanism is crucial in guiding the agent’s learning path, enabling it to develop and refine strategies that effectively balance return and risk.

  • Environment: In financial RL, the environment represents the complexities of market dynamics, regulatory frameworks, and economic indicators, similar to the intricate and evolving environments encountered in robotics. The financial environment is inherently non-stationary and only partially observable, posing significant challenges comparable to those faced in robotic navigation and interaction. Designing an RL model for finance requires a deep understanding of market intricacies and nuances. The environment should mirror specific investment timeframes and strategies, with considerations varying greatly between applications such as high-frequency trading and long-term asset management. It’s imperative that the environment is realistically modeled to encapsulate market behaviors and constraints, ensuring the RL model’s relevance and effectiveness in actual financial scenarios.

As we explore Reinforcement Learning (RL) for portfolio optimization, understanding its mathematical foundations is crucial. RL algorithms, though complex, rely on well-established mathematical principles and computational procedures. Creating these systems necessitates expertise in AI, finance, and a deep understanding of regulatory and risk considerations. This skill set is essential for successful deployment of robust Reinforcement Learning systems in finance. Our aim is to present these mathematical concepts clearly and comprehensively, particularly for readers with a finance background. While equations are our focus, the accompanying explanations provide an insight into their underlying principles. Designing RL systems for portfolio optimization requires more than technical proficiency; it entails integrating financial knowledge, AI capabilities, and an understanding of real-world regulations and risks. As we delve into Markov Decision Processes (MDPs), Partially Observable Markov Decision Processes (POMDPs), and key aspects of Deep Learning, our goal is to maintain a balance between technical rigor and accessibility. Subsequent sections will cover mathematical equations while contextualizing them within financial portfolio management.

2.3 Mathematical Formalism: MDPs, POMDPs, Deep Learning

Advancing Investment Frontiers: Industry-grade Deep Reinforcement Learning for Portfolio Optimization (1)

We now delve into the mathematical formalism of Deep Reinforcement Learning (RL) shown in Fig. 2. This section is structured to first introduce intuitive explanations of Markov Decision Processes (MDPs) and Partially Observable Markov Decision Processes (POMDPs), followed by their formal definitions, and then a discussion of Deep Learning and neural networks. A key distinction between broader Reinforcement Learning (RL) approaches and Deep Reinforcement Learning (DRL) is the use of deep neural networks in the latter, enabling them to handle more complex, high-dimensional and noisy financial market data. Although an in-depth technical exploration exceeds the scope of this paper, we introduce neural networks as the key driving force behind our RL algorithms. Analogous to computational brains555It is essential to understand that neural networks, while biologically inspired, do not replicate the full complexity of biological brains. These statistical learning algorithms are abstractions, simplifying key principles from our evolving understanding of neuroscience. Interestingly, there appears to be a correlation between advancements in computational neuroscience and the development of increasingly performant neural network architectures. However, this should not be mistaken for a comprehensive understanding of the brain’s intricacies or an implication of neural networks’ complete mimicry of biological processes [71, 72]., these biologically inspired algorithms are indispensable due to their ability as universal function approximators, essential for effective planning, prediction, and decision-making within MDP and POMDP frameworks. In our examination of Markov Decision Processes (MDPs) and Partially Observable Markov Decision Processes (POMDPs), let’s offer simplified interpretations before presenting formal definitions. MDPs represent decision-making frameworks with complete market information. Imagine making investment decisions where you have full visibility into asset prices, economic indicators, and market trends. This scenario aligns with an MDP framework. Conversely, POMDPs apply when some market information is hidden or uncertain. Picture a situation where certain variables, such as insider information or future market movements, are unknown. Despite this uncertainty, decisions must be made based on available information. POMDPs provide a framework for decision-making under such conditions, allowing adaptation to market dynamics even with incomplete visibility:

Definition 2.3 (Markov Decision Processes in Portfolio Optimization).

In portfolio optimization, an MDP is characterized by the tuple (S,A,T,T0,R,γ)𝑆𝐴𝑇subscript𝑇0𝑅𝛾(S,A,T,T_{0},R,\gamma)( italic_S , italic_A , italic_T , italic_T start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT , italic_R , italic_γ ). Here, S𝑆Sitalic_S represents the array of possible market states, encompassing variables like asset prices and economic indicators. A𝐴Aitalic_A denotes the range of possible investment actions, reflecting decisions across a portfolio of n𝑛nitalic_n assets, thus A[0,1]n𝐴superscript01𝑛A\subseteq[0,1]^{n}italic_A ⊆ [ 0 , 1 ] start_POSTSUPERSCRIPT italic_n end_POSTSUPERSCRIPT. The transition function T:S×AΔ(S)normal-:𝑇normal-→𝑆𝐴normal-Δ𝑆T:S\times A\rightarrow\Delta(S)italic_T : italic_S × italic_A → roman_Δ ( italic_S ) models the probability of moving from one market state to another, given a specific investment action. T0:S[0,1]normal-:subscript𝑇0normal-→𝑆01T_{0}:S\rightarrow[0,1]italic_T start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT : italic_S → [ 0 , 1 ] defines the initial distribution of states in the market. The reward function R:S×Anormal-:𝑅normal-→𝑆𝐴R:S\times A\rightarrow\mathbb{R}italic_R : italic_S × italic_A → blackboard_R quantifies the financial impact of each action, considering both returns and associated risks. Unlike finite-horizon models, the time horizon in this setting is considered infinite, and the discount factor γ[0,1)𝛾01\gamma\in[0,1)italic_γ ∈ [ 0 , 1 ) reflects the long-term strategy of the investment approach. The primary objective is to determine an optimal policy π*:SAnormal-:superscript𝜋normal-→𝑆𝐴\pi^{*}:S\rightarrow Aitalic_π start_POSTSUPERSCRIPT * end_POSTSUPERSCRIPT : italic_S → italic_A that maximizes expected returns over this infinite horizon, as formalized in the following equation:

π*=argmaxπ𝔼st,at,rtT,π[t=0γtrt+1|s0].superscript𝜋subscript𝜋subscript𝔼formulae-sequencesimilar-tosubscript𝑠𝑡subscript𝑎𝑡subscript𝑟𝑡𝑇𝜋delimited-[]conditionalsuperscriptsubscript𝑡0superscript𝛾𝑡subscript𝑟𝑡1subscript𝑠0\pi^{*}=\arg\max_{\pi}\mathbb{E}_{s_{t},a_{t},r_{t}\sim T,\pi}\left[\sum_{t=0}%^{\infty}\gamma^{t}r_{t+1}\;|\;s_{0}\right].italic_π start_POSTSUPERSCRIPT * end_POSTSUPERSCRIPT = roman_arg roman_max start_POSTSUBSCRIPT italic_π end_POSTSUBSCRIPT blackboard_E start_POSTSUBSCRIPT italic_s start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT , italic_a start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT , italic_r start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ∼ italic_T , italic_π end_POSTSUBSCRIPT [ ∑ start_POSTSUBSCRIPT italic_t = 0 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT ∞ end_POSTSUPERSCRIPT italic_γ start_POSTSUPERSCRIPT italic_t end_POSTSUPERSCRIPT italic_r start_POSTSUBSCRIPT italic_t + 1 end_POSTSUBSCRIPT | italic_s start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT ] .(7)

The objective of the RL agent in this framework is to learn the optimal mapping of states to actions (portfolio allocations), thereby maximizing long-term investment returns while adapting to the dynamic nature of financial markets.

Definition 2.4 (Partially Observable Markov Decision Processes in Financial Markets).

POMDPs extend the MDP framework to scenarios where full knowledge of market states is not directly observable by investors. A POMDP in finance is characterized by the tuple (S,A,O,T,T0,R,γ)𝑆𝐴𝑂𝑇subscript𝑇0𝑅𝛾(S,A,O,T,T_{0},R,\gamma)( italic_S , italic_A , italic_O , italic_T , italic_T start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT , italic_R , italic_γ ). Here, S𝑆Sitalic_S denotes the complete range of market states, A𝐴Aitalic_A represents the range of investment actions, and O𝑂Oitalic_O encompasses observable market factors. The transition function T:S×AΔ(S)normal-:𝑇normal-→𝑆𝐴normal-Δ𝑆T:S\times A\rightarrow\Delta(S)italic_T : italic_S × italic_A → roman_Δ ( italic_S ) models the dynamics of market state changes following investment actions, and T0:S[0,1]normal-:subscript𝑇0normal-→𝑆01T_{0}:S\rightarrow[0,1]italic_T start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT : italic_S → [ 0 , 1 ] defines the initial state distribution. The observation function O:S×AΔ(O)normal-:𝑂normal-→𝑆𝐴normal-Δ𝑂O:S\times A\rightarrow\Delta(O)italic_O : italic_S × italic_A → roman_Δ ( italic_O ) determines the likelihood of observing specific market indicators given the state and action. The reward function R:S×Anormal-:𝑅normal-→𝑆𝐴R:S\times A\rightarrow\mathbb{R}italic_R : italic_S × italic_A → blackboard_R quantifies the financial impact of actions. The decision-making policy in a POMDP, π:Anormal-:𝜋normal-→𝐴\pi:\mathcal{H}\rightarrow Aitalic_π : caligraphic_H → italic_A, where \mathcal{H}caligraphic_H is the history of observations, actions, and rewards, depends on both the current and past observations. This history up to time t𝑡titalic_t is represented as τ0:t=(o0,a0,o1,r1,,at1,ot,rt)subscript𝜏normal-:0𝑡subscript𝑜0subscript𝑎0subscript𝑜1subscript𝑟1normal-…subscript𝑎𝑡1subscript𝑜𝑡subscript𝑟𝑡\tau_{0:t}=(o_{0},a_{0},o_{1},r_{1},\dots,a_{t-1},o_{t},r_{t})italic_τ start_POSTSUBSCRIPT 0 : italic_t end_POSTSUBSCRIPT = ( italic_o start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT , italic_a start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT , italic_o start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT , italic_r start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT , … , italic_a start_POSTSUBSCRIPT italic_t - 1 end_POSTSUBSCRIPT , italic_o start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT , italic_r start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ). The objective in a POMDP setting is to identify an optimal policy π*superscript𝜋\pi^{*}italic_π start_POSTSUPERSCRIPT * end_POSTSUPERSCRIPT that maximizes expected returns over an infinite horizon, considering both observable and unobservable market factors:

π*=argmaxπ𝔼st,at,ot,rtT,O,π[t=0γtrt+1|o0].superscript𝜋subscript𝜋subscript𝔼formulae-sequencesimilar-tosubscript𝑠𝑡subscript𝑎𝑡subscript𝑜𝑡subscript𝑟𝑡𝑇𝑂𝜋delimited-[]conditionalsuperscriptsubscript𝑡0superscript𝛾𝑡subscript𝑟𝑡1subscript𝑜0\pi^{*}=\arg\max_{\pi}\mathbb{E}_{s_{t},a_{t},o_{t},r_{t}\sim T,O,\pi}\left[%\sum_{t=0}^{\infty}\gamma^{t}r_{t+1}\;|\;o_{0}\right].italic_π start_POSTSUPERSCRIPT * end_POSTSUPERSCRIPT = roman_arg roman_max start_POSTSUBSCRIPT italic_π end_POSTSUBSCRIPT blackboard_E start_POSTSUBSCRIPT italic_s start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT , italic_a start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT , italic_o start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT , italic_r start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ∼ italic_T , italic_O , italic_π end_POSTSUBSCRIPT [ ∑ start_POSTSUBSCRIPT italic_t = 0 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT ∞ end_POSTSUPERSCRIPT italic_γ start_POSTSUPERSCRIPT italic_t end_POSTSUPERSCRIPT italic_r start_POSTSUBSCRIPT italic_t + 1 end_POSTSUBSCRIPT | italic_o start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT ] .(8)

In financial markets, where complete state information is often unavailable, this model becomes crucial, necessitating decision-making based on partial observations [73, 74].

In our exploration of decision-making frameworks, such as MDPs and POMDPs within dynamic financial environments, we have encountered foundational concepts that underpin adaptive learning. Now, we pivot our focus to a fundamental concept in Reinforcement Learning (RL) that serves as a linchpin for adaptive decision-making: the Bellman equation. Developed by Richard Bellman in 1957 [75], the Bellman equation plays a pivotal role in dissecting decision-making processes into interconnected steps, enabling the evaluation of both immediate rewards and the anticipated utility of future actions. At the heart of RL methodologies lies the concept of the value function, which encapsulates the expected cumulative future rewards an agent can attain from a given state. In the context of portfolio optimization, this value function serves as the cornerstone of the RL agent’s learning process, guiding it towards determining the optimal portfolio allocation strategy. By iteratively updating the value function based on observed rewards and transitions between states, the RL agent gradually hones its understanding of the market dynamics and learns to make informed decisions that maximize long-term returns while balancing risk. With this understanding of the value function’s significance in RL, let us delve into the formulation of the Bellman equation and its implications for portfolio optimization:

V(st)=maxat[R(st,at)+γt=t+1P(st|st,at)V(st)]𝑉subscript𝑠𝑡subscriptsubscript𝑎𝑡𝑅subscript𝑠𝑡subscript𝑎𝑡𝛾superscriptsubscriptsuperscript𝑡𝑡1𝑃conditionalsubscript𝑠superscript𝑡subscript𝑠𝑡subscript𝑎𝑡𝑉subscript𝑠superscript𝑡V(s_{t})=\max_{a_{t}}\left[R(s_{t},a_{t})+\gamma\sum_{t^{\prime}=t+1}^{\infty}%P(s_{t^{\prime}}|s_{t},a_{t})V(s_{t^{\prime}})\right]italic_V ( italic_s start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ) = roman_max start_POSTSUBSCRIPT italic_a start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT end_POSTSUBSCRIPT [ italic_R ( italic_s start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT , italic_a start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ) + italic_γ ∑ start_POSTSUBSCRIPT italic_t start_POSTSUPERSCRIPT ′ end_POSTSUPERSCRIPT = italic_t + 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT ∞ end_POSTSUPERSCRIPT italic_P ( italic_s start_POSTSUBSCRIPT italic_t start_POSTSUPERSCRIPT ′ end_POSTSUPERSCRIPT end_POSTSUBSCRIPT | italic_s start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT , italic_a start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ) italic_V ( italic_s start_POSTSUBSCRIPT italic_t start_POSTSUPERSCRIPT ′ end_POSTSUPERSCRIPT end_POSTSUBSCRIPT ) ](9)

where:

  • Value Function V(st)𝑉subscript𝑠𝑡V(s_{t})italic_V ( italic_s start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ): Serves as a predictive model that evaluates the expected future performance of a portfolio from a given state stsubscript𝑠𝑡s_{t}italic_s start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT, representing current market conditions, asset valuations, and portfolio composition. By calculating the cumulative expected returns, the value function incorporates both immediate gains and future prospects, considering the probabilities of various market scenarios and their timings. It guides the RL agent in assessing the long-term effectiveness of different investment strategies, taking into account factors such as market volatility, portfolio diversification, and the risk-return balance. Essentially, V(st)𝑉subscript𝑠𝑡V(s_{t})italic_V ( italic_s start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ) aids in algorithmic decision-making by projecting the future value of a portfolio under different conditions and strategies, facilitating the selection of optimal investment actions.

  • Immediate Reward R(st,at)𝑅subscript𝑠𝑡subscript𝑎𝑡R(s_{t},a_{t})italic_R ( italic_s start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT , italic_a start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ): Quantifies the instantaneous gain or loss resulting from an action atsubscript𝑎𝑡a_{t}italic_a start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT taken in state stsubscript𝑠𝑡s_{t}italic_s start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT. It reflects the immediate financial impact of trading decisions in portfolio management, such as the realized profit and loss after executing a trade. Within reinforcement learning, the concept of reward shaping is crucial, involving the design of the reward function to align with specific investment objectives and risk profiles. For instance, a hedge fund targeting absolute returns might shape R(st,at)𝑅subscript𝑠𝑡subscript𝑎𝑡R(s_{t},a_{t})italic_R ( italic_s start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT , italic_a start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ) based on log returns to emphasize direct profitability. Conversely, a strategy focused on risk-adjusted returns might utilize a reward function derived from the Sharpe ratio or its variants, thereby incorporating risk considerations. Effective reward shaping plays a pivotal role in guiding the RL agent toward investment strategies that align with investor goals, striking a balance between immediate returns and risk management.

  • Discount Factor γ𝛾\gammaitalic_γ: Determines how the RL agent weighs immediate versus future rewards. In finance, it functions similarly to the discounting of future cash flows to their present value. A higher value of γ𝛾\gammaitalic_γ indicates a stronger emphasis on long-term gains, mirroring a strategic long-term investment approach in finance. By adjusting γ𝛾\gammaitalic_γ, the RL model can be fine-tuned to prioritize either short-term gains or long-term investment objectives, aligning with the specific goals and risk tolerance of the investor.

  • Transition Probability P(st+1|st,at)𝑃conditionalsubscript𝑠𝑡1subscript𝑠𝑡subscript𝑎𝑡P(s_{t+1}|s_{t},a_{t})italic_P ( italic_s start_POSTSUBSCRIPT italic_t + 1 end_POSTSUBSCRIPT | italic_s start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT , italic_a start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ): Represents the probability of moving from the current state stsubscript𝑠𝑡s_{t}italic_s start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT to a new state st+1subscript𝑠𝑡1s_{t+1}italic_s start_POSTSUBSCRIPT italic_t + 1 end_POSTSUBSCRIPT after taking action atsubscript𝑎𝑡a_{t}italic_a start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT. In the financial context, this is similar to the probability of market shifts resulting from specific investment actions. It captures the essence of market volatility and the inherent unpredictability in financial decision-making. This concept in RL reflects the need to account for the variable nature of markets, where each action can lead to multiple potential future scenarios, each with its own likelihood and financial implications.

  • Value of Subsequent State V(st+1)𝑉subscript𝑠𝑡1V(s_{t+1})italic_V ( italic_s start_POSTSUBSCRIPT italic_t + 1 end_POSTSUBSCRIPT ): Reflects the expected utility for future states resulting from current actions. In financial terms, it is comparable to projecting the performance of a portfolio into the future, taking into account the potential outcomes of present investment decisions. This aspect of RL encapsulates the idea of forward-looking analysis in portfolio management, where the focus is not only on immediate results but also on how current choices shape future financial landscapes. It highlights the importance of strategic planning and anticipatory decision-making in both RL and financial investment, considering the long-term implications of actions taken today.

Transitioning from the traditional state-value focus of the Bellman equation, shown in Equation 9, our approach in financial portfolio optimization embraces a broader perspective. In classical Reinforcement Learning, the concept of an optimal policy, denoted as π*superscript𝜋\pi^{*}italic_π start_POSTSUPERSCRIPT * end_POSTSUPERSCRIPT, is central. This policy is typically formulated to maximize the expected cumulative reward over time, as shown in the standard RL formulation:

π*=argmaxπ𝔼π[t=0γtRt]superscript𝜋subscriptargmax𝜋subscript𝔼𝜋delimited-[]superscriptsubscript𝑡0superscript𝛾𝑡subscript𝑅𝑡\pi^{*}=\operatorname*{arg\,max}_{\pi}\mathbb{E}_{\pi}\left[\sum_{t=0}^{\infty%}\gamma^{t}R_{t}\right]italic_π start_POSTSUPERSCRIPT * end_POSTSUPERSCRIPT = start_OPERATOR roman_arg roman_max end_OPERATOR start_POSTSUBSCRIPT italic_π end_POSTSUBSCRIPT blackboard_E start_POSTSUBSCRIPT italic_π end_POSTSUBSCRIPT [ ∑ start_POSTSUBSCRIPT italic_t = 0 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT ∞ end_POSTSUPERSCRIPT italic_γ start_POSTSUPERSCRIPT italic_t end_POSTSUPERSCRIPT italic_R start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ](10)

Within the financial context, our emphasis shifts to identifying an optimal policy π*superscript𝜋\pi^{*}italic_π start_POSTSUPERSCRIPT * end_POSTSUPERSCRIPT that maximizes expected utility over an infinite horizon. This shift is encapsulated in the following formulation for Financial RL, where the optimal policy is interpreted as the optimal algorithmic investment strategy:

π*=argmaxπ𝔼π[t=0γtU(rt,σt,Xt,wt)]superscript𝜋subscriptargmax𝜋subscript𝔼𝜋delimited-[]superscriptsubscript𝑡0superscript𝛾𝑡𝑈subscript𝑟𝑡subscript𝜎𝑡subscript𝑋𝑡subscript𝑤𝑡\pi^{*}=\operatorname*{arg\,max}_{\pi}\mathbb{E}_{\pi}\left[\sum_{t=0}^{\infty%}\gamma^{t}U(r_{t},\sigma_{t},X_{t},w_{t})\right]italic_π start_POSTSUPERSCRIPT * end_POSTSUPERSCRIPT = start_OPERATOR roman_arg roman_max end_OPERATOR start_POSTSUBSCRIPT italic_π end_POSTSUBSCRIPT blackboard_E start_POSTSUBSCRIPT italic_π end_POSTSUBSCRIPT [ ∑ start_POSTSUBSCRIPT italic_t = 0 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT ∞ end_POSTSUPERSCRIPT italic_γ start_POSTSUPERSCRIPT italic_t end_POSTSUPERSCRIPT italic_U ( italic_r start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT , italic_σ start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT , italic_X start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT , italic_w start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ) ](11)

In this formulation, π*superscript𝜋\pi^{*}italic_π start_POSTSUPERSCRIPT * end_POSTSUPERSCRIPT is not merely focused on maximizing immediate returns but integrates a broader spectrum of financial metrics, including returns (rtsubscript𝑟𝑡r_{t}italic_r start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT), risk (σtsubscript𝜎𝑡\sigma_{t}italic_σ start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT), external market factors (Xtsubscript𝑋𝑡X_{t}italic_X start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT), and portfolio weights (wtsubscript𝑤𝑡w_{t}italic_w start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT). This multi-faceted approach reflects the intricate nature of financial decision-making, where balancing risk, return, and adapting to market dynamics are paramount:

  • Optimal Policy π*superscript𝜋\pi^{*}italic_π start_POSTSUPERSCRIPT * end_POSTSUPERSCRIPT: Denotes the strategy that maximizes expected utility over an infinite horizon, which is crucial in financial contexts. In Financial RL, this is comparable to identifying the most effective algorithmic investment strategy, tailored to the dynamics and complexities of financial markets.

  • Expected Utility Maximization 𝔼πsubscript𝔼𝜋\mathbb{E}_{\pi}blackboard_E start_POSTSUBSCRIPT italic_π end_POSTSUBSCRIPT: Represents the expectation of utility under policy π𝜋\piitalic_π, quantifying the long-term effectiveness of an investment strategy. This concept aligns with our earlier discussion on maximizing expected outcomes in MPT, subject to real-world constraints as detailed in Equation 6.

  • Recursive Utility Function U(rt,σt,Xt,wt)𝑈subscript𝑟𝑡subscript𝜎𝑡subscript𝑋𝑡subscript𝑤𝑡U(r_{t},\sigma_{t},X_{t},w_{t})italic_U ( italic_r start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT , italic_σ start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT , italic_X start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT , italic_w start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ): Integrates critical financial metrics at each time step t𝑡titalic_t, including returns (rtsubscript𝑟𝑡r_{t}italic_r start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT), risk (σtsubscript𝜎𝑡\sigma_{t}italic_σ start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT), external market factors (Xtsubscript𝑋𝑡X_{t}italic_X start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT), and portfolio weights (wtsubscript𝑤𝑡w_{t}italic_w start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT). This function is pivotal in reflecting the nuanced trade-offs between risk and return, aligning closely with investor objectives and their specific financial goals.

In our previous discussions, we differentiated between standard Reinforcement Learning (RL) and Deep Reinforcement Learning (DRL), noting the integration of neural networks in the latter. Now, we delve deeper into this distinction, focusing specifically on the role of neural networks and Deep Learning architectures within DRL. These sophisticated computational frameworks are pivotal in advancing portfolio management strategies, offering enhanced predictive and adaptive capabilities far beyond the scope of traditional RL. The following discussion will delve into the complexities of these neural networks and Deep Learning architectures, emphasizing their critical role in financial portfolio optimization. Renowned for their efficacy, Deep Learning architectures form the backbone of DRL, processing complex, high-dimensional, and non-stationary financial data with unparalleled efficiency. The advent of diverse architectures such as Convolutional Neural Networks (CNNs) for image and time-series data, Recurrent Neural Networks (RNNs) for sequential data, and Transformers666Transformers have gained popularity not only in time series forecasting but also in language processing tasks. They are widely used in Large Language Models (LLMs), such as those popularized in ChatGPT by OpenAI, for tasks like text generation, translation, and summarization. for complex text and language processing tasks, has revolutionized how large and noisy financial datasets are handled. These architectures are adept at extracting pertinent patterns and insights from a wide array of data types, ranging from numeric and textual to audio and visual formats, significantly broadening their applicability across diverse financial contexts. This adaptability is evidenced by their demonstrated superiority over traditional econometric models when applied to asset classes and sectors such as energy, real estate, equities, and exotic derivatives [76, 77, 78, 79, 80]. Furthermore, recent scientific advancements have extended Deep Learning’s capabilities in long-term forecasting, challenging prevailing industry misconceptions [81, 82]. This progress in Deep Learning, underpinned by a generalized mathematical formulation of neural networks, lays the foundation for understanding their operational framework in financial portfolio optimization:

Definition 2.5 (Generic Representation of Deep Neural Networks [81, 82]).

A deep neural network with L𝐿Litalic_L layers can be mathematically represented as a composition of L𝐿Litalic_L functions fi:Ei×HiEi+1normal-:subscript𝑓𝑖normal-→subscript𝐸𝑖subscript𝐻𝑖subscript𝐸𝑖1f_{i}:E_{i}\times H_{i}\to E_{i+1}italic_f start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT : italic_E start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT × italic_H start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT → italic_E start_POSTSUBSCRIPT italic_i + 1 end_POSTSUBSCRIPT, where Ei,Hi,subscript𝐸𝑖subscript𝐻𝑖E_{i},H_{i},italic_E start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT , italic_H start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT , and Ei+1subscript𝐸𝑖1E_{i+1}italic_E start_POSTSUBSCRIPT italic_i + 1 end_POSTSUBSCRIPT are inner product spaces for each i[L]𝑖delimited-[]𝐿i\in[L]italic_i ∈ [ italic_L ]. The state variables are denoted as xiEisubscript𝑥𝑖subscript𝐸𝑖x_{i}\in E_{i}italic_x start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ∈ italic_E start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT and the parameters as θiHisubscript𝜃𝑖subscript𝐻𝑖\theta_{i}\in H_{i}italic_θ start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ∈ italic_H start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT. The network’s output for an input xE1𝑥subscript𝐸1x\in E_{1}italic_x ∈ italic_E start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT is defined by a function F:E1×(H1××HL)EL+1normal-:𝐹normal-→subscript𝐸1subscript𝐻1normal-⋯subscript𝐻𝐿subscript𝐸𝐿1F:E_{1}\times(H_{1}\times\cdots\times H_{L})\to E_{L+1}italic_F : italic_E start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT × ( italic_H start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT × ⋯ × italic_H start_POSTSUBSCRIPT italic_L end_POSTSUBSCRIPT ) → italic_E start_POSTSUBSCRIPT italic_L + 1 end_POSTSUBSCRIPT, expressed as:

F(x;θ)=(fLf1)(x),𝐹𝑥𝜃subscript𝑓𝐿subscript𝑓1𝑥F(x;\theta)=(f_{L}\circ\cdots\circ f_{1})(x),italic_F ( italic_x ; italic_θ ) = ( italic_f start_POSTSUBSCRIPT italic_L end_POSTSUBSCRIPT ∘ ⋯ ∘ italic_f start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT ) ( italic_x ) ,(12)

where θ𝜃\thetaitalic_θ encapsulates the parameter set {θ1,,θL}subscript𝜃1normal-…subscript𝜃𝐿\{\theta_{1},\ldots,\theta_{L}\}{ italic_θ start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT , … , italic_θ start_POSTSUBSCRIPT italic_L end_POSTSUBSCRIPT }. This high-level formulation adapts to various architectures. Each layer function fisubscript𝑓𝑖f_{i}italic_f start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT boosts adaptability and learning, crucial for tasks like optimal portfolio allocations in RL. The head map αisubscript𝛼𝑖\alpha_{i}italic_α start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT and tail map ωisubscript𝜔𝑖\omega_{i}italic_ω start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT aid in understanding network behavior and facilitate backpropagation for neural training. In financial portfolio optimization, the final layer fLsubscript𝑓𝐿f_{L}italic_f start_POSTSUBSCRIPT italic_L end_POSTSUBSCRIPT might act as the strategic mastermind, comparable to an advanced trader, in an MDP or POMDP setting, positioning the neural network as the strategic core of these decision processes.

In summary, our exploration of Reinforcement Learning (RL) in financial portfolio management has been comprehensive. Beginning with foundational concepts such as states, actions, and rewards, we progressed to understanding the recursive nature of decision-making through the Bellman equation. This journey led us through the intricacies of Markov Decision Processes (MDPs) and Partially Observable Markov Decision Processes (POMDPs), demonstrating their application in modeling dynamic financial markets. The integration of neural networks as computational engines within these frameworks was pivotal, enabling nuanced predictions essential for portfolio optimization. However, transitioning theoretical models to real-world financial markets presents challenges beyond technical complexities. It necessitates a blend of financial expertise, AI knowledge, and regulatory understanding to ensure system robustness, rigorous model validation, and compliance with regulatory and game theoretic constraints. While RL offers solutions beyond Modern Portfolio Theory (MPT), its application in finance is not without complexities and risks, demanding careful design and implementation to balance innovation with risk management. Having established a solid understanding of RL’s theoretical and practical aspects in finance, we now turn to historical parallels between technology and finance. This exploration sets the stage for our novel sim-to-real frameworks, grounded in RL advancements, aiming to revolutionize portfolio management. Through empirical testing via simulations, we aim to validate our theoretical propositions and demonstrate the real-world viability of our innovative approaches.

3 Advancing Portfolio Management: Novel Frameworks in Reinforcement Learning

As we embark on presenting our novel frameworks in the realm of financial portfolio management, this section first establishes a foundational context, underscoring how modern finance has historically drawn inspiration from various scientific and technological disciplines. By delving into the intersections between technology and finance, we aim to provide a well-grounded rationale for our innovative approaches, rooted in the rich tapestry of interdisciplinary influences that have shaped financial strategies over time. Our exploration begins by highlighting the parallels between financial strategies and technological advancements, with a particular focus on recommendation systems. Both fields have navigated similar optimization challenges, employing Reinforcement Learning (RL) as a transformative tool. This convergence underscores RL’s potential to significantly impact finance, a domain that has traditionally embraced methods from physics and probability theory, notably in areas such as options pricing and risk management [70, 83]. Building on this historical context, we then introduce our novel simulation-to-real (sim-to-real) frameworks. These frameworks, inspired by advancements in robotics and mathematical physics, are specifically designed to address the complexities of modern financial portfolio management. They represent a strategic and innovative response, reflecting finance’s history of adapting cross-disciplinary methodologies. This is crucial given the high-stakes nature of and unique challenges faced by the financial sector. The final computational experiments aim to rigorously evaluate our sim-to-real frameworks, gauging their potential impacts on the industry. Positioned at the cutting edge of current Artificial Intelligence research, these frameworks aim to stimulate discussion and the development of scalable, efficient, and practical solutions in finance. Our approach, rooted in historical precedents and cross-disciplinary insights, anticipates a future where finance is increasingly influenced by advanced algorithms. In advocating for their responsible and effective application, we emphasize the need for these innovative technologies to be meticulously designed to meet the complex and highly regulated demands of modern international finance.

3.1 Parallel Evolution: Reinforcement Learning in Finance and Technology

Building on our earlier discussions of Modern Portfolio Theory (MPT) in finance, this subsection examines the striking similarities between optimization processes in finance and technology, especially when confronted with predefined constraints. In finance, Modern Portfolio Theory (MPT) aims to maximize returns by strategically allocating investments across a mix of assets such as stocks, bonds, and derivatives, balancing risk and potential gains. Each choice is subject to a myriad of real-world constraints including risk profiles, market conditions, and regulatory frameworks. In finance, the optimization challenge resembles the task encountered in technology’s recommendation systems. These systems are designed to construct portfolios of digital content, encompassing a diverse range of offerings from products to movies and social media advertisem*nts. Their objective is twofold. Firstly, to maximize expected revenue for the firm by strategically aligning offerings with user interests. Secondly, to enhance the user’s experience by precisely tailoring recommendations. This user-centric approach aims to ensure that users are presented with content that resonates with their preferences, geographical locations, and personal interests. The success of these recommendation engines is measured not only by traditional engagement metrics like click-through rates and viewing time but also by the degree to which they fulfill user expectations and drive revenue growth. This meticulous strategy in content curation mirrors the principles of portfolio optimization in finance, where the goal is to balance risk and return to meet specific investment objectives. The historical development of both finance and technology reveals profound methodological and mathematical parallels. In finance, numerical optimization methods under MPT are employed to derive the best risk-adjusted returns for portfolios. This mirrors technology’s early recommendation systems, which used mathematical techniques such as matrix factorization to predict user preferences for digital content [84, 85, 86, 87]. Central to both domains is the concept of utility maximization. In finance, utility functions are employed to balance risk against return, aligning with investor objectives. Conversely, in technology, utility functions gauge user satisfaction, focusing on digital engagement and content preferences. Furthermore, both fields confront analogous challenges. The cold start problem in finance, characterized by initiating portfolio allocations without prior data (in the case of new financial products), is comparable to the challenge in technology of recommending content to new users who lack engagement history. Additionally, the issue of sparse rewards is a common obstacle in both domains [88, 89]. In finance, especially in contexts such as options trading, the concept of sparse rewards is manifested in the form of delayed financial outcomes. Until the expiry of an option or its exercise, the profit and loss (PnL) remains uncertain. This delay in realizing PnL can range from minutes to days or even months, depending on the nature of the investment. Similarly, in the realm of technology, particularly in recommendation systems, sparse rewards are seen as delayed monetary returns. The financial benefit for a technology firm materializes only when a user takes a definitive action, such as clicking on an advertisem*nt or making a purchase from recommended products. Here too, the delay before seeing PnL can vary significantly, from immediate responses to prolonged periods of user engagement. These parallels in problem-solving structures and strategies between financial portfolio management and technology-based recommendation systems underscore a fundamental mathematical symmetry. This symmetry provides a logical foundation for hypothesizing the potential for Reinforcement Learning (RL) to transform finance, drawing on its successful applications in technology. This analysis not only reinforces the connection between these domains but also highlights the potential for cross-domain insights to inform future advancements in financial portfolio optimization.

For financial professionals, this parallel offers valuable insights. The evolution and resolution of these challenges in the technology sector, particularly through the adoption of advanced Reinforcement Learning techniques, suggest a potential roadmap for similar advancements in financial portfolio management. By acknowledging and learning from the technological sector’s approach to these analogous problems, finance professionals can envisage and prepare for potential revolutionary shifts in portfolio management strategies. For those within the finance sector who may harbor skepticism about the transformative potential of advanced algorithms, the evolution and successes of RL in technology present a compelling case. As advancements in technology continued to evolve, the classical methods in both portfolio optimization and recommendation systems encountered significant limitations in addressing the dynamic and complex nature of their respective domains. This is particularly evident in portfolio optimization, where conventional methodologies rooted in Modern Portfolio Theory (MPT) often face challenges adapting to the realistic constraints and high-dimensional, noisy data characteristics of modern markets. The technology sector, in its quest for more adaptive and scalable solutions, increasingly embraced Deep Learning and Reinforcement Learning. RL’s capability to learn from user interactions and dynamically adapt strategies proved to be evolutionary in addressing the intricate challenges of recommendation systems. The shift from static, algebraic solutions to dynamic, data-driven RL approaches indicates a potential evolutionary path for financial portfolio management. Notably, in the technology sector, issues once deemed computationally intractable have been effectively resolved by elite teams of mathematicians and physicists at major technology firms. These groups leverage innovative RL systems, achieving high efficiency, low latency, and scalability, thus enabling real-time recommendations for millions of users [90, 91, 92, 93, 94, 95].

This historical parallel, where two distinct fields faced similar mathematical challenges and evolved towards RL-based solutions, strongly suggests that the future of finance is poised to be reshaped by these advanced algorithmic approaches [92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102]. Much like how advanced algorithms revolutionized the advertising industry - a field once dominated by marketing and business experts - finance is poised to undergo a similar transformation. Sophisticated algorithmic solutions may challenge and reshape traditional practices, potentially altering the roles and tasks traditionally associated with investors. In envisioning a future increasingly driven by advanced algorithms in finance, we assert the need for prudence and sustainability in their application, especially given the stringent regulatory environment and the immense responsibility of managing client funds. Given the unique complexities of the financial sector and the responsibility of managing substantial funds, the deployment of Reinforcement Learning techniques must be conducted with meticulous care. Unlike recommendations in social media or e-commerce, errors in financial portfolio optimization can lead to risk exposures amounting to hundreds of millions or billions of dollars, with far-reaching consequences for investment firms, financial institutions, and sovereign wealth funds. Thus, the integration of RL into finance necessitates not only advanced algorithmic development but also a profound comprehension of market nuances and robust risk management protocols. Particularly in leveraged scenarios, the potential financial ramifications underscore the importance of rigorous risk assessment and stress testing practices.

Collaboration is crucial in these practices, necessitating alignment among scientific, quantitative, risk, and regulatory teams to ensure RL’s responsible utilization in high-stakes environments. Furthermore, academic researchers in this field are obliged to communicate assumptions and potential risks transparently. In integrating advanced technologies like Reinforcement Learning into financial portfolio optimization, effective collaboration among engineers, scientists, finance and legal professionals is paramount. This multidisciplinary synergy is vital, not only to drive innovation but also to ensure thoughtful application of RL within the highly regulated and complex landscape of finance. In the technology sector, the popular adage move fast and break things encapsulates a culture that values rapid innovation and risk-taking. However, when considering the application of similar technological advancements in finance, this approach necessitates cautious reconsideration. A reckless importation of this ethos into the financial domain could lead to dire consequences, potentially sparking severe financial crises and triggering stringent regulatory responses, which could hinder future innovation. This paper advocates for a more deliberate and sustainable approach - move prudently and build sustainably - as a guiding principle in the development of financial technology. Our aim is to spark and contribute to discussions about the future of financial technology, highlighting the need for a balanced approach that recognizes the unique risks and responsibilities of the financial sector. As we progress, we will delve into hypothetical case studies that offer simplified yet insightful perspectives on the unique challenges faced in finance as opposed to technology, emphasizing the importance of an interdisciplinary, cautious and collaborative approach.

Practical Implications in Financial Portfolio Management: In applying Reinforcement Learning to portfolio optimization, it’s crucial for academic and industry research to transparently justify methodologies and reasoning. The diverse nature of financial markets and regulatory environments means that theoretically sound strategies might vary significantly based on specific market conditions and investor objectives. Consequently, the developers of such mathematical and computational artifacts must clearly articulate the assumptions underpinning the chosen approach and their respective rationale. This clarity is vital for understanding the real-world applicability and constraints of these sophisticated mathematical models in the complex landscape of financial portfolio management and optimization777The development of RL systems for financial applications is an intricate endeavor that encompasses detailed mathematical modeling and considerable computational efforts. It demands not only expertise in algorithmic design but also adherence to stringent software engineering standards. Minor implementation errors can lead to significant operational complications, underscoring the need for precision in both the mathematical and software engineering facets of RL. This highlights the importance of creating robust, reliable, and efficient systems for high-stakes financial applications.. To illustrate these points, we present a few simplified open-ended questions, highlighting the significant complexity of building industry-grade RL systems in portfolio optimization and financial RL more broadly:

  1. 1.

    Understanding Time Horizon & Market Microstructure: The transferability of an RL agent from one market domain to another poses significant challenges. Take, for instance, an RL agent (algorithm) developed for portfolio optimization in the energy derivatives market. The question arises: is it appropriate or even feasible to apply this agent to a disparate context such as a cross-asset strategy? Such a shift would involve transitioning from a market with specific characteristics to one with potentially vastly different microstructures and trading dynamics. This case underscores the need for domain-specific knowledge and strategy alignment. Moreover, the time horizon of the trading strategy plays a crucial role. An RL agent trained for high-frequency trading, where decisions are made on a millisecond basis, may not be suitable for long-term, long-only investment strategies. This distinction is rooted in the unique market microstructures and investment objectives associated with different trading styles and asset classes. It exemplifies the importance of understanding both the time horizon and the microstructural nuances of the market when deploying RL agents in finance [103, 104, 105].

  2. 2.

    Understanding Dynamical Systems & Non-Stationarity: The adaptability of RL agents to dynamic market conditions is put to the test in scenarios such as the onset of the COVID-19 pandemic. An agent trained in pre-pandemic market conditions might face significant challenges in adjusting to the subsequent, drastically altered market dynamics. This raises critical questions: has the agent been rigorously stress-tested to handle the volatility and shifts in market behavior seen post-COVID? Are the risk metrics utilized by the agent, sufficiently robust and adaptable to suit different investment styles and client risk profiles in this new market landscape? This scenario accentuates the inherent difficulty in dealing with non-stationary environments in financial markets, a problem that even state-of-the-art RL models grapple with. It underscores the importance of a comprehensive design of experiments [44, 45, 46, 47, 48], and continual adaptation and stress testing in financial RL applications. Such a meticulous approach is crucial for developing RL systems that are not only theoretically sound but also resilient to market evolution.

  3. 3.

    Understanding Asymmetric Risk & Game Theory Dynamics: A critical aspect of applying RL in finance is navigating asymmetric risk profiles and the game-theoretic nature of financial markets. This complexity is rooted in the constantly changing rules and dynamics that govern financial systems, unlike the more stable natural laws observed in the natural sciences. In the financial world, the introduction of new regulations, sudden market shocks, or strategic moves by major market players can drastically alter the market environment. This fluidity presents unique challenges for RL agents, which must be capable of understanding and responding to asymmetric risks and the strategic behavior of other market participants. It leads to a scenario where the data available for training and validation of these agents is not only limited but also rapidly becomes outdated as market dynamics evolve. This situation renders financial RL arguably more complex than many other domains, making the development of these respective systems among the most intricate. The crucial question then becomes: are the RL agents designed to account for and adapt to these asymmetric risks and game-theoretic complexities? [106, 107, 108]

This discussion sheds light on the nuanced notion of fully autonomous agents in finance, which, upon closer examination, appears somewhat of a misnomer. Finance, with its intricate web of risk management and regulatory compliance, necessitates a human-in-the-loop (HIL) approach, essential for overseeing and validating algorithmic decisions [109, 110, 111, 112]. Taylor et al. (2023) articulate this necessity well: “[One should] not think of RL as a fully autonomous paradigm, but instead as an iterative learning and development process involving both learning algorithms and humans. While better algorithms may chip away at this assumption, full autonomy in terms of problem identification, construction, and deployment is unlikely in the near-future. Instead, we argue that it is critical to consider how humans and RL agents can work together to solve sequential decision tasks[110]. In concluding this section, we have emphasized the vital role of human insight and expertise in guiding and understanding the application of Reinforcement Learning (RL) in portfolio optimization and broader financial contexts. Effective RL systems in finance must navigate a landscape filled with uncertainties that often extend beyond the scope of sophisticated statistical algorithms. The presented scenarios highlight this blend as critical for the successful and responsible application of RL in finance.

3.2 Simulation-to-Real Transfer in financial RL: Bridging the Reality Gap

Drawing upon the successful application of Reinforcement Learning in high-risk domains like mathematical physics and robotics, we introduce our novel Simulation-to-Real (Sim-to-Real) frameworks for financial portfolio optimization. This advancement is underpinned by the interdisciplinary convergence of applied mathematics, computational science, and finance. These fields collectively reveal synergies and analogous methodologies, shaping our algorithmic solutions for financial Reinforcement Learning. Central to our approach is Sim-to-Real transfer learning, a method pivotal in the development of high-stakes, autonomous systems such as self-driving cars. Here, RL systems are exhaustively tested in simulated settings that closely mimic diverse real-world scenarios. This methodical simulation is critical in preparing the systems to navigate and adapt to the multifaceted and unpredictable nature of real-world environments, effectively narrowing the reality gap – the often observed divergence between simulated models and real-world outcomes. In the realm of finance, the reality gap is characterized by the unique challenges posed by fluctuating market conditions, regulatory changes, and trading constraints. Our RL models for financial portfolio optimization are meticulously crafted to be resilient, adaptable, and progressively evolving in response to the dynamic nature of financial markets. We emphasize a strategic application of RL, deeply rooted in the practical realities of market dynamics and regulatory frameworks. By integrating Sim-to-Real principles from sectors like robotics and mathematical physics, our goal is to significantly enhance the practicality and flexibility of RL in financial portfolio optimization. This approach not only aligns with our broader research objectives of fusing theoretical insights with real-world market complexities but also advocates for meticulous engineering, comprehensive domain expertise, and the responsible implementation of financial RL systems. Such a strategy echoes the critical insights gained from robotics, underscoring the importance and intricacies of deploying Sim-to-Real transfer in the financial sector:

"There was wide agreement that the ultimate goal is to design robotic systems that live in the real-world …A key aspect in sim-to-real transfer is the choice of simulation. Independently of the techniques utilized for efficiently transferring knowledge to real robots, the more realistic a simulation is the better results that can be expected …Reinforcement learning algorithms often rely on simulated data to meet their need for vast amounts of labeled experiences. The mismatch between the simulation environments and real-world scenarios, however, requires further attention to be put to methods for sim-to-real transfer of the knowledge acquired in simulation" - Perspectives on sim2real transfer for robotics, Höfer et al (2020) [113].

The advancements in high-risk robotics, as discussed by Höfer et al (2020), have been achieved through addressing significant technical challenges and innovations. As we apply sim-to-real constraints and methodologies similar to those in robotics to our financial RL models, we anticipate encountering comparable challenges. This underscores the need for innovation and meticulous application in finance, a domain with considerably high stakes. Our literature review on Reinforcement Learning and Artificial Intelligence in robotics has led to the formulation of three fundamental principles. These principles are not only grounded in advanced computational science but are also acutely aware of the complexities and challenges unique to financial markets. Our objective is to establish a preliminary framework tailored to the specific needs of financial portfolio management, while also adhering to the stringent regulatory frameworks that govern this sector. Drawing on insights from a wide array of fields, including robotics and mathematical physics [114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144], we now introduce our novel Simulation-to-Real (Sim-to-Real) frameworks for financial RL. These frameworks are distinctively designed to address the specific challenges and nuances of financial RL use-cases:

  1. 1.

    Realistic Reward Shaping & Environmental Modeling: Central to our sim-to-real transfer methods in high-risk financial applications is the precise construction of financial environments mirroring real-world market conditions. This requires in-depth modeling of transaction costs, market impact, and regulatory constraints, essential for effective RL strategies. Recent academic efforts, like using Generative Adversarial Networks (GANs) for market data generation, highlight the challenge of accurately replicating market dynamics. Without realistic simulation, such techniques risk being ineffective. The domain randomization approach in robotics, where RL models face diverse simulated conditions to build resilience, is enhanced by incorporating concepts like concentrability, which ensures training on a representative sample of the state space, and coverability, which guarantees exposure to a wide range of market scenarios including rare but critical (black-swan) events. These additions are crucial for RL systems to effectively navigate and adapt to the volatile nature of financial markets. Reward shaping is equally crucial, requiring the selection of metrics such as log returns or the Sortino ratio, tailored to specific financial goals and investor profiles. This domain-specific knowledge is imperative for aligning RL agent actions with investor objectives. Furthermore, robust software engineering is key, particularly with open-source financial modeling libraries. The final RL software must be theoretically sound, resilient, and scalable, capable of operating efficiently in diverse computational environments, including industrial-grade investment settings.

  2. 2.

    Robust Risk Analysis & Statistical Stress Testing: In the diverse and complex landscape of finance, it is imperative to design Reinforcement Learning systems that are specifically tailored to the unique characteristics of different asset classes and market microstructures, rather than attempting to create universally applicable models. This nuanced approach necessitates a comprehensive risk management strategy that incorporates conventional risk metrics, fine-tuned for particular asset classes and market sectors. A critical aspect of this approach is conducting extensive statistical analyses of performance metrics across a wide range of RL training scenarios. An active area of RL research also includes advanced causality-driven model explainability techniques, such as counterfactual explanations, which can significantly enhance understanding and trust in these systems. Emphasizing the importance of rigorous software engineering, especially for those transitioning from the computer science domain to financial applications, our methodology includes the utilization of advanced tools for serialization and cloud-based storage of performance data. This practice not only facilitates easy retrieval of crucial data during regulatory audits or client interactions but also ensures a high degree of transparency and accountability. By carefully considering these factors, we aim to develop RL systems that are not only robust in theoretical simulations but also adaptable and reliable in the real-world financial ecosystem, respecting the intricacies of each asset class and market microstructure.

  3. 3.

    Strategic Resilience & Fault Tolerance-Aware Engineering: Drawing parallels from high-risk robotics, such as autonomous vehicle systems, financial RL systems require robust fault tolerance mechanisms to anticipate and effectively manage market extremities and external shocks. This entails engineering contingency protocols for scenarios where algorithms might malfunction or underperform, particularly during volatile market conditions. Leveraging strategies utilized in autonomous robotics to handle unpredictable events, financial RL systems need proactive design considerations for navigating market anomalies and systemic fluctuations. The burgeoning threat of adversarial attacks, particularly relevant for RL models processing sentiment data from fluctuating sources like social media, must be addressed. These models require rigorous vulnerability assessments to prevent potential exploitations that could adversely affect financial outcomes. A judicious approach might involve selectively excluding data sources where the risk-to-reward ratio is unfavorable. This proactive stance on system security and fault tolerance mandates exhaustive risk assessments, encompassing both traditional financial risks and those unique to algorithmic investments, such as data manipulation and model over-fitting. Integrating these comprehensive risk management strategies from the outset ensures the development of financial RL systems that are not only resilient but which also adhere to the stringent safety and adaptability standards observed in high-risk robotics applications.

In our exploration of sim-to-real transfer methodologies from high-risk domains such as robotics to financial Reinforcement Learning (RL), we have established preliminary yet foundational guidelines. These guidelines are devised to guide the development of sophisticated RL strategies in the nuanced domains of portfolio optimization and broader financial RL. While they represent a significant initial contribution, it is essential to recognize that they are just the inception of a broader journey in this evolving field. These principles, we hope, will ignite ongoing discussions and foster safe yet innovative advancements, especially resonating with stakeholders in financial institutions, governments, and policymaking bodies worldwide. As the financial sector increasingly integrates complex algorithms, comprehensively understanding and addressing the multifaceted nature of these systems is vital. Our framework, while thorough in its current form, is not exhaustive and represents an initial step in aligning theoretical RL models with the dynamic and unpredictable nature of financial markets. With a focus on environmental realism, adaptability to market changes, and resilient system design, our work is committed to developing RL methodologies that can effectively meet the diverse challenges of the financial sector. Designed for both theoretical robustness and practical applicability, these methodologies reflect our dedication to pragmatic and effective solutions in financial management. As we conclude this part of our research, we approach a crucial stage; empirical validation of our theories through computational experiments with RL agents and our proprietary systems and market environments.

4 Computational Experiments

Experiment Focus & Evaluation Metrics: Our computational experiments aim to rigorously evaluate the novel simulation-to-real framework, as delineated in previous sections. This framework is not only centered on applying state-of-the-art (SOTA) Reinforcement Learning algorithms, recognized for their effectiveness in literature under conditions of low or negligible transaction costs, but also on a comprehensive consideration of real-world financial factors. Our experimental design integrates key real-world factors, including market frictions and broker fees, surpassing traditional academic models to provide a realistic simulation of financial market conditions. It’s important to note that certain implementation details are omitted due to the proprietary nature of our algorithms and system design. This led us to adopt a simplified experimental design that emphasizes a holistic, realism-driven approach to financial market simulations. We believe that this approach, while possibly affecting algorithmic performance, provides invaluable insights for both researchers and practitioners in industry and academia. As discussed in the previous sections, we posit that from a practitioner’s standpoint, an algorithm rigorously stress-tested in realistic settings holds greater value, even if it demonstrates suboptimal performance. This is in contrast to algorithms that excel in academic research but may not withstand the complexities of real-world financial environments. Our work aims to bridge this gap, offering a framework that balances theoretical robustness with practical applicability. Additionally, our experiment focus, is on specifically tailored to long-term horizon asset managers and market makers. This distinction is vital, as the strategies and constraints relevant to these entities often differ markedly from those employed by independent retail proprietary traders or entities seeking pure alpha returns. By being transparent about this focus, we ensure that our experimental design aligns closely with the operational realities and regulatory considerations unique to these segments of the financial industry. These experiments utilized high-performance cloud computing and advanced GPUs, conforming to industrial-grade software engineering practices. As part of our evaluation methodology, we incorporate a comprehensive set of performance metrics for asset-class agnostic evaluation, as suggested by Lim, et. al (2019) [145] and Zhang et. al (2020) [100]. These metrics collectively offer a comprehensive and nuanced view of portfolio performance, addressing various dimensions of risk and return, crucial for a holistic evaluation of any trading strategy:

MetricDefinitionExplanation

𝔼[R]𝔼𝑅\operatorname{\mathbb{E}}[R]blackboard_E [ italic_R ]

Annualized expected trade returns.

Indicates potential profitability and sets a baseline for performance expectation.

std(R)std𝑅\text{std}(R)std ( italic_R )

Annualized standard deviation of trade returns.

Measures the volatility of returns, crucial for understanding the risk level of the strategy.

DD

Annualized standard deviation of negative trade returns.

Captures the strategy’s downside risk, critical for gauging potential losses.

Sharpe

Sharpe ratio: 𝔼[R]/std(R)𝔼𝑅std𝑅\operatorname{\mathbb{E}}[R]/\text{std}(R)blackboard_E [ italic_R ] / std ( italic_R ).

Evaluates risk-adjusted returns, allowing comparison of performance against the risk taken.

Sortino

Sortino ratio: 𝔼[R]/DD𝔼𝑅DD\operatorname{\mathbb{E}}[R]/\text{DD}blackboard_E [ italic_R ] / DD.

Focuses on downside risk, offering a refined view of risk-adjusted performance.

MDD

Maximum drawdown

Indicates the largest observed loss from peak to trough, essential for assessing strategy resilience.

Calmar

Calmar ratio: 𝔼[R]/MDD𝔼𝑅MDD\operatorname{\mathbb{E}}[R]/\text{MDD}blackboard_E [ italic_R ] / MDD.

Relates annual return to maximum drawdown, providing a perspective on performance during adverse conditions.

% +ve

Percentage of positive trade returns.

Reflects the strategy’s consistency in yielding profits, a key aspect for investment decision-making.

Avg+/Avg-

Ratio of average positive to negative trade
returns.

Shows the balance between average gains and losses, indicating overall trade effectiveness.

Reward Function: In evaluating performance, we adopted the Differential Sharpe Ratio (DSR) as our reward function. Articulated in the classic Financial RL paper by Moody and Saffell (1998), the DSR can be formulated as follows [146]:

Dt=Bt1ΔAt12At1ΔBt(Bt1At12)32,subscript𝐷𝑡subscript𝐵𝑡1Δsubscript𝐴𝑡12subscript𝐴𝑡1Δsubscript𝐵𝑡superscriptsubscript𝐵𝑡1subscriptsuperscript𝐴2𝑡132D_{t}=\frac{B_{t-1}\Delta A_{t}-\frac{1}{2}A_{t-1}\Delta B_{t}}{(B_{t-1}-A^{2}%_{t-1})^{\frac{3}{2}}},italic_D start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT = divide start_ARG italic_B start_POSTSUBSCRIPT italic_t - 1 end_POSTSUBSCRIPT roman_Δ italic_A start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT - divide start_ARG 1 end_ARG start_ARG 2 end_ARG italic_A start_POSTSUBSCRIPT italic_t - 1 end_POSTSUBSCRIPT roman_Δ italic_B start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT end_ARG start_ARG ( italic_B start_POSTSUBSCRIPT italic_t - 1 end_POSTSUBSCRIPT - italic_A start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_t - 1 end_POSTSUBSCRIPT ) start_POSTSUPERSCRIPT divide start_ARG 3 end_ARG start_ARG 2 end_ARG end_POSTSUPERSCRIPT end_ARG ,(13)

where Atsubscript𝐴𝑡A_{t}italic_A start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT and Btsubscript𝐵𝑡B_{t}italic_B start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT represent the exponential moving averages of the returns and squared returns, respectively. ΔAtΔsubscript𝐴𝑡\Delta A_{t}roman_Δ italic_A start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT and ΔBtΔsubscript𝐵𝑡\Delta B_{t}roman_Δ italic_B start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT are the changes in these averages at time t𝑡titalic_t. This metric, utilizing exponential moving averages, is particularly suited for the dynamic and online nature of RL environments, offering immediate, risk-adjusted performance feedback vital for RL decision-making888The DSR’s denominator, raised to the power of 3232\frac{3}{2}divide start_ARG 3 end_ARG start_ARG 2 end_ARG, emerges from its first-order expansion, capturing the immediate impact of current returns on the Sharpe Ratio. This formulation is unique to the DSR, specifically designed for online applications [147].. The DSR’s focus on consistent returns and effective risk management aligns well with our goals of guiding the agent toward risk-adjusted strategies similar to those of a long-only asset manager. Different reward metrics might be preferred for strategies focusing on absolute returns or conservative approaches, emphasizing the need to tailor the reward function to specific investment objectives and regulatory requirements.

Data Selection Rationale: The data selection process for our study was meticulously planned to ensure both the validity and practical applicability of our models. We selected 11 high-volume U.S. equities and indexes from varied sectors including industrials, energy, financials, and consumer goods. This choice was driven by two main factors: Firstly, to minimize market impact and transaction costs. High-volume stocks were selected for their liquidity, which is crucial in real-world trading environments, reducing the potential market impact and costs associated with trading activities. Secondly, our focus was on enabling effective long-term horizon portfolio allocation strategies. These stocks and indexes are well-suited for strategies involving real-world Exchange Traded Funds (ETFs) and individual stocks. By opting for a low cardinality set of assets, as supported by existing literature, we aimed to focus on the sim-to-real dynamics within a controlled yet realistic market setting, avoiding excessive market frictions. Additionally, we started with a moderate initial cash amount and explicitly excluded the use of leverage or borrowing, to further reduce market impact and align with real-world trading constraints.

Reinforcement Learning Models and Experimental Results: We have carefully chosen a set of Reinforcement Learning (RL) algorithms that have been demonstrated to surpass traditional investment strategies, such as those based on Modern Portfolio Theory (MPT), as discussed in earlier sections999While the detailed technical intricacies of these algorithms are beyond the scope of this paper, the provided intuitive explanations aim to highlight their unique attributes and operational analogies in the context of financial portfolio management. It is vital to note the challenges inherent in replicating financial RL experiments, especially regarding computational resources. Studies from Wang et. al (2021) and others [148, 149, 150, 151] highlight the importance of advanced computing power, vectorized training, and sophisticated memory buffer systems for scaling experiments and improving computational efficiency. Variations in hardware, software, and statistical experimentation design can lead to different outcomes, even with seemingly similar methodologies, features and model parameters.. For clarity, especially for readers well-versed in finance terminology, we reiterate that the term policy within this context refers to an algorithmic investment strategy:

  • Advantage Actor-Critic (A2C): Technically, A2C operates with a dual architecture, combining a policy model (the actor) and a value model (the critic). It simultaneously learns a policy and a value function, balancing immediate rewards with long-term value. This makes A2C efficient in high-dimensional action spaces. Intuitively, think of A2C as having a trader (actor) making decisions, with a risk management team (critic) evaluating and guiding these decisions for optimal outcomes.

  • AlphaOptimizerNet Simplified (AONS): Our proprietary AONS algorithm, which is based on more recent developments in the state-of-the-art (SOTA) Artificial Intelligence literature was originally designed for proprietary high-frequency intra-day research. The neural architecture was adapted and simplified for our long-term asset management experiments.

  • Proximal Policy Optimization (PPO): PPO is innovative in its approach to updating policies, using a clipped objective function to prevent large, destabilizing updates. It maintains a balance between exploring new strategies and exploiting known ones, crucial in unpredictable financial markets. Intuitively, PPO can be seen as an adaptive trader, constantly refining strategies based on market feedback and risk-return profiles.

  • Soft Actor-Critic (SAC): SAC employs a stochastic policy framework and off-policy learning, effectively exploring diverse strategies. This makes it adept at handling market uncertainty and volatility. Intuitively, SAC resembles a trader using probabilistic models to anticipate and adapt to various market conditions, enhancing strategy adaptability.

Table 2 details the performance metrics for each RL agent during this out-of-sample backtesting phase, providing insights into their ability to navigate the complexities of the financial market landscape. Our out-of-sample backtesting yields a nuanced perspective on the applicability of reinforcement learning (RL) models under real-world market constraints. Although RL models are often touted for their potential to surpass classical investment strategies, our findings suggest that their efficacy might be compromised under realistic market conditions. These results preliminarily support our hypothesis that our novel sim-to-real considerations may be crucial for the practical deployment of RL models in financial environments, beyond being merely additive.

Model𝐄[𝐑]𝐄delimited-[]𝐑\mathbf{E[R]}bold_E [ bold_R ]std(R)DDSharpeSortinoMDDCalmarAvg+/Avg-
A2C0.13820.13820.13820.13820.01520.01520.01520.01520.01510.01510.01510.01510.79270.79270.79270.79271.17381.17381.17381.17380.22760.2276-0.2276- 0.22760.52480.52480.52480.52480.98260.9826-0.9826- 0.9826
AONS0.23090.2309\mathbf{0.2309}bold_0.23090.01610.01610.01610.01610.01570.0157\mathbf{0.0157}bold_0.01571.14591.1459\mathbf{1.1459}bold_1.14591.72441.7244\mathbf{1.7244}bold_1.72440.23950.2395\mathbf{-0.2395}- bold_0.23950.82910.8291\mathbf{0.8291}bold_0.82911.03731.0373\mathbf{-1.0373}- bold_1.0373
PPO0.14160.14160.14160.14160.01520.01520.01520.01520.01520.01520.01520.01520.80940.80940.80940.80941.18971.18971.18971.18970.25110.2511-0.2511- 0.25110.48760.48760.48760.48760.98410.9841-0.9841- 0.9841
SAC0.01150.01150.01150.01150.01560.01560.01560.01560.01620.01620.01620.01620.18100.18100.18100.18100.25570.25570.25570.25570.27760.2776-0.2776- 0.27760.03600.03600.03600.03600.91200.9120-0.9120- 0.9120
Advancing Investment Frontiers: Industry-grade Deep Reinforcement Learning for Portfolio Optimization (2)

In particular, our proprietary AlphaOptimizerNet Simplified (AONS) has demonstrated significant adaptability within recent market regimes. AONS posted an impressive 23.09% return (𝔼[R]𝔼𝑅\operatorname{\mathbb{E}}[R]blackboard_E [ italic_R ]), indicating its strong performance in generating profits. This high return rate is noteworthy, especially when compared with other models like A2C, PPO, and SAC, as it highlights AONS’s ability to adapt to market shifts and achieve superior returns. When considering risk-adjusted returns, AONS stands out with a Sharpe ratio of 1.1459 and a Sortino ratio of 1.7244. These metrics suggest that AONS efficiently balances returns with the risks it takes. Moreover, its Calmar ratio of 0.8291, despite a Maximum Drawdown (MDD) of -23.95%, signals a strong recovery potential from losses. It’s important to note that MDD, a crucial risk metric, is not annualized. Instead, it measures the largest single drop from peak to trough in the portfolio’s value, offering insight into the downside risk over the entire investment period. AONS’s negative Avg+/Avg- ratio indicates an aggressive trading strategy, contrasting with the more passive approaches often found in classical Modern Portfolio Theory (MPT). This aggressive approach, focusing on capitalizing on market trends, is a significant departure from traditional strategies and highlights the dynamic nature of RL strategies in finance. Overall, while models like A2C, PPO, and SAC also achieved positive returns, they did not match AONS’s high returns and risk-adjusted performance metrics. This underscores the effectiveness of RL in adapting to market changes and outperforming traditional single-period MVO solutions. Figure 5 illustrates AONS’s dynamic portfolio allocation strategy, exemplifying the model’s proficiency in adjusting to short-term market trends. Such visualizations, coupled with an emphasis on model explainability, are crucial in a field where many financial AI studies tend to neglect the temporal evolution of trading strategies. Our replication attempts of these studies have revealed a trend where models often default to simpler strategies like Uniform Buy and Hold, emphasizing the importance of incorporating realistic market dynamics into financial AI research. Despite the promising results, they warrant cautious scrutiny, particularly in high-stakes financial domains. The aggressive stance of AONS may elicit concerns within traditional institutional risk management frameworks, reinforcing the need for judicious deployment of such models. The success of AONS, along with the positive outcomes from other RL models, calls for ongoing refinement and empirical validation to ensure their practical applicability in the ever-changing financial market landscape. Our study contributes to the discourse on integrating sophisticated AI techniques into portfolio management, advocating for a balanced approach that fuses technical ingenuity with pragmatic execution.

Advancing Investment Frontiers: Industry-grade Deep Reinforcement Learning for Portfolio Optimization (3)

5 Conclusion

In conclusion, this paper marks a substantial step towards achieving our initial goals. We have successfully developed the AlphaOptimizerNet Reinforcement Learning agent. Furthermore, we have explored the critical aspects of our novel sim-to-real-world transfer frameworks, contributing to the discourse on advanced AI applications in portfolio management. Our preliminary frameworks and extensive computational experiments are designed to serve as a cornerstone for future research and practical implementations in this field. We aspire for our findings and methodologies to elevate financial management standards and inform future explorations. Particularly for financial institutions managing substantial assets, our work offers a methodically designed reference point, potentially enhancing their wealth safeguarding strategies and ability to limit their exposure to risk. Our commitment to advancing this domain remains unwavering, as we continue to innovate in neural network architectures and refine our proposed frameworks. Our ultimate goal is to foster a financial ecosystem that is not only efficient and resilient but also attuned its broader impact on individual prosperity and collective economic growth. We hope that our current and future contributions will significantly advance investment frontiers, bridging theoretical research and practical applications.

Disclaimer: The authors make no representation or warranty, express or implied, and disclaim all liability regarding the completeness, accuracy, or reliability of the information contained in this paper. This document is not intended as investment research or advice, nor as a recommendation, offer, or solicitation for the purchase or sale of any security, financial instrument, product, or service.

References

References

  • [1]Philip Ndikum“Machine learning algorithms for financial asset priceforecasting”In arXiv preprint arXiv:2004.01504, 2020
  • [2]Paul Wilmott“Where quants go wrong: a dozen basic lessons in commonsensefor quants and risk managers and the traders who rely on them”In Wilmott Journal 1.1Wiley Online Library, 2009, pp. 1–22
  • [3]David H Bailey and Marcos Prado“Finance is Not Excused: Why Finance Should Not Flout BasicPrinciples of Statistics”In Forthcoming, Significance (Royal Statistical Society), 2021
  • [4]Humphrey K. K. Tung and Michael C. S. Wong“Financial Risk Forecasting with Non-Stationarity”In Financial Risk ForecastingPalgrave Macmillan UK, 2011
  • [5]Thomas Guhr“Non-stationarity in Financial Markets: Dynamics of MarketStates Versus Generic Features”In Acta Physica Polonica B 46, 2015, pp. 1625
  • [6]Mazin AM Al Janabi“Optimization algorithms and investment portfolio analyticswith machine learning techniques under time-varying liquidity constraints”In Journal of Modelling in ManagementEmerald Publishing Limited, 2021
  • [7]Marcos López De Prado“The 10 reasons most machine learning funds fail”In The Journal of Portfolio Management 44.6Institutional Investor Journals Umbrella, 2018, pp. 120–133
  • [8]Adrian Millea“Deep reinforcement learning for trading—A critical survey”In Data 6.11Multidisciplinary Digital Publishing Institute, 2021, pp. 119
  • [9]Leif Andersen“Regulation, Capital, and Margining: Quant Angle”Bank of America Merrill Lynch, 2014
  • [10]Christos Makridis and Alberto G Rossi“Rise of the’Quants’ in Financial Services: Regulation andCrowding Out of Routine Jobs”In Available at SSRN 3218031, 2018
  • [11]Sean D Holcomb et al.“Overview on deepmind and its alphago zero ai”In Proceedings of the 2018 international conference on bigdata and education, 2018, pp. 67–71
  • [12]Johannes Heinrich and David Silver“Deep reinforcement learning from self-play inimperfect-information games”In arXiv preprint arXiv:1603.01121, 2016
  • [13]Philip Paquette et al.“No-press diplomacy: Modeling multi-agent gameplay”In Advances in Neural Information Processing Systems 32, 2019
  • [14]Noam Brown, Anton Bakhtin, Adam Lerer and Qucheng Gong“Combining deep reinforcement learning and search forimperfect-information games”In Advances in Neural Information Processing Systems 33, 2020, pp. 17057–17069
  • [15]Jian Yao et al.“Solving imperfect information poker games using Monte Carlosearch and POMDP models”In 2020 IEEE 9th Data Driven Control and Learning SystemsConference (DDCLS), 2020, pp. 1060–1065IEEE
  • [16]Anton Bakhtin, David Wu, Adam Lerer and Noam Brown“No-press diplomacy from scratch”In Advances in Neural Information Processing Systems 34, 2021, pp. 18063–18074
  • [17]B Ravi Kiran et al.“Deep reinforcement learning for autonomous driving: A survey”In IEEE Transactions on Intelligent TransportationSystems 23.6IEEE, 2021, pp. 4909–4926
  • [18]Jithin Jagannath, Anu Jagannath, Sean Furman and Tyler Gwin“Deep learning and reinforcement learning for autonomousunmanned aerial systems: Roadmap for theory to deployment”In Deep Learning for Unmanned SystemsSpringer, 2021, pp. 25–82
  • [19]Makram Chahine et al.“Robust flight navigation out of distribution with liquidneural networks”In Science Robotics 8.77American Association for the Advancement of Science, 2023, pp. eadc8892
  • [20]Zhiwei Qin et al.“Ride-hailing order dispatching at didi via reinforcementlearning”In INFORMS Journal on Applied Analytics 50.5INFORMS, 2020, pp. 272–286
  • [21]Terrence J Sejnowski“Large language models and the reverse turing test”In Neural computation 35.3MIT Press, 2023, pp. 309–342
  • [22]Adaku Uchendu et al.“TURINGBENCH: A benchmark environment for Turing test in theage of neural text generation”In arXiv preprint arXiv:2109.13296, 2021
  • [23]Baolin Peng et al.“Instruction tuning with gpt-4”In arXiv preprint arXiv:2304.03277, 2023
  • [24]Shijie Wu et al.“Bloomberggpt: A large language model for finance”In arXiv preprint arXiv:2303.17564, 2023
  • [25]Xiao-Yang Liu et al.“FinRL: A deep reinforcement learning library for automatedstock trading in quantitative finance”In arXiv preprint arXiv:2011.09607, 2020
  • [26]Xiao-Yang Liu et al.“FinRL-Meta: Market environments and benchmarks fordata-driven financial reinforcement learning”In Advances in Neural Information Processing Systems 35, 2022, pp. 1835–1849
  • [27]Zechu Li et al.“FinRL-Podracer: High performance and scalable deepreinforcement learning for quantitative finance”In Proceedings of the Second ACM International Conference onAI in Finance, 2021, pp. 1–9
  • [28]Zitao Song, Xuyang Jin and Chenliang Li“Safe-FinRL: A Low Bias and Variance Deep ReinforcementLearning Implementation for High-Freq Stock Trading”In arXiv preprint arXiv:2206.05910, 2022
  • [29]Antonin Raffin et al.“Stable-Baselines3: Reliable Reinforcement LearningImplementations”In Journal of Machine Learning Research 22.268, 2021, pp. 1–8URL: http://jmlr.org/papers/v22/20-1364.html
  • [30]Shengyi Huang et al.“Cleanrl: High-quality single-file implementations of deepreinforcement learning algorithms”In Journal of Machine Learning Research 23.274, 2022, pp. 1–18
  • [31]Mark Towers et al.“Gymnasium”Zenodo, 2023DOI: 10.5281/zenodo.8127026
  • [32]Jiayi Weng et al.“Envpool: A highly parallel reinforcement learning environmentexecution engine”In Advances in Neural Information Processing Systems 35, 2022, pp. 22409–22421
  • [33]Ngoc Duy Nguyen et al.“Towards designing a generic and comprehensive deepreinforcement learning framework”In Applied Intelligence 53.3Springer, 2023, pp. 2967–2988
  • [34]Matei Zaharia et al.“The Shift from Models to Compound AI Systems”, https://bair.berkeley.edu/blog/2024/02/18/compound-ai-systems/, 2024
  • [35]Theis Ingerslev Jensen, Bryan Kelly and Lasse Heje Pedersen“Is there a replication crisis in finance?”In The Journal of Finance 78.5Wiley Online Library, 2023, pp. 2465–2518
  • [36]Theis Ingerslev Jensen, Bryan T Kelly and Lasse Heje Pedersen“Is there a replication crisis in finance?”, 2021
  • [37]Kewei Hou, Chen Xue and Lu Zhang“Replicating anomalies”In The Review of financial studies 33.5Oxford University Press, 2020, pp. 2019–2133
  • [38]Florian Echtler and Maximilian Häußler“Open source, open science, and the replication crisis in HCI”In Extended abstracts of the 2018 CHI conference on humanfactors in computing systems, 2018, pp. 1–8
  • [39]Matthew Hutson“Artificial intelligence faces reproducibility crisis”American Association for the Advancement of Science, 2018
  • [40]Elizabeth Gibney“Is AI fuelling a reproducibility crisis in science”In Nature 608, 2022, pp. 250–251
  • [41]Yann-Gaël Guéhéneuc and Foutse Khomh“Empirical software engineering”In Handbook of Software EngineeringSpringer, 2019, pp. 285–320
  • [42]Jessica Hullman et al.“The worst of both worlds: A comparative analysis of errors inlearning from data in psychology and machine learning”In Proceedings of the 2022 AAAI/ACM Conference on AI,Ethics, and Society, 2022, pp. 335–348
  • [43]Christopher Tong“Statistical inference enables bad science; statisticalthinking enables good science”In The American Statistician 73.sup1Taylor & Francis, 2019, pp. 246–261
  • [44]Roger Mead, Steven G Gilmour and Andrew Mead“Statistical principles for the design of experiments:applications to real experiments”Cambridge University Press, 2012
  • [45]Michael Parkinson and Carlos Oscar Sánchez Sorzano“Why Do We Need a Statistical Experiment Design?”In Experimental Design and Reproducibility in PreclinicalAnimal StudiesSpringer, 2021, pp. 129–146
  • [46]Hartmut Schiefer, Felix Schiefer, Hartmut Schiefer and Felix Schiefer“Statistical Design of Experiments (DoE)”In Statistics for Engineers: An Introduction withExamples from PracticeSpringer, 2021, pp. 1–20
  • [47]Xiaoning Kang and Xinwei Deng“Design and analysis of computer experiments with quantitativeand qualitative inputs: A selective review”In Wiley Interdisciplinary Reviews: Data Mining andKnowledge Discovery 10.3Wiley Online Library, 2020, pp. e1358
  • [48]Tyler Munger and Subhas Desa“A statistical design of experiments approach to machinelearning model selection in engineering applications”In Journal of Computing and Information Science inEngineering 21.1American Society of Mechanical Engineers, 2021, pp. 011008
  • [49]Frank J Fabozzi, Francis Gupta and Harry M Markowitz“The legacy of modern portfolio theory”In The journal of investing 11.3Institutional Investor Journals Umbrella, 2002, pp. 7–22
  • [50]Warren B Powell“Reinforcement Learning and Stochastic Optimization: A unifiedframework for sequential decisions”John Wiley & Sons, 2022
  • [51]Zvi Bodie, Alex Kane and Alan Marcus“Investments: Global Edition”McGraw Hill, 2020
  • [52]Harry M Markowitz and Harry M Markowitz“Portfolio selection: efficient diversification ofinvestments”J. Wiley, 1967
  • [53]Franco Modigliani and Merton H Miller“The cost of capital, corporation finance and the theory ofinvestment”In The American economic review 48.3JSTOR, 1958, pp. 261–297
  • [54]William F Sharpe“Capital asset prices: A theory of market equilibrium underconditions of risk”In The journal of finance 19.3Wiley Online Library, 1964, pp. 425–442
  • [55]Harry M Markowitz“Foundations of portfolio theory”In The journal of finance 46.2JSTOR, 1991, pp. 469–477
  • [56]T-J Chang, Nigel Meade, John E Beasley and Yazid M Sharaiha“Heuristics for cardinality constrained portfoliooptimisation”In Computers & Operations Research 27.13Elsevier, 2000, pp. 1271–1302
  • [57]Francesco Cesarone, Andrea Scozzari and Fabio Tardella“A new method for mean-variance portfolio optimization withcardinality constraints”In Annals of Operations Research 205Springer, 2013, pp. 213–234
  • [58]Dimitris Kouzoupis, Gianluca Frison, Andrea Zanelli and Moritz Diehl“Recent advances in quadratic programming algorithms fornonlinear model predictive control”In Vietnam Journal of Mathematics 46.4Springer, 2018, pp. 863–882
  • [59]Markus Hirschberger, Yue Qi and Ralph E Steuer“Large-scale MV efficient frontier computation via a procedureof parametric quadratic programming”In European Journal of Operational Research 204.3Elsevier, 2010, pp. 581–588
  • [60]Stephen Boyd et al.“Multi-period trading via convex optimization”In Foundations and Trends® inOptimization 3.1Now Publishers, Inc., 2017, pp. 1–76
  • [61]Olga Bourachnikova and Thierry Burger-Helmchen“Investor’s behaviour and the relevance of asymmetric riskmeasures”In Banks & bank systems, 2012, pp. 87–94
  • [62]Borjana Racheva-Iotova and Stoyan Stoyanov“Post-modern approaches for portfolio optimization”In Handbook on information technology in financeSpringer, 2008, pp. 613–634
  • [63]Edmond Lezmi, Thierry Roncalli and Jiali Xu“Multi-Period Portfolio Optimization”In Available at SSRN, 2022
  • [64]Petter N Kolm, Reha Tütüncü and Frank J Fabozzi“60 Years of portfolio optimization: Practical challenges andcurrent trends”In European Journal of Operational Research 234.2Elsevier, 2014, pp. 356–371
  • [65]Jean-Philippe Bouchaud“Economics needs a scientific revolution”In Nature 455.7217Nature Publishing Group UK London, 2008, pp. 1181–1181
  • [66]Cristina Geambasu, Robert Sova, Iulia Jianu and Liviu Geambasu“Risk measurement in post-modern portfolio theory: differencesfrom modern portfolio theory”In Economic Computation & Economic Cybernetics Studies& Research 47.1, 2013, pp. 113–132
  • [67]Lionel Martellini“Toward the design of better equity benchmarks: Rehabilitatingthe tangency portfolio from modern portfolio theory”In The Journal of Portfolio Management 34.4Institutional Investor Journals Umbrella, 2008, pp. 34–41
  • [68]Alexander Kempf, Olaf Korn and Sven Saßning“Portfolio optimization using forward-looking information”In Review of Finance 19.1Oxford University Press, 2015, pp. 467–490
  • [69]Daniele D’Alvia“Uncertainty: The Necessary Unknowable Road to Speculation”In The Speculator of Financial Markets: How FinancialInnovation and Supervision Made the Modern WorldCham: Springer International Publishing, 2023, pp. 119–169
  • [70]Erdinç Akyıldırım and Halil Mete Soner“A brief history of mathematics in finance”In Borsa Istanbul Review 14.1Elsevier, 2014, pp. 57–63
  • [71]Samuel Schmidgall et al.“Brain-inspired learning in artificial neural networks: areview”In arXiv preprint arXiv:2305.11252, 2023
  • [72]Fahad Sarfraz, Elahe Arani and Bahram Zonooz“A Study of Biologically Plausible Neural Network: The Roleand Interactions of Brain-Inspired Mechanisms in Continual Learning”In arXiv preprint arXiv:2304.06738, 2023
  • [73]Karl Johan Åström“Optimal control of Markov processes with incomplete stateinformation”In Journal of mathematical analysis and applications 10.1Academic Press, 1965, pp. 174–205
  • [74]Michael L Littman“A tutorial on partially observable Markov decision processes”In Journal of Mathematical Psychology 53.3Elsevier, 2009, pp. 119–125
  • [75]Richard Bellman“A Markovian decision process”In Journal of mathematics and mechanicsJSTOR, 1957, pp. 679–684
  • [76]Huaizhi Wang et al.“A review of deep learning for renewable energy forecasting”In Energy Conversion and Management 198Elsevier, 2019, pp. 111799
  • [77]Alejandro J Real, Fernando Dorado and Jaime Durán“Energy demand forecasting using deep learning: Applicationsfor the French grid”In Energies 13.9MDPI, 2020, pp. 2242
  • [78]Ioannis Boukas et al.“A deep reinforcement learning framework for continuousintraday market bidding”In Machine Learning 110Springer, 2021, pp. 2335–2387
  • [79]Yuanhang Zheng, Zeshui Xu and Anran Xiao“Deep learning in economics: a systematic and critical review”In Artificial Intelligence ReviewSpringer, 2023, pp. 1–43
  • [80]Jay Cao, Jacky Chen, John Hull and Zissis Poulos“Deep learning for exotic option valuation”In The Journal of Financial Data ScienceInstitutional Investor Journals Umbrella, 2021
  • [81]Anthony L Caterini and Dong Eui Chang“Generic Representation of Neural Networks”In Deep Neural Networks in a Mathematical FrameworkSpringer, 2018, pp. 23–34
  • [82]Salvatore Cuomo et al.“Scientific Machine Learning through Physics-Informed NeuralNetworks: Where we are and What’s next”In arXiv preprint arXiv:2201.05624, 2022
  • [83]James Owen Weatherall“The physics of wall street: a brief history of predicting theunpredictable”Houghton Mifflin Harcourt, 2013
  • [84]Alex Tuzhilin et al.“Large-scale recommender systems and the netflix prizecompetition”In KDD Proceedings, 2008, pp. 1–34
  • [85]Robert M Bell and Yehuda Koren“Lessons from the Netflix prize challenge”In Acm Sigkdd Explorations Newsletter 9.2ACM New York, NY, USA, 2007, pp. 75–79
  • [86]Harald Steck et al.“Deep learning for recommender systems: A Netflix case study”In AI Magazine 42.3, 2021, pp. 7–18
  • [87]Hong-Jian Xue et al.“Deep matrix factorization models for recommender systems.”In IJCAI 17, 2017, pp. 3203–3209Melbourne, Australia
  • [88]Lucas de Azevedo Takara, André Alves Portela Santos, Viviana Cocco Mariani and Leandro Santos Coelho“Deep Reinforcement Learning Applied to a Sparse-RewardTrading Environment with Intraday Data”In Available at SSRN 4411793
  • [89]Chen Gao et al.“A survey of graph neural networks for recommender systems:Challenges, methods, and directions”In ACM Transactions on Recommender Systems 1.1ACM New York, NY, USA, 2023, pp. 1–51
  • [90]Zheqing Zhu et al.“Pearl: A Production-ready Reinforcement Learning Agent”In arXiv preprint arXiv:2312.03814, 2023
  • [91]Andrew Bennett, Nathan Kallus, Lihong Li and Ali Mousavi“Off-policy evaluation in infinite-horizon reinforcementlearning with latent confounders”In International Conference on Artificial Intelligence andStatistics, 2021, pp. 1999–2007PMLR
  • [92]Rachana Mehta and Keyur Rana“A review on matrix factorization techniques in recommendersystems”In 2017 2nd International Conference on CommunicationSystems, Computing and IT Applications (CSCITA), 2017, pp. 269–274IEEE
  • [93]Shuai Zhang, Lina Yao, Aixin Sun and Yi Tay“Deep learning based recommender system: A survey and newperspectives”In ACM computing surveys (CSUR) 52.1ACM New York, NY, USA, 2019, pp. 1–38
  • [94]M Mehdi Afsar, Trafford Crump and Behrouz Far“Reinforcement learning based recommender systems: A survey”In ACM Computing Surveys 55.7ACM New York, NY, 2022, pp. 1–38
  • [95]Yuanguo Lin et al.“A survey on reinforcement learning for recommender systems”In IEEE Transactions on Neural Networks and LearningSystemsIEEE, 2023
  • [96]Saud Almahdi and Steve Y Yang“An adaptive portfolio trading system: A risk-return portfoliooptimization using recurrent reinforcement learning with expected maximumdrawdown”In Expert Systems with Applications 87Elsevier, 2017, pp. 267–279
  • [97]Yoshiharu Sato“Model-free reinforcement learning for financial portfolios: abrief survey”In arXiv preprint arXiv:1904.04973, 2019
  • [98]Amine Mohamed Aboussalah and Chi-Guhn Lee“Continuous control with stacked deep dynamic recurrentreinforcement learning for portfolio optimization”In Expert Systems with Applications 140Elsevier, 2020, pp. 112891
  • [99]Hui Niu, Siyuan Li and Jian Li“MetaTrader: An reinforcement learning approach integratingdiverse policies for portfolio optimization”In Proceedings of the 31st ACM International Conference onInformation & Knowledge Management, 2022, pp. 1573–1583
  • [100]Zihao Zhang, Stefan Zohren and Roberts Stephen“Deep reinforcement learning for trading”In The Journal of Financial Data ScienceInstitutional Investor Journals Umbrella, 2020
  • [101]Leonardo Kanashiro Felizardo, Francisco Caio Lima Paiva, Anna Helena Reali Costa and Emilio Del-Moral-Hernandez“Reinforcement Learning Applied to Trading Systems: A Survey”In arXiv preprint arXiv:2212.06064, 2022
  • [102]Gordon Ritter“Machine learning for trading”In Available at SSRN 3015609, 2017
  • [103]Bruno Biais, Larry Glosten and Chester Spatt“Market microstructure: A survey of microfoundations,empirical results, and policy implications”In Journal of Financial Markets 8.2Elsevier, 2005, pp. 217–264
  • [104]David Easley, Marcos Prado, Maureen O’Hara and Zhibai Zhang“Microstructure in the machine age”In The Review of Financial Studies 34.7Oxford University Press, 2021, pp. 3316–3363
  • [105]Bastien Baldacci“Quantitative finance at the microstructure scale: algorithmictrading and regulation”, 2021
  • [106]Tobias Galla and J Doyne Farmer“Complex dynamics in learning complicated games”In Proceedings of the National Academy of Sciences 110.4National Acad Sciences, 2013, pp. 1232–1236
  • [107]Reidar B Bratvold and Frank Koch“Game Theory in the Oil and Gas Industry”In The Way Ahead 7.01SPE, 2011, pp. 18–20
  • [108]Vivek K Pandey and Chen Y Wu“Investors May Take Heart: A Game Theoretic View of HighFrequency Trading”In Journal of Financial Planning 28.5, 2015, pp. 53–57
  • [109]Prajit T Rajendran, Huascar Espinoza, Agnes Delaborde and Chokri Mraidha“Human-in-the-Loop Learning Methods Toward Safe DL-BasedAutonomous Systems: A Review”In Computer Safety, Reliability, and Security. SAFECOMP 2021Workshops: DECSoS, MAPSOD, DepDevOps, USDAI, and WAISE, York, UK, September7, 2021, Proceedings 40, 2021, pp. 251–264Springer
  • [110]Matthew E Taylor“Reinforcement Learning Requires Human-in-the-Loop Framing andApproaches.”In HHAI, 2023, pp. 351–360
  • [111]Alvaro HC Correia and Freddy Lecue“Human-in-the-loop feature selection”In Proceedings of the AAAI Conference on ArtificialIntelligence 33.01, 2019, pp. 2438–2445
  • [112]Jingda Wu et al.“Human-in-the-loop deep reinforcement learning withapplication to autonomous driving”In arXiv preprint arXiv:2104.07246, 2021
  • [113]Sebastian Höfer et al.“Perspectives on sim2real transfer for robotics: A summary ofthe r: Ss 2020 workshop”In arXiv preprint arXiv:2012.03806, 2020
  • [114]Saminda Wishwajith Abeyruwan et al.“i-sim2real: Reinforcement learning of robotic policies intight human-robot interaction loops”In Conference on Robot Learning, 2023, pp. 212–224PMLR
  • [115]Xiaoyu Chen et al.“Understanding domain randomization for sim-to-real transfer”In arXiv preprint arXiv:2110.03239, 2021
  • [116]Wenshuai Zhao, Jorge Peña Queralta and Tomi Westerlund“Sim-to-real transfer in deep reinforcement learning forrobotics: a survey”In 2020 IEEE symposium series on computational intelligence(SSCI), 2020, pp. 737–744IEEE
  • [117]Andrei A Rusu et al.“Sim-to-real robot learning from pixels with progressive nets”In Conference on robot learning, 2017, pp. 262–270PMLR
  • [118]Jiachen Hu, Han Zhong, Chi Jin and Liwei Wang“Provable sim-to-real transfer in continuous domain withpartial observations”In arXiv preprint arXiv:2210.15598, 2022
  • [119]Marc G Bellemare, Will Dabney and Mark Rowland“Distributional reinforcement learning”MIT Press, 2023
  • [120]Moloud Abdar et al.“A review of uncertainty quantification in deep learning:Techniques, applications and challenges”In Information fusion 76Elsevier, 2021, pp. 243–297
  • [121]Owen Lockwood and Mei Si“A review of uncertainty for deep reinforcement learning”In Proceedings of the AAAI Conference on ArtificialIntelligence and Interactive Digital Entertainment 18.1, 2022, pp. 155–162
  • [122]Yi Zhu, Jing Dong and Henry Lam“Uncertainty Quantification and Exploration for ReinforcementLearning”In Operations ResearchINFORMS, 2023
  • [123]Gianluca Bianchin, Yin-Chen Liu and Fabio Pasqualetti“Secure navigation of robots in adversarial environments”In IEEE Control Systems Letters 4.1IEEE, 2019, pp. 1–6
  • [124]Lifeng Zhou and Pratap Tokekar“Multi-robot coordination and planning in uncertain andadversarial environments”In Current Robotics Reports 2Springer, 2021, pp. 147–157
  • [125]Josh Tobin et al.“Domain randomization for transferring deep neural networksfrom simulation to the real world”In 2017 IEEE/RSJ international conference on intelligentrobots and systems (IROS), 2017, pp. 23–30IEEE
  • [126]Georgios D Kontes et al.“High-speed collision avoidance using deep reinforcementlearning and domain randomization for autonomous vehicles”In 2020 IEEE 23rd international conference on IntelligentTransportation Systems (ITSC), 2020, pp. 1–8IEEE
  • [127]Quan Vuong et al.“How to pick the domain randomization parameters forsim-to-real transfer of reinforcement learning policies?”In arXiv preprint arXiv:1903.11774, 2019
  • [128]Kimin Lee, Kibok Lee, Jinwoo Shin and Honglak Lee“Network randomization: A simple technique for generalizationin deep reinforcement learning”In arXiv preprint arXiv:1910.05396, 2019
  • [129]Dylan J Foster, Akshay Krishnamurthy, David Simchi-Levi and Yunzong Xu“Offline reinforcement learning: Fundamental barriers forvalue function approximation”In arXiv preprint arXiv:2111.10919, 2021
  • [130]Tengyang Xie et al.“Policy finetuning: Bridging sample-efficient offline andonline reinforcement learning”In Advances in neural information processing systems 34, 2021, pp. 27395–27407
  • [131]Tengyang Xie et al.“The role of coverage in online reinforcement learning”In arXiv preprint arXiv:2210.04157, 2022
  • [132]Lindsay Wells and Tomasz Bednarz“Explainable ai and reinforcement learning—a systematicreview of current approaches and trends”In Frontiers in artificial intelligence 4Frontiers Media SA, 2021, pp. 550030
  • [133]Prashan Madumal, Tim Miller, Liz Sonenberg and Frank Vetere“Explainable reinforcement learning through a causal lens”In Proceedings of the AAAI conference on artificialintelligence 34.03, 2020, pp. 2493–2500
  • [134]Jasmina Gajcin and Ivana Dusparic“Redefining Counterfactual Explanations for ReinforcementLearning: Overview, Challenges and Opportunities”In ACM Computing SurveysACM New York, NY, 2024
  • [135]Petar Kormushev, Sylvain Calinon and Darwin G Caldwell“Reinforcement learning in robotics: Applications andreal-world challenges”In Robotics 2.3MDPI, 2013, pp. 122–148
  • [136]Julian Ibarz et al.“How to train your robot with deep reinforcement learning:lessons we have learned”In The International Journal of Robotics Research 40.4-5SAGE Publications Sage UK: London, England, 2021, pp. 698–721
  • [137]Gabriel Dulac-Arnold et al.“Challenges of real-world reinforcement learning: definitions,benchmarks and analysis”In Machine Learning 110.9Springer, 2021, pp. 2419–2468
  • [138]Henry Zhu et al.“The ingredients of real-world robotic reinforcement learning”In arXiv preprint arXiv:2004.12570, 2020
  • [139]Avi Singh et al.“End-to-end robotic reinforcement learning without rewardengineering”In arXiv preprint arXiv:1904.07854, 2019
  • [140]Serkan Cabi et al.“Scaling data-driven robotics with reward sketching and batchreinforcement learning”In arXiv preprint arXiv:1909.12200, 2019
  • [141]Abhishek Gupta et al.“Unpacking reward shaping: Understanding the benefits ofreward engineering on sample complexity”In Advances in Neural Information Processing Systems 35, 2022, pp. 15281–15295
  • [142]Dylan Hadfield-Menell et al.“Inverse reward design”In Advances in neural information processing systems 30, 2017
  • [143]Ryan Julian et al.“Never stop learning: The effectiveness of fine-tuning inrobotic reinforcement learning”In arXiv preprint arXiv:2004.10190, 2020
  • [144]Lukas Brunke et al.“Safe learning in robotics: From learning-based control tosafe reinforcement learning”In Annual Review of Control, Robotics, and AutonomousSystems 5Annual Reviews, 2022, pp. 411–444
  • [145]Bryan Lim, Stefan Zohren and Stephen Roberts“Enhancing time-series momentum strategies using deep neuralnetworks”In The Journal of Financial Data ScienceInstitutional Investor Journals Umbrella, 2019
  • [146]John Moody, Lizhong Wu, Yuansong Liao and Matthew Saffell“Performance functions and reinforcement learning for tradingsystems and portfolios”In Journal of forecasting 17.5-6Wiley Online Library, 1998, pp. 441–470
  • [147]Zhiyu Zhang, David Bombara and Heng Yang“Discounted Adaptive Online Prediction”In arXiv preprint arXiv:2402.02720, 2024
  • [148]Zhicheng Wang et al.“DeepTrader: a deep reinforcement learning approach forrisk-return balanced portfolio management with market conditions Embedding”In Proceedings of the AAAI conference on artificialintelligence 35.1, 2021, pp. 643–650
  • [149]Zhengyao Jiang, Dixing Xu and Jinjun Liang“A deep reinforcement learning framework for the financialportfolio management problem”In arXiv preprint arXiv:1706.10059, 2017
  • [150]Pengqian Yu et al.“Model-based deep reinforcement learning for dynamic portfoliooptimization”In arXiv preprint arXiv:1901.08740, 2019
  • [151]Srijan Sood, Kassiani Papasotiriou, Marius Vaiciulis and Tucker Balch“Deep Reinforcement Learning for Optimal Portfolio Allocation:A Comparative Study with Mean-Variance Optimization”In FinPlan 2023, 2023, pp. 21
Advancing Investment Frontiers: Industry-grade Deep Reinforcement Learning for Portfolio Optimization (2024)
Top Articles
Latest Posts
Article information

Author: Carmelo Roob

Last Updated:

Views: 5493

Rating: 4.4 / 5 (45 voted)

Reviews: 92% of readers found this page helpful

Author information

Name: Carmelo Roob

Birthday: 1995-01-09

Address: Apt. 915 481 Sipes Cliff, New Gonzalobury, CO 80176

Phone: +6773780339780

Job: Sales Executive

Hobby: Gaming, Jogging, Rugby, Video gaming, Handball, Ice skating, Web surfing

Introduction: My name is Carmelo Roob, I am a modern, handsome, delightful, comfortable, attractive, vast, good person who loves writing and wants to share my knowledge and understanding with you.