Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Artif. Intell., 30 January 2026

Sec. AI in Finance

Volume 8 - 2025 | https://doi.org/10.3389/frai.2025.1709493

Portfolio management based on value distribution reinforcement learning algorithm

Yan Yang
Yan Yang1*Tian WangTian Wang1Yiding FuYiding Fu1Jingna HuangJingna Huang2Dong ZhouDong Zhou3
  • 1Strategic Development Department Southern Power Grid Capital Holding Co., Ltd., Guangzhou, China
  • 2Southern Power Grid Financial Leasing Co., Ltd., Guangzhou, China
  • 3Southern Power Grid Private Fund Management Co., Ltd., Southern Power Grid Jianxin Fund Management, Guangzhou, China

Introduction: In the face of high uncertainty and complexity in financial markets, achieving portfolio return maximization while effectively controlling risk remains a critical challenge.

Methods: We propose a novel portfolio management framework based on the value distribution maximum entropy actor-critic (VD-MEAC) reinforcement learning algorithm. We establish a framework where the agent’s actions represent portfolio weight adjustments and stock factors serve as state observations. For risk management, the critic network learns the complete distribution of future returns. For return enhancement, we incorporate entropy regularization.

Results: We conduct extensive experiments using real market data from the Chinese stock market. Results demonstrate that our VD-MEAC strategy achieves an average return of 2.490 and an average Sharpe ratio of 2.978, significantly outperforming benchmark strategies.

Discussion: These results validate the effectiveness of our approach in practical portfolio management scenarios.

1 Introduction

Portfolio management remains one of the most challenging problems in financial mathematics and quantitative investment, requiring sophisticated approaches to balance return maximization against risk minimization in highly complex and non-stationary market environments (Rezaei and Nezamabadi-Pour, 2025; Sattar et al., 2025; Xu, 2025). Traditional portfolio optimization methods, from Markowitz’s mean–variance framework to various factor models, often rely on restrictive assumptions about return distributions and market behavior that may not hold in practice (Li and Hai, 2024). With the advancement of artificial intelligence and the increasing availability of high-dimensional financial data, reinforcement learning (RL) has emerged as a promising approach to portfolio management, enabling the development of adaptive investment strategies through interaction with financial markets (Jiang et al., 2024).

In practice, prior studies have applied RL to portfolio management from different perspectives. For instance, Day et al. (2024) employed policy gradient algorithms to build trading frameworks, while Fu and Huang (2025) used Q-learning to design an intelligent portfolio management system. However, these works relied on shallow neural networks, which are insufficient to handle the increasing complexity of financial markets. With the development of reinforcement learning theory, the Actor-Critic (AC) framework, which combines the benefits of both value-based and policy-based methods, has been introduced into quantitative investment. Specifically, Kitchat et al. (2024) applied the deterministic policy gradient (DPG) method to allocate a set of cryptocurrency weights and proposed a model-free convolutional network for feature extraction. Building on DPG, Cui et al. (2024) designed a state-augmentation approach to address data heterogeneity. In addition, Cheng and Sun (2024) applied the deep deterministic policy gradient (DDPG) algorithm within the AC framework to portfolio problems, while Belyakov and Sizykh (2024) introduced a proximal optimization method under the AC framework to handle portfolio optimization with transaction costs. Furthermore, Pippas et al. (2025) integrated fuzzy representations with the AC framework and proposed an adaptive fuzzy reinforcement learning approach.

Recent applications of deep reinforcement learning in portfolio management have demonstrated promising results but continue to face critical challenges (Junfeng et al., 2024). First, conventional RL algorithms typically optimize for expected returns using point estimates, which fail to capture the full uncertainty inherent in financial returns and can lead to risk-seeking behavior unsuitable for investment applications (Betancourt and Chen, 2021; Wu et al., 2021; Jang and Seong, 2023). Second, most existing approaches suffer from overestimation bias in value functions, potentially resulting in overly aggressive investment strategies and substantial drawdowns during market downturns (Aminifar et al., 2022; Koratamaddi et al., 2021). Third, the exploration-exploitation tradeoff in financial markets presents a unique challenge, as insufficient exploration may lead to strategies that perform well historically but fail to adapt to changing market conditions (Teoh et al., 2021; Pallathadka et al., 2023).

To address these limitations, we propose a Value Distribution Maximum Entropy Actor-Critic (VD-MEAC) framework that fundamentally reimagines the application of reinforcement learning to portfolio management. Our framework makes three key innovations: (1) Instead of modeling expected returns, our Critic network learns the entire distribution of future returns, providing a more comprehensive risk assessment; (2) We implement a novel mechanism to filter out overconfident decision information in the value distribution, explicitly reducing overestimation risk; and (3) We incorporate maximum entropy reinforcement learning principles to encourage strategy diversification and robust exploration of the investment action space.

The remainder of this paper is organized as follows: Section II formulates the portfolio management problem within a reinforcement learning framework. Section III introduces our VD-MEAC algorithm, detailing its theoretical foundations and implementation. Section IV presents experimental results on real market data. Section V discusses the implications of our findings and concludes the paper.

2 Reinforcement learning framework for portfolio optimization

2.1 Portfolio problem description

Portfolio optimization involves the adjustment of asset weights by investors seeking to maximize utility at the end of an investment period (Du and Ghavidel, 2022). This problem can be expressed in the following optimization form:

max x [ u ( W ( x , p ( ε ) ) ) ]     (1)

where x represents the investor’s trading strategy (i.e., the vector of asset weights), u ( · ) is the utility function, W ( · ) denotes the terminal wealth value, ε represents random factors, and p ( ε ) denotes asset prices.

To rigorously describe the problem, we make the following assumptions about investors and the financial environment (Jin et al., 2024; de López Prado et al., 2025):

• Investors are risk-neutral, meaning the utility function is linear.

• The asset pool consists of a fixed set of N risky assets and one risk-free asset, with no addition of new risky assets during the investment period.

• No minimum trading unit exists, meaning assets can be infinitely divisible.

• Trading prices are closing prices for each period, without consideration of bid-ask spreads.

• Trading costs can be represented as proportional costs.

For the optimal decision problem in Equation 1, we can employ a reinforcement learning framework for the solution. Reinforcement learning is built upon a Markov Decision Process (S, A, R, ρ, γ), where S represents the state space, A is the action space, R denotes the reward function, ρ is the state transition matrix (dependent on the specific policy π and the environment), and γ is the reward discount factor.

As shown in Figure 1, investors observe the state information from the financial market, take actions to adjust weights, and the financial environment provides rewards in the form of portfolio gains or losses. The ultimate goal of reinforcement learning is to train an investor that acts based on long-term benefits rather than myopic behavior.

Figure 1
Flowchart illustrating interactions between a financial market and an investor. Arrows connect the market and investor in a cycle. Steps: 1. Observe State S_t, 2. Trading Action A_t, 3. Change in Wealth R_t.

Figure 1. Reinforcement learning portfolio framework.

2.2 Reinforcement learning framework design

State space design. Investors need to observe the state information of the financial market to make trading decisions. In this study, we choose to describe the financial market using factor information (Dong et al., 2024). The state S t can be represented as:

S t = [ s 01 t s 02 t s 0 n t s 11 t s 12 t s 1 n t s N 1 t s N 2 t s Nn t ]     (2)

where S t is a matrix composed of asset factor information, and s ij t represents the value of factor j for asset i at time t .

We use the Light Gradient Boosting Machine (LightGBM) (Gong et al., 2024) method to select important factors from the factor library. Figure 2 shows the factor importance. Investors select the top 15 important factors as observations of the financial environment before each trade, i.e., n = 15. The specific state information is shown in Table 1. Some factors in Table 1, such as the 20-day turnover rate, already incorporate historical information, so investors only observe the current period’s factor information rather than using a three-dimensional tensor.

Figure 2
Bar chart displaying feature importance for various financial metrics. PEG score is the most important, followed by net profit to total operate revenue TTM, Variance20, and others. Liquidity has the least importance.

Figure 2. Factor importance ranking generated by LightGBM.

Table 1
www.frontiersin.org

Table 1. Specific state information.

Action space design. Investors adjust asset weights at the beginning of each period. The action At at time t is defined as the target portfolio weight vector xt for the end of the period, after rebalancing:

A t = x t = ( x t , 0 , x t , 1 , , x t , N ) T     (3)

where x t , i represents the target weight of the i-th asset (with i = 0 being the risk-free asset). The action space is the set of all valid weight vectors, i.e., i = 0 N x t , i = 1 , x t , i 0 (for a long-only portfolio). This constraint explicitly forbids the use of leverage or short-selling, ensuring that all performance gains are derived solely from the agent’s methodological advantages. The policy network outputs the parameters for this continuous action vector.

Reward function design. The design of the reward function is crucial. We follow the framework presented in Jiang et al. (2017) to define the reward as the one-period portfolio log return, net of transaction costs, which directly relates to maximizing terminal wealth.

Let W t be the portfolio value after rebalancing at time t, and x t = ( x t , 0 , , x t , N ) T be the corresponding weight vector. Let P t + 1 = ( P t + 1 , 0 , , P t + 1 , N ) T be the gross relative price vector from time t to t + 1 (i.e., price t + 1 / price t ). We assume P t + 1 , 0 = 1 for the risk-free asset. The portfolio value before rebalancing at t + 1 is W t + 1 = W t · ( x t T P t + 1 ) .

At this point, the agent observes state S t + 1 and takes action A t + 1 to choose a new weight vector x t + 1 . Before this rebalancing, the drifted weights (due to market movement) are x t = ( x t P t + 1 ) / ( x t T P t + 1 ) , where is element-wise multiplication.

Following Jiang et al. (2017), a proportional transaction cost C is incurred on the change in weights for risky assets. The portfolio value after rebalancing at t + 1 is: W t + 1 = W t + 1 ( 1 C · i = 1 N x t + 1 , i x t , i ) . The one-period reward R t + 1 is then defined as the log return:

R t + 1 = ln ( W t + 1 / W t ) = ln ( ( x t T P t + 1 ) · ( 1 C · i = 1 N x t + 1 , i x t , i ) )     (4)

This reward function directly optimizes the cumulative log return, while correctly accounting for the friction of transaction costs as defined in Jiang et al. (2017).

3 Methodology

Classical reinforcement learning handles uncertainty in long-term decision processes by calculating expectations, specifically:

Q ( s , a ) = E [ R ( s , a ) ] + γ E [ max a Q ( s , a ) ]     (5)

where Q ( s , a ) is used to evaluate the maximum expected return that the current state-action pair ( s , a ) can generate. In solving optimal decision problems, whether using value function-based or policy-based reinforcement learning algorithms, the accuracy of Q ( s , a ) directly affects the algorithm’s performance. However, from the definition in Equation 5, we can see that Q ( s , a ) only utilizes the expectation information from the distribution Z π ( s , a ) , and expectation values are easily influenced by extreme values. Furthermore, the maximization in Equation 5 and the bootstrapping TD (temporal difference) algorithm used in training inevitably produce overestimation (Zhao et al., 2024; Li et al., 2024), which in investment manifests as overconfidence, potentially leading to investment losses.

To address this, we define a stochastic policy π : S ( A ) as a mapping from states to a probability distribution over actions. We then model the full distribution of the random return Z π ( s , a ) , which is defined as the discounted sum of future rewards: Z π ( s , a ) k = 0 γ k R t + k + 1 S t = s , A t = a , π . The Bellman equation for this random return is:

Z π ( s , a ) = R ( s , a ) + γ Z π ( S , A )     (6)

where ( S , A ) ~ π ( · S ) . Our goal is to learn the distribution of Z π ( s , a ) , not just its expectation Q π ( s , a ) = E [ Z π ( s , a ) ] .

3.1 Distribution function parameterization

The first issue to address is how to parameterize the distribution Z π ( s , a ) . We follow the Fully Parameterized Quantile Function (FQF) approach (Yang et al., 2019) to parameterize Z π ( s , a ) . According to Yang et al. (2019), any cumulative distribution function (CDF) FZ and its inverse (quantile function) F Z 1 satisfy the following relationship for the expected value E [ Z ] = 0 1 F Z 1 . This fundamental result allows us to represent a distribution by discretizing its quantile function. Following the FQF parameterization, we represent the return distribution as:

Z θ , τ ( s , a ) i = 0 N 1 ( τ i + 1 τ i ) δ θ i ( s , a )     (7)

where δ θ i ( s , a ) is the Dirac function, and τ represents the quantiles with 0 = τ 0 < τ i 1 < τ i < τ N = 1 . Our approach employs two neural networks: (1) Quantile Proposal Network τ: Takes state-action pair (s, a) as input and outputs adaptive quantile fractions τ = τ(s, a). (2) Quantile Value Network θ: Takes (s, a, τ) as input and outputs the quantile values θ = θ(s, a, τ).

For each state-action pair ( s , a ) , the quantile proposal network outputs quantiles τ , and the quantile value network outputs quantile values θ for each set of τ inputs.

If F Z ( z ) = P ( Z < z ) is the cumulative distribution function of Z π ( s , a ) , then its inverse function is F Z 1 ( p ) inf { z : p F Z ( z ) } . According to Equation 7, we can derive the expression for the quantile values:

F Z θ , τ 1 ( ω ) = θ 0 + i = 0 N 1 ( θ i + 1 θ i ) H τ i + 1 ( ω )     (8)

where H τ i + 1 ( ω ) is the unit step function.

For the quantile proposal network, the closer the output quantiles τ are to the actual quantiles, the better. Thus, we define the loss function as:

W 1 ( Z , τ ) = i = 0 N 1 τ i τ i + 1 F Z 1 ( ω ) F Z 1 ( τ i ) a     (9)

where τ i = ( τ i + τ i + 1 ) / 2 , using the Wasserstein distance.

The gradient information can be obtained by differentiating the parametric variable integral (Yang et al., 2019):

W 1 τ i = 2 F Z 1 ( τ i ) F Z 1 ( τ i ) F Z 1 ( τ i 1 )     (10)

Equation 10 can be simplified to avoid integral calculations, reducing the difficulty of network training. For the quantile value network, combining quantile regression with the Bellman equation, we have the TD error:

δ ij t = r t + γ F Z , w 1 1 ( τ i ) F Z , w 1 1 ( τ j )     (11)

where w1 denotes the target network parameters.

The loss function is chosen as the Huber quantile regression function:

L ( s t , a t , r t , s t + 1 ) = i = 0 N 1 j = 0 N 1 ρ τ j κ ( δ ij t )     (12)

where ρ τ j κ ( δ ij t ) = τ I ( δ ij t < 0 ) L κ ( δ ij t ) κ , I ( · ) is the indicator function, L κ ( · ) is the Huber loss function, and κ is the threshold value. When δ ij κ , it is the squared error, otherwise, it is the linear error.

3.2 Value distribution reinforcement learning

After parameterizing the distribution Z π ( s , a ) , we need to consider how to utilize the distribution information. We adopt the Actor-Critic framework, where the distributional critic guides the actor. Our approach is based on the Soft Actor-Critic (SAC) framework (Haarnoja et al., 2018), which incorporates a maximum entropy objective to encourage exploration. Unlike SAC, our critic learns a quantile-parameterized return distribution rather than an expected Q-value.

High quantiles imply higher estimates of future returns for the current state-action pair ( s , a ) , which in finance can lead to risk due to overconfidence. Due to the overestimation problem inherent in network training, we need to discard information that might cause overestimation. We define the utilization of distribution information as:

Q π ( s , a ) = i = 0 ( N 1 ) β ( τ i + 1 τ i ) F Z , w 2 1 ( τ i )     (13)

where Q π ( s , a ) is the guidance information from the Critic network, transmitted to the Actor network, w2 denotes the quantile-value network parameters, β is the distribution information utilization coefficient with ( N 1 ) β + . The coefficient β ( 0 , 1 ] controls the fraction of quantile information used when aggregating the learned distribution. A smaller β filters out upper-tail quantiles to mitigate overestimation. The state value function V π ( s ) is defined as:

V π ( s ) = Q π ( s , a ) + αH ( π ( · s ) )     (14)

In terms of returns, we add entropy regularization H ( π ( · s ) ) , using the maximum entropy principle to encourage investors to explore the action space and find more profitable trading decisions. α is the regularization coefficient; a larger α indicates stronger exploration (Zhu et al., 2024). Unlike fixed game scenarios, the financial market is a complex environment with multiple suboptimal or optimal decisions. Therefore, we prefer learning a stochastic policy to adapt to the complex financial market.

The maximum entropy objective modifies the standard reinforcement learning objective to:

π = argma x π E π [ t = 0 γ t ( R ( s t , a t ) + αH ( π ( · s t ) ) ) ]     (15)

This objective encourages exploration in a principled way by maximizing both the expected return and the entropy of the policy. The entropy term H ( π ( · s ) ) is defined as:

H ( π ( · s ) ) = E a ~ π ( · s ) [ log π ( a s ) ]     (16)

By incorporating this entropy term, the agent is incentivized to maintain diverse action selection probabilities, preventing premature convergence to potentially suboptimal deterministic policies. This is particularly valuable in financial markets where:

• Multiple near-optimal strategies may exist.

• Market conditions change over time.

• Deterministic policies are more vulnerable to adversarial conditions.

• Exploration is necessary to discover new profitable opportunities.

3.3 Synergistic benefits of value distribution and maximum entropy

The true innovation of our VD-MEAC algorithm lies in the synergistic integration of value distribution learning and maximum entropy exploration. These two components complement each other in several ways:

Risk-aware exploration: The value distribution component provides rich uncertainty information that guides the entropy-based exploration toward regions with both high expected returns and manageable risk.

Robust uncertainty estimation: The maximum entropy component encourages the agent to explore diverse states, which in turn improves the quality and coverage of the learned return distributions.

Adaptive risk–return tradeoff: The combination allows for dynamic adjustment of the risk–return tradeoff based on the full distribution information rather than just point estimates.

Market regime adaptation: By maintaining policy stochasticity while capturing return distributions, the agent can quickly adapt to changing market conditions and regime shifts.

The Actor network in VD-MEAC follows a stochastic policy parameterization:

π ϕ ( a s ) = 1 2 π σ ϕ ( s ) 2 exp ( ( a μ ϕ ( s ) ) 2 2 σ ϕ ( s ) 2 )     (17)

where μ ϕ ( s ) and σ ϕ ( s ) are the mean and standard deviation of the action distribution, respectively, produced by the Actor network. This Gaussian policy allows for controlled stochasticity in portfolio weight adjustments.

3.4 Theoretical convergence properties

The theoretical convergence properties of VD-MEAC are founded on the established convergence guarantees of its core components. The distributional critic, based on FQF, inherits the convergence properties of distributional RL in the 1-Wasserstein metric, which is shown to be a contraction (Yang et al., 2019). The actor and its entropy-regularized objective are based on the Soft Actor-Critic framework, which provides its own policy improvement and convergence guarantees (Haarnoja et al., 2018).

While a unified convergence proof for the combined VD-MEAC framework is non-trivial and left for future work, the robust empirical convergence demonstrated in our experiments (Figure 3) validates the stability and effectiveness of this synergistic approach.

Figure 3
Four graphs display model performance metrics over 500,000 training steps. Graph (a) shows a decreasing learning rate, (b) shows an increasing Sharpe ratio, (c) shows a decreasing loss with fluctuations in red, and (d) shows a similar decreasing loss pattern in purple.

Figure 3. Model training results. (a) Learning rate, (b) Sharpe ratio, (c) Loss (Critic), and (d) Loss (Actor).

Algorithm 1 and Figure 4 illustrate the VD-MEAC algorithm flow. Our approach combines the strengths of distributional reinforcement learning with maximum entropy reinforcement learning to create a robust portfolio management system that effectively balances risk and return considerations.

Figure 4
Diagram of a reinforcement learning framework. The environment provides a state \(s\) and a reward \(r\) to the Critic Network, which includes a Quantile Proposal Network and a Quantile Value Network. This network outputs a quantile value \(\theta\) and an action \(a\), which are evaluated as \(Q^\pi(s,a)\). The Actor Network, linked with entropy regularization, uses a Policy Network to determine actions based on state \(s\). The loop continues with interactions between the environment, Critic Network, and Actor Network.

Figure 4. VD-MEAC algorithm flow diagram.

ALGORITHM 1 Value distribution maximum entropy actor-critic

Initialize actor network π ϕ with random parameters ϕ

Initialize quantile proposal network τ ψ with random parameters ω

Initialize quantile value network θ ω with random parameters ω

Initialize target networks: ψ ψ , ω ω

Initialize replay buffer 𝓓

For each episode:

Initialize state s 0

For each time step t:

Observe state s t

Sample action a t ~ π ϕ ( · s t )

Execute action a t , observe next state s t + 1 and reward r t

Store transition ( s t , a t , r t , a s t + 1 ) in 𝓓

If time to update:

1. Sample mini-batch of N transitions (s, a, r, s′) from 𝓓

2. Generate quantiles τ = τ ψ ( s , a )

3. Compute quantile values θ = θ ω ( s , a , τ )

4. Compute target quantiles τ = τ ψ ( s , a ) with a ~ π ϕ ( · s )

5. Compute target quantile values θ = θ ω ( s , a , τ )

6. Update quantile proposal network by minimizing W 1 ( Z , τ ) in Equation 9

7. Update quantile value network by minimizing the loss ( Equation 12 )

8. Compute filtered Q π ( s , a ) using Eq. 13

9. Update actor network by maximizing Q π ( s , a ) + αH ( π ϕ ( · s ) )

10. Update target networks:

ψ τ ψ + ( 1 τ ) ψ

ω τ ω + ( 1 τ ) ω

4 Experiments and analysis

4.1 Model training

To thoroughly evaluate our VD-MEAC algorithm, we designed a comprehensive experimental framework comparing against both traditional portfolio strategies and state-of-the-art reinforcement learning methods (Jeribi et al., 2024; Alzaman, 2025; Aritonang et al., 2025; Cui et al., 2025). The comparative methods are:

• Equal-weight (EW): A naive baseline that assigns equal weights to all assets, requiring no optimization but serving as a surprisingly effective benchmark in many portfolio studies.

• CSI 300 Index: A market capitalization-weighted index tracking the 300 largest stocks in China, representing the market benchmark.

• DDPG: A model-free, off-policy actor-critic algorithm using deep function approximators for continuous action spaces. DDPG combines the actor-critic approach with insights from DQN.

• TD3 (Twin Delayed DDPG): An improved version of DDPG that addresses function approximation errors by using twin critics and delayed policy updates, reducing overestimation bias.

• SAC (Soft Actor-Critic): A state-of-the-art off-policy algorithm that maximizes both expected return and entropy, encouraging exploration and robustness.

For each portfolio, we selected constituent stocks from the CSI 300 Index with minimal missing data. Any intermittent missing values (e.g., due to trading halts) within the selected stocks were filled using the forward-fill method, carrying over the last known value. This ensures continuity in price series while maintaining the most recent available information for suspended stocks. Before training, all 15 state factors were normalized using z-score normalization based on the mean and standard deviation of the training dataset (July 1, 2017–July 1, 2020). This ensures that all input features have a mean of approximately 0 and a standard deviation of 1, preventing features with larger scales from dominating the learning process. The stock list is presented in Table 2.

Table 2
www.frontiersin.org

Table 2. Stock list.

To ensure the robustness of our results and avoid selection bias, we conducted experiments on three different portfolios sampled from CSI 300 constituents:

Portfolio A (Original Portfolio): Nine stocks selected based on data completeness and sector diversity, plus one risk-free asset (government bonds).

Portfolio B (Financial & Consumer Sectors): Ten stocks from financial services and consumer goods sectors, representing defensive and stable growth characteristics.

Portfolio C (Technology & Healthcare Sectors): Ten stocks from technology and healthcare sectors, representing high-growth and innovative industries.

We implemented the VD-MEAC strategy using TensorFlow 2.4 with Python 3.8. The experiments were conducted on a high-performance computing workstation equipped with an Intel Xeon E5-2698 v4 CPU, an NVIDIA Tesla V100 GPU, and 128 GB of DDR4 RAM. The system ran on Ubuntu 20.04 LTS, ensuring a stable Linux-based environment for deep learning training. The training period spanned from July 1, 2017, to July 1, 2020, ensuring sufficient historical data to capture various market conditions. The testing period was from July 1, 2020, to September 1, 2021, encompassing both bull and bear market phases. The main parameter settings for the model are presented in Table 3, where we utilized the Adam optimizer with ReLU activation functions. The 5 × 105 training steps for the VD-MEAC model took approximately 8.5 h to complete. The computational complexity of the agent at each time step is dominated by the forward passes of the actor and critic networks, which is efficient for real-time decision-making.

Table 3
www.frontiersin.org

Table 3. Model main parameter settings.

Additionally, we used the Adam optimizer and ReLU activation functions. The model was trained for 5 × 10 5 steps, with the training results shown in Figure 3. Figure 3a shows the learning rate, which incorporates decay to prevent non-convergence due to excessive learning rates. Figures 3c,d display the loss values for the Actor and Critic networks, respectively, indicating that the network training has stabilized. It’s important to note that the interpretation of loss values in reinforcement learning differs from that in deep learning; stable network training does not necessarily signify that the model has learned a profitable trading strategy. However, examining Figure 3b, we observe that the Sharpe ratio per episode increases continuously as training progresses and eventually stabilizes, suggesting model convergence.

4.2 Model testing

It is critical to note that all strategies compared in this section, including our own VD-MEAC, are evaluated under the strict long-only, no-leverage constraint defined in Section II-B. The superior performance of VD-MEAC is therefore derived entirely from its methodological advantages in risk modeling and exploration, not from financial engineering or hidden leverage.

The testing period spans from July 1, 2020, to September 1, 2021, with transaction costs set at 0.25%. Since VD-MEAC learns a stochastic policy, we conducted 100 trading simulations during the test period to avoid evaluation bias from extreme performances. The trading results are presented in Figure 5. Across these 100 trading simulations, the VD-MEAC strategy achieved an average Sharpe ratio of 2.978 with a variance of 0.015, and an average wealth ratio of 2.490 with a variance of 0.011. These results demonstrate the remarkable stability of the VD-MEAC stochastic policy, with even the worst-performing test achieving a Sharpe ratio of approximately 2.70.

Figure 5
Two histograms with density plots. (a) Shows a blue histogram for Sharpe Ratio with an overlayed red line, mean 2.965, and standard deviation 0.111. (b) Displays a green histogram for Wealth Ratio with a purple line, mean 2.492, and standard deviation 0.100.

Figure 5. Distribution of VD-MEAC trading results across 100 runs.

We selected one representative test result for comparison with other strategies. The comparative trading results are illustrated in Figure 6. Figure 6 clearly shows that although the VD-MEAC strategy lagged behind other strategies in the initial trading period, it significantly outperformed both baseline comparison groups (CSI 300, Equal-weight) and classical reinforcement learning algorithms (TD3, DDPG, SAC) throughout the remainder of the testing period.

Figure 6
Line chart depicting cumulative wealth growth from August 2020 to August 2021. VD-MEAC (blue line) shows significant growth, far surpassing other strategies like TD3 (orange), DDPG (green), SAC (red), CSI 300 (purple dashed), and Equal-weight (brown dashed).

Figure 6. Performance comparison of different portfolio strategies.

To conduct a more comprehensive comparison, we introduced additional quantitative metrics to evaluate portfolio performance, as shown in Table 4. The bold values indicate the best performance under each metric. From Table 4, we observe that the Equal-weight strategy’s annualized return of 0.0707 underperforms the CSI 300 index, while all reinforcement learning strategies surpass the CSI 300 index in terms of returns. The VD-MEAC strategy demonstrates superior performance with an annualized return of 1.1944, highlighting its strong profitability. Regarding risk management, VD-MEAC also significantly outperforms other strategies in terms of Sharpe ratio and Calmar ratio, confirming that leveraging more distribution information effectively enhances risk resistance.

Table 4
www.frontiersin.org

Table 4. Comparison of strategy evaluation metrics.

In investment, particular attention must be paid to drawdown metrics, as maximum drawdown measures the largest potential loss investors may experience, while drawdown duration affects investor confidence and subsequent trading decisions. As shown in Table 5, although DDPG slightly outperforms VD-MEAC in terms of maximum drawdown, DDPG never recovered to its highest wealth point by the end of the testing period, reflecting its inferior profitability compared to VD-MEAC. Crucially, durations in Table 5 marked with a > symbol (e.g., “>419 days”) indicate that the strategy failed to recover to its previous peak by the end of the testing period. Our VD-MEAC, in contrast, was one of only two strategies to achieve a full recovery, and it did so in only 169 days, demonstrating superior resilience. Under conditions where VD-MEAC’s wealth value is significantly higher than other strategies, VD-MEAC’s maximum drawdown period is shorter, demonstrating its exceptional recovery capability. Overall, considering multiple dimensions of assessment, VD-MEAC performs better than other strategies in terms of maximum drawdown.

Table 5
www.frontiersin.org

Table 5. Comparison of maximum drawdown periods.

4.3 Factor portfolio analysis

As shown in Figure 7, the factor importance analysis provides deep insights into the decision-making mechanics of the VD-MEAC algorithm. The dominance of risk-related factors, particularly Variance120 and Sharpe_ratio120, at the top of the ranking confirms our theoretical framework, which emphasizes comprehensive risk assessment as the primary determinant of portfolio allocation. The algorithm systematically places higher weight on long-horizon risk metrics (120-day measures) compared to short-term indicators, thereby filtering out market noise and focusing on persistent patterns of risk. The notable importance assigned to momentum (importance value = 0.118) reveals that VD-MEAC has internalized the predictive value of trend-following signals. At the same time, the strong contribution of fundamental measures such as the PEG ratio (importance value = 0.093) demonstrates that the algorithm integrates both technical and fundamental domains. This balance suggests the emergence of a sophisticated multi-factor framework that captures non-trivial interactions among diverse signals, without requiring explicit programming of factor interrelationships.

Figure 7
Bar chart showing various financial factors and their importance scores. Factors are categorized into Risk, Style, Growth, and Sentiment, distinguished by color. Variance120 has the highest importance score at 0.142, while VOSC has the lowest at 0.008.

Figure 7. Factor importance in VD-MEAC model.

To further enhance interpretability, we provide a visual example of the agent’s decision-making at a specific time step in Figure 8. While the factor importance analysis in Figure 7 provides a global view of which factors the model values most, Figure 7 offers a local interpretation for a single decision. Using a contribution analysis (akin to SHAP or LIME), we can visualize the factors that pushed the agent to increase or decrease its allocation to a specific asset. In this example, the agent’s decision to significantly increase allocation to ‘Yanzhou Coal Mining’ (YCM) on February 22, 2021, was primarily driven by a very strong ‘Sharpe_ratio120’ and a high ‘momentum’ factor, which offset the negative contribution from its ‘Variance120’ (which was high, but deemed acceptable given the risk-adjusted returns). This local-level insight is crucial for building practitioner trust, as it allows for an audit of the agent’s “reasoning” at critical market junctures.

Figure 8
Bar chart showing factors contributing to the action

Figure 8. Factor importance in VD-MEAC model.

The portfolio weight evolution, visualized in Figure 9, reveals that VD-MEAC adapts allocation strategies in a manner consistent with prevailing market conditions. During bullish phases, the algorithm increased exposure to cyclical sectors such as energy and materials (YCM, CMM), while simultaneously reducing allocations to technology (TU). This sectoral rotation aligns with the cyclical structure of financial markets. Conversely, in bearish conditions, the model exhibited a defensive posture, reducing cyclical exposures and reallocating toward more stable, defensive sectors. The periodicity observed in rebalancing suggests that the algorithm has implicitly discovered near-optimal rebalancing frequencies, despite the absence of explicit programming to that effect.

Figure 9
Stacked area chart showing portfolio weights from August 2020 to August 2021 with identifiable events: Tech Rotation, Market Correction, and Sector Reallocation. Legend indicates segments: YCM, YTO, ZNC, CMM, SJS, HT, TO, SOS, AE.

Figure 9. VD-MEAC portfolio weight dynamics.

The correlation matrix of factors, illustrated in Figure 10, highlights VD-MEAC’s ability to internalize interdependencies among explanatory variables. The emergence of distinct correlation clusters, especially within factors of the same type, indicates that the algorithm systematically accounts for redundancy in information. This suggests that VD-MEAC not only recognizes the presence of collinearity but also adjusts its weighting to prevent the double-counting of equivalent sources of risk.

Figure 10
Correlation heatmap displaying relationships among variables like Variance120, Sharpe_ratio120, and others. Colors range from blue (negative correlation) to red (positive correlation), with intensity indicating strength. A color bar on the right shows correlation values from negative one to positive one.

Figure 10. Factor correlation matrix.

Finally, the decision boundary analysis of the two most influential factors (Variance120 and Sharpe_ratio120), as depicted in Figure 11, offers an interpretable view of the algorithm’s internal logic. The non-linear geometry of the boundary confirms that VD-MEAC captures complex, non-linear relationships between risk characteristics and allocation choices. Importantly, the identified decision regions correspond closely to financial intuition: the algorithm increases exposure when variance is elevated but compensated by a high Sharpe ratio, and decreases exposure when variance is high but not accompanied by sufficient risk-adjusted return.

Figure 11
Scatter plot showing variance on the x-axis and Sharpe ratio on the y-axis. Dots are colored based on portfolio weight change, with a color bar ranging from blue (-2.4) to red (7.2). Annotations indicate

Figure 11. Decision boundary of top two factors.

4.4 Extended experiments

To address concerns about the generalizability of our findings from a single portfolio and to validate the individual contributions of our model’s components, we conducted three additional experiments: (1) hyperparameter sensitivity analysis, (2) performance under different market conditions, and (3) an ablation study examining the individual contributions of value distribution and maximum entropy components.

We examined the sensitivity of VD-MEAC to two key hyperparameters: the distribution information utilization coefficient ( β ) and the entropy regularization coefficient ( α ). Figure 12 shows how these parameters affect the Sharpe ratio and annualized return. The results reveal that the performance of VD-MEAC is relatively stable across a range of parameter values, with optimal performance achieved when β is around 0.75 and α is approximately 0.05. Too small values of β lead to insufficient utilization of distribution information, while too large values can include noisy extreme quantiles. Similarly, very small values of α result in insufficient exploration, while excessive values may lead to overly random policies.

Figure 12
Two line graphs display the relationship between financial metrics. The left graph plots the Sharpe Ratio and Annualized Return against the Distribution Information Utilization Coefficient (β), highlighting an optimal β with a 3.2 Sharpe Ratio and 1.35 Annualized Return. The right graph shows similar metrics for the Entropy Regularization Coefficient (α), with the Sharpe Ratio remaining around 3.17 and the Annualized Return near 1.28. Both graphs indicate optimal levels with marked peaks and labeled axes.

Figure 12. Hyperparameter sensitivity.

To assess the robustness of VD-MEAC across varying market conditions, we divided our test period into three market regimes: bullish (uptrend), bearish (downtrend), and sideways (neutral). Table 6 presents the performance metrics under each condition. The results demonstrate that VD-MEAC significantly outperforms other strategies across all market conditions, with particular strength during bearish markets where it maintains positive returns while other strategies experience losses. This highlights the algorithm’s robustness to varying market conditions, which is crucial for real-world portfolio management.

Table 6
www.frontiersin.org

Table 6. Performance under different market conditions.

To understand the individual contributions of the value distribution and maximum entropy components, we conducted an ablation study comparing four variants: (1) VD-MEAC (full algorithm), (2) VD-AC (without maximum entropy), (3) ME-AC (with maximum entropy but using traditional Q-learning), and (4) AC (basic actor-critic). The results are presented in Table 7. The ablation study confirms that both the value distribution and maximum entropy components contribute significantly to the algorithm’s performance. While each component individually improves performance over the basic actor-critic approach, their combination in VD-MEAC yields synergistic benefits, particularly in terms of risk-adjusted returns as measured by the Sharpe and Calmar ratios. This study provides definitive evidence that the superior, high-return performance of VD-MEAC is a direct result of its novel architecture, not an artifact of external factors such as leverage, which were explicitly forbidden.

Table 7
www.frontiersin.org

Table 7. Ablation study results.

To address concerns about selection bias from a single small portfolio, we conducted additional experiments on three different 10-stock portfolios randomly sampled from CSI 300 constituents:

Portfolio A: Original portfolio (Table 2).

Portfolio B: 10 stocks from financial and consumer sectors.

Portfolio C: 10 stocks from the technology and healthcare sectors.

Table 8 presents the performance comparison across all three portfolios. VD-MEAC consistently outperforms benchmarks across all portfolios, with average Sharpe ratios of 2.98 (Portfolio A), 2.76 (Portfolio B), and 2.85 (Portfolio C). This multi-portfolio validation demonstrates that our algorithm’s superior performance is not an artifact of a single favorable stock selection.

Table 8
www.frontiersin.org

Table 8. Multi-portfolio performance comparison.

4.5 Analysis and discussion

Combining evaluations across multiple dimensions, the VD-MEAC strategy demonstrates superior performance in both risk management and return generation compared to baseline strategies. Several key insights emerge from our experimental results:

First, the value distribution approach significantly enhances risk management by capturing the full uncertainty of returns rather than just point estimates. This is particularly evident in the reduced maximum drawdowns and shorter recovery periods exhibited by VD-MEAC.

Second, the maximum entropy component effectively encourages exploration of the action space, leading to the discovery of more profitable trading strategies. This is reflected in the substantially higher annualized returns achieved by VD-MEAC compared to other reinforcement learning algorithms.

Third, the stability of performance across 100 test runs (with very low variance in Sharpe and wealth ratios) demonstrates the robustness of VD-MEAC as a stochastic policy. This stability is crucial for real-world applications where consistent performance is valued.

Fourth, the outperformance of VD-MEAC across different market conditions (bullish, bearish, and sideways) highlights its adaptability to changing market environments, a critical advantage over traditional strategies that may perform well in certain market conditions but poorly in others.

Finally, the hyperparameter sensitivity analysis reveals that while the algorithm’s performance can be optimized through careful parameter tuning, it maintains strong performance across a reasonable range of parameter values, indicating robustness to hyperparameter settings.

In summary, our comprehensive experimental evaluation validates that the VD-MEAC algorithm effectively addresses the risk–return tradeoff in portfolio management, achieving superior risk-adjusted returns compared to both traditional investment strategies and state-of-the-art reinforcement learning methods.

5 Conclusion

This study proposes a portfolio management strategy built upon the VD-MEAC framework, which shifts the focus from single-point return predictions to modeling the entire distribution of outcomes. Such a formulation enhances the capacity to evaluate both profitability and downside risk, thereby strengthening the role of distributional reinforcement learning in financial decision-making. Empirical tests on real stock market data confirm the algorithm’s promising profitability and resilience under uncertainty, underscoring its alignment with practical investment logic. Nevertheless, these results are derived under a set of experimental assumptions that simplify real-world trading environments. Future research should narrow this gap by incorporating more realistic market frictions, transaction costs, and dynamic constraints. Moreover, while the model demonstrates strong performance, the opacity of the agent’s decision process remains a key limitation. While our Factor Portfolio Analysis in Section IV-C provides a significant degree of transparency into the model’s learned logic and decision-making process, addressing interpretability, potentially by integrating advances in explainable AI, will be critical for building investor trust and enabling deployment in live trading systems.

Furthermore, the VD-MEAC framework opens several avenues for future work. Extending the model to a multi-agent reinforcement learning (MARL) setting, where different agents manage different asset classes or cooperate/compete to optimize a joint portfolio, could capture more complex market dynamics. Additionally, exploring cross-market transfer learning, for instance, pre-training the agent on a data-rich market (e.g., the U.S. stock market) and subsequently fine-tuning it on another (e.g., the Chinese market), could significantly improve data efficiency and model generalization, aligning with current AI-in-finance trends.

Reproducibility: To ensure full reproducibility and facilitate further research, the complete source code, experimental framework, and trained models for this paper have been made publicly available at: https://github.com/YanYang/VD-MEAC

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author/s.

Author contributions

YY: Conceptualization, Writing – original draft, Writing – review & editing. TW: Methodology, Software, Writing – original draft, Writing – review & editing. YF: Methodology, Software, Writing – original draft, Writing – review & editing. JH: Methodology, Software, Writing – review & editing. DZ: Methodology, Software, Writing – review & editing.

Funding

The author(s) declared that financial support was not received for this work and/or its publication.

Conflict of interest

YY, TW and YF were employed by Strategic Development Department Southern Power Grid Capital Holding Co., Ltd., JH were employed by Southern Power Grid Financial Leasing Co., Ltd., DZ were employed by Southern Power Grid Private Fund Management Co., Ltd.

Generative AI statement

The author(s) declared that Generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Alzaman, C. (2025). Optimizing portfolio selection through stock ranking and matching: a reinforcement learning approach. Expert. Syst. Appl. 269:126430. doi: 10.1016/j.eswa.2025.126430

Crossref Full Text | Google Scholar

Aminifar, F., Abedini, M., Amraee, T., Jafarian, P., Samimi, M. H., and Shahidehpour, M. (2022). A review of power system protection and asset management with machine learning techniques. Energy Syst. 13, 855–892. doi: 10.1007/s12667-021-00448-6

Crossref Full Text | Google Scholar

Aritonang, P. K., Wiryono, S. K., and Faturohman, T. (2025). Hidden-layer configurations in reinforcement learning models for stock portfolio optimization. Intell Syst. Appl. 25:200467. doi: 10.1016/j.iswa.2024.200467

Crossref Full Text | Google Scholar

Belyakov, B., and Sizykh, D. (2024). Adaptive algorithm for selecting the optimal trading strategy based on reinforcement learning for managing a hedge fund. IEEE Access. 12, 189047–189063.

Google Scholar

Betancourt, C., and Chen, W.-H. (2021). Deep reinforcement learning for portfolio management of markets with a dynamic number of assets. Expert. Syst. Appl. 164:114002. doi: 10.1016/j.eswa.2020.114002

Crossref Full Text | Google Scholar

Cheng, L.-C., and Sun, J.-S. (2024). Multiagent-based deep reinforcement learning framework for multi-asset adaptive trading and portfolio management. Neurocomputing 594:127800. doi: 10.1016/j.neucom.2024.127800

Crossref Full Text | Google Scholar

Cui, Y., Han, X., Chen, J., Zhang, X., Yang, J., and Zhang, X. (2025). FraudGNN-RL: a graph neural network with reinforcement learning for adaptive financial fraud detection. IEEE Open J. Comput. Soc. 6:426–437. doi: 10.1109/OJCS.2025.3543450

Crossref Full Text | Google Scholar

Cui, T., Yang, X., Jia, F., Jin, J., Ye, Y., and Bai, R. (2024). Mobile robot sequential decision making using a deep reinforcement learning hyper-heuristic approach. Expert. Syst. Appl. 257:124959. doi: 10.1016/j.eswa.2024.124959

Crossref Full Text | Google Scholar

Day, M.-Y., Yang, C.-Y., and Ni, Y. (2024). Portfolio dynamic trading strategies using deep reinforcement learning. Soft. Comput. J. 28, 8715–8730.

Google Scholar

de López Prado, M., Simonian, J., Fabozzi, F. A., and Fabozzi, F. J. (2025). Enhancing Markowitz's portfolio selection paradigm with machine learning. Ann. Oper. Res. 346, 319–340. doi: 10.1007/s10479-024-06257-1

Crossref Full Text | Google Scholar

Dong, Z., Fan, X., and Peng, Z. (2024). “Fnspid: a comprehensive financial news dataset in time series” in Proceedings of the 30th ACM SIGKDD conference on knowledge discovery and data mining (New York: ACM Digital Library), 4918–4927.

Google Scholar

Du, A., and Ghavidel, A. (2022). Parameterized deep reinforcement learning-enabled maintenance decision-support and life-cycle risk assessment for highway bridge portfolios. Struct. Saf. 97:102221. doi: 10.1016/j.strusafe.2022.102221

Crossref Full Text | Google Scholar

Fu, Y.-T., and Huang, W.-C. (2025). Optimizing stock investment strategies with double deep Q-networks: exploring the impact of oil and gold price signals. Appl. Soft. Comput. 180:113264. doi: 10.1016/j.asoc.2025.113264,

PubMed Abstract | Crossref Full Text | Google Scholar

Gong, X., Xie, F., Zhou, Z., and Zhang, C. (2024). The enhanced benefits of ESG in portfolios: a multi-factor model perspective based on LightGBM. Pac. Basin Financ. J. 85:102365. doi: 10.1016/j.pacfin.2024.102365

Crossref Full Text | Google Scholar

Haarnoja, T., Zhou, A., Abbeel, P., and Levine, S. (2018). “Soft actor-critic: off-policy maximum entropy deep reinforcement learning with a stochastic actor” in International conference on machine learning (New York: Pmlr), 1861–1870.

Google Scholar

Jang, J., and Seong, N. Y. (2023). Deep reinforcement learning for stock portfolio optimization by connecting with modern portfolio theory. Expert. Syst. Appl. 218:119556. doi: 10.1016/j.eswa.2023.119556

Crossref Full Text | Google Scholar

Jeribi, F., Martin, R. J., Mittal, R., Jari, H., Alhazmi, A. H., Malik, V., et al. (2024). A deep learning based expert framework for portfolio prediction and forecasting. IEEE Access 12, 103810–103829. doi: 10.1109/ACCESS.2024.3434528

Crossref Full Text | Google Scholar

Jiang, Y., Olmo, J., and Atwi, M. (2024). Deep reinforcement learning for portfolio selection. Glob. Financ. J. 62:101016. doi: 10.1016/j.gfj.2024.101016

Crossref Full Text | Google Scholar

Jiang, Z., Xu, D., and Liang, J. (2017). A deep reinforcement learning framework for the financial portfolio management problem. arXiv preprint arXiv:1706.10059.

Google Scholar

Jin, L., Kim, D., Chan, K. Y., and Abu-Siada, A. (2024). Deep machine learning-based asset management approach for oil-immersed power transformers using dissolved gas analysis. IEEE Access 12, 27794–27809. doi: 10.1109/ACCESS.2024.3366905

Crossref Full Text | Google Scholar

Junfeng, W., Li, Y., Tan, W., and Chen, Y. (2024). Portfolio management based on a reinforcement learning framework. J. Forecast. 43, 2792–2808.

Google Scholar

Kitchat, K., Lin, M.-H., Chen, H.-S., Sun, M.-T., Sakai, K., Ku, W.-S., et al. (2024). A deep reinforcement learning system for the allocation of epidemic prevention materials based on DDPG. Expert. Syst. Appl. 242:122763. doi: 10.1016/j.eswa.2023.122763

Crossref Full Text | Google Scholar

Koratamaddi, P., Wadhwani, K., Gupta, M., and Sanjeevi, S. G. (2021). Market sentiment-aware deep reinforcement learning approach for stock portfolio allocation. Eng. Sci. Technol. Int. J. 24, 848–859. doi: 10.1016/j.jestch.2021.01.007

Crossref Full Text | Google Scholar

Li, H., and Hai, M. (2024). Deep reinforcement learning model for stock portfolio management based on data fusion. Neural Process. Lett. 56:108. doi: 10.1007/s11063-024-11582-4

Crossref Full Text | Google Scholar

Li, F., Qu, H., Zhang, L., Fu, M., Chen, W., and Yi, Z. (2024). Q-ADER: an effective Q-learning for recommendation with diminishing action space. IEEE Trans. Neural. Netw. Learn. Syst. 56–108.

Google Scholar

Pallathadka, H., Mustafa, M., Sanchez, D. T., Sajja, G. S., Gour, S., and Naved, M. (2023). Impact of machine learning on management, healthcare and agriculture. Mater. Today Proc. 80, 2803–2806. doi: 10.1016/j.matpr.2021.07.042

Crossref Full Text | Google Scholar

Pippas, N., Ludvig, E. A., and Turkay, C. (2025). The evolution of reinforcement learning in quantitative finance: a survey. ACM Comput. Surv. 57, 1–51. doi: 10.1145/3733714

Crossref Full Text | Google Scholar

Rezaei, M., and Nezamabadi-Pour, H. (2025). A taxonomy of literature reviews and experimental study of deepreinforcement learning in portfolio management. Artif. Intell. Rev. 58:94. doi: 10.1007/s10462-024-11066-w

Crossref Full Text | Google Scholar

Sattar, A., Sarwar, A., Gillani, S., Bukhari, M., Rho, S., and Faseeh, M. (2025). A novel RMS-driven deep reinforcement learning for optimized portfolio management in stock trading. IEEE Access 13:42813–42835. doi: 10.1109/ACCESS.2025.3546099

Crossref Full Text | Google Scholar

Teoh, Y. K., Gill, S. S., and Parlikad, A. K. (2021). IoT and fog-computing-based predictive maintenance model for effective asset management in industry 4.0 using machine learning. IEEE Internet Things J. 10, 2087–2094. doi: 10.1109/JIOT.2021.3050441,

PubMed Abstract | Crossref Full Text | Google Scholar

Wu, M.-E., Syu, J.-H., Lin, J. C.-W., and Ho, J.-M. (2021). Portfolio management system in equity market neutral using reinforcement learning. Appl. Intell. 51, 8119–8131. doi: 10.1007/s10489-021-02262-0

Crossref Full Text | Google Scholar

Xu, Y. (2025). Deep reinforcement learning-driven intelligent portfolio management with green computing: sustainable portfolio optimization and management. Sustain. Comput. Inf. Syst. 46:101125. doi: 10.1016/j.suscom.2025.101125,

PubMed Abstract | Crossref Full Text | Google Scholar

Yang, D., Zhao, L., Lin, Z., Qin, T., Bian, J., and Liu, T.-Y. (2019). Fully parameterized quantile function for distributional reinforcement learning. Adv. Neural Inf. Process. Syst. 32.

Google Scholar

Zhao, L.-Y., Chang, T.-Q., Guo, L.-B., Zhang, J., Zhang, L., and Ma, J.-D. (2024). An overestimation reduction method based on the multi-step weighted double estimation using value-decomposition multi-agent reinforcement learning. Neural Process. Lett. 56:152. doi: 10.1007/s11063-024-11611-2

Crossref Full Text | Google Scholar

Zhu, Z., Gao, X., Bu, S., Chan, K. W., Zhou, B., and Xia, S. (2024). Cooperative dispatch of renewable-penetrated microgrids alliances using risk-sensitive reinforcement learning. IEEE Trans. Sustain. Energy 15, 2194–2208. doi: 10.1109/TSTE.2024.3406590

Crossref Full Text | Google Scholar

Keywords: portfolio optimization, reinforcement learning, value distribution risk management, quantitative finance, actor-critic algorithm

Citation: Yang Y, Wang T, Fu Y, Huang J and Zhou D (2026) Portfolio management based on value distribution reinforcement learning algorithm. Front. Artif. Intell. 8:1709493. doi: 10.3389/frai.2025.1709493

Received: 23 September 2025; Revised: 13 November 2025; Accepted: 02 December 2025;
Published: 30 January 2026.

Edited by:

Jiancheng Jiang, University of North Carolina at Charlotte, United States

Reviewed by:

Xiang Shi, Millennium Management, United States
Bertrand Kian Hassani, University College London, United Kingdom

Copyright © 2026 Yang, Wang, Fu, Huang and Zhou. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Yan Yang, d3QxOTg1ZXN3YUAxNjMuY29t

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.