Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Robot. AI, 01 August 2025

Sec. Computational Intelligence in Robotics

Volume 12 - 2025 | https://doi.org/10.3389/frobt.2025.1645927

This article is part of the Research TopicGenerative AI and Intelligent Control in Robotics for Deployment in Challenging EnvironmentsView all articles

Intelligent maneuver decision-making for UAVs using the TD3–LSTM reinforcement learning algorithm under uncertain information

Tongle Zhou
Tongle Zhou*Ziyi LiuZiyi LiuWenxiao Jin&#x;Wenxiao Jin Zengliang HanZengliang Han
  • College of Automation Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing, China

Aiming to address the complexity and uncertainty of unmanned aerial vehicle (UAV) aerial confrontation, a twin delayed deep deterministic policy gradient (TD3)–long short-term memory (LSTM) reinforcement learning-based intelligent maneuver decision-making method is developed in this paper. A victory/defeat adjudication model is established, considering the operational capability of UAVs based on an aerial confrontation scenario and the 3-degree-of-freedom (3-DOF) UAV model. For the purpose of assisting UAVs in making maneuvering decisions in continuous action space, a model-driven state transition update mechanism is designed. The uncertainty is represented using the Wasserstein distance and memory nominal distribution methods to estimate the detection noise of the target. On the basis of TD3, an LSTM network is utilized to extract features from high-dimensional aerial confrontation situations with uncertainty. The effectiveness of the proposed method is verified by conducting four different aerial confrontation simulation experiments.

1 Introduction

Unmanned aerial vehicles (UAVs) have undergone significant development over recent years, offering advantages such as cost-effectiveness, strong maneuverability, good stealth performance, and the ability to be recycled and reused. It will gradually replace manned aircraft in future complex environments for performing regional reconnaissance, attacking targets, and other tasks (Shin et al., 2018; Zhou et al., 2020b). The process by which UAVs automatically make control decisions by simulating pilots and commanders who respond to various air combat situations is a crucial component of the autonomous decision-making system for aerial confrontations (Zhou et al., 2020a; Wang et al., 2020). As a result, it is critical to develop an intelligent maneuver decision-making approach to enhance UAV autonomy, intelligence, and air combat capability and adapt to the real-time demands of unmanned aerial confrontation.

The OODA (observation, orientation, decision, and action) loop theory governs maneuver decision-making in UAV confrontations (Virtanen et al., 2006). The UAV maneuver decision-making theory has gained significant attention and has been widely studied over the past decade due to advancements in the autonomy and intelligence of UAVs. At the moment, research on the intelligent maneuver decision-making method of UAVs is centered on three areas: expert systems, game theory, and learning algorithms.

Expert system-based maneuver decision-making formulates predicate logic production rules following “if–else–then” principles, upon which UAVs base tactical maneuver selections (Virtanen et al., 2006). The expert system is widely used in actual maneuver decision-making systems due to its simplicity in design and the interpretability of the decision outcomes. However, the expert system is overly reliant on rule dependability and lacks scalability. To improve the adaptability of expert systems in aerial combat, Liu et al. (2016) developed a receding horizon control-based maneuver decision-making method. Tan et al. (2022) developed a fuzzy expert system for UAV short-range escape maneuvers by learning tactical information.

Based on optimization theory, aerial confrontation is considered a strategic game involving decision-makers. The state transition during this process is described by a differential equation (payment function), and the maneuver decision-making problem is subsequently resolved through numerical optimization techniques (Park et al., 2016; Duan et al., 2015). Li Q. et al. (2022) established a decision-making model for maneuver games. It was based on positional situation information, fighter performance, intentional threat, and the collaborative effects of multiple fighters. The optimal decision scheme for the game was determined using the double game tree distributed Monte Carlo search strategy. Duan et al. (2023) designed a game with a mixed objective function for UAV autonomous maneuver decision-making problems, and the optimal solution was obtained using improved pigeon-inspired optimization.

With the accelerated advancement of artificial intelligence technology and computer processing power, deep learning and reinforcement learning algorithms have grown in popularity and are widely employed in unmanned systems. For the deep learning-based maneuver decision-making method, the situation information and UAV performance parameters are input into the deep networks, and the maneuvering action or control command is obtained after training and learning (Zhang and Huang, 2020). Relatively, the idea of reinforcement learning is more aligned with maneuver decision-making of UAVs. Based on the reinforcement learning Markov decision process, a UAV can select the corresponding maneuvering action through the real-time assessment of the environmental situation (Zhang et al., 2018; Tu et al., 2021). A heuristic deep deterministic policy gradient (DDPG) algorithm was introduced to improve the exploration capability in continuous action space for the UAV maneuver decision-making problem (Wang et al., 2022). A deep reinforcement learning and Monte Carlo tree search-based maneuver decision-making method, independent of human knowledge, was proposed by Zhang et al. (2022). Utilizing self-play, the system initiates with random actions to generate air combat training samples (including states, actions, and rewards).

Due to the complexity of the environment in actual aerial confrontations, the traditional discrete maneuver action library struggles to meet the demand for maneuvering diversity. The learning and training of continuous maneuvers require higher algorithm real-time efficiency. The twin delayed deep deterministic policy gradient algorithm (TD3) is a deterministic strategy reinforcement learning algorithm designed for high-dimensional continuous action spaces. It offers significant advantages in offline training plasticity and the real-time usage of neural networks (Hong et al., 2021). Furthermore, the long short-term memory (LSTM) network can transform the aerial confrontation state with uncertainty into a high-dimensional perceptual situation and improve the neural network learning ability of the target state time series data (Kant et al., 2023). Hence, the TD3–LSTM reinforcement learning-based intelligent algorithm is developed to address the UAV maneuver decision-making problem under uncertain information. The following are the major contributions:

A victory/defeat adjudication model is established based on the actual UAV aerial confrontation scenario, which could ensure the validity of maneuver decision-making.

A model-driven state transition update mechanism is developed based on the 3-degree-of-freedom (3-DOF) UAV model to ensure the efficiency of the continuous action space.

A Wasserstein distance-based model aims to describe uncertainty in confrontation, which can enhance the robustness of maneuver decision-making.

A reinforcement learning intelligent algorithm is proposed based on TD3–LSTM to improve the efficiency of maneuver decision-making.

The remainder of this paper is organized as follows. Section 2 details the problem formulation, including the one-to-one confrontation model, the victory/defeat adjudication model, and the maneuver decision-making system structure. The deep reinforcement learning-based UAV maneuver decision-making method, which consists of a model-driven state transition update mechanism, reward function design, uncertainty description, and the TD3–LSTM algorithm, is introduced in Section 3. In Section 4, we provide the simulation results. Finally, the conclusion is presented in Section 5.

2 Problem description

Maneuver decision-making is the process by which a UAV selects maneuvering actions based on the current aerial confrontation situation and environmental information, aiming to gain operational superiority by altering the aerial confrontation dynamics. The maneuver decision problem belongs to the top-level decision problem of UAVs.

2.1 Confrontation model of UAV

The UAV one-on-one confrontation scenario is shown in Figure 1 (Yang et al., 2019).

Figure 1
Diagram showing a 3D coordinate system with an orange target and a blue UAV. The target is positioned at \( P_R(x_R, y_R, z_R) \) with velocity vector \( \vec{v}_R \). The UAV is at \( P_B(x_B, y_B, z_B) \) with velocity vector \( \vec{v}_B \), forming an angle \( \alpha \) with the line of sight vector \( \vec{D} \). Axes labeled X, Y, and Z converge at point \( O \).

Figure 1. UAV one-on-one confrontation scenario.

In Figure 1, vB and PB(xB,yB,zB) denote the velocity vector and position vector of the UAV, respectively, while vR and PR(xR,yR,zR) represent the velocity vector and position vector of the target, respectively. α represents the relative azimuth angle.

Defining D as the relative distance of the UAV with respect to the target—where the direction represents the UAV pointing toward the target and the magnitude is given by dD and d can be calculated by Equations 1, 2 (Li B. et al., 2022)

D=DRDB,(1)
d=D2.(2)

Thus, the relative azimuth angle α can be calculated as Equation 3 (Li B. et al., 2022):

α=arccosD×vB/D2vB2.(3)

2.2 Victory/defeat adjudication model

Generally, the operational capability of a UAV is constrained by the capabilities of its weapon system (Luo et al., 2022). A schematic showing the UAV attacking and locking onto the target is presented in Figure 2.

Figure 2
Diagram showing a blue UAV on the left targeting an orange aircraft on the right with a cone of surveillance extending from the UAV. Labels include positions \(P_B\) for the UAV, \(P_R\) for the target, with maximum velocity \(v_{max}\) and engagement distances \(D_{max}\) and \(t_{min}^{lock}\) for locking onto the target.

Figure 2. Schematic of the UAV attacking and locking the target.

In Figure 2, υmax denotes the maximum off-axis emission angle of the UAV weapon system, Dmaxa stands for the maximum firing distance, and tminlock indicates the minimum locking time. Moreover, the UAV wins the confrontation when it locks onto the target for more than tminlock seconds within the UAV firing range. Specifically, the victory adjudication condition can be described as Equation 4:

dDmaxaαυmaxtintminlock,(4)

where tin represents the time the UAV locks onto the target.

Based on the points discussed above, the objective of this paper is to design an intelligent algorithm that enables UAVs to make autonomous decisions to achieve victory adjudication conditions in advance, based on environmental and situational information.

2.3 Maneuver decision-making system structure of UAVs

The UAV maneuver decision-making system is comprised of three components: the situation assessment module, the maneuver decision-making module, and the flight drive module. The situation assessment module obtains relative situational information from the environment and determines whether the conditions for adjudicating victory are met. If not, the maneuver decision module provides the maneuver command based on the situation assessment result. Afterward, the flight drive module updates the UAV state and provides feedback to the environment. This cycle would continue until one side achieves victory through adjudication. The interactive procedure is shown in Figure 3.

Figure 3
Flowchart depicting a battlefield decision-making process. It consists of four modules:

Figure 3. Interactive procedure of the UAV maneuver decision-making system.

3 Deep reinforcement learning-based UAV maneuver decision-making algorithm

To ensure that UAVs meet the victory adjudication condition ahead of schedule, a deep reinforcement learning-based algorithm is proposed in this paper for UAV maneuver decision-making. The basic framework is shown in Figure 4.

Figure 4
Diagram illustrating a UAV maneuver decision-making system integrated with deep reinforcement learning. It shows a flow of state, state update, and reward between the UAV, confrontation environment, and target. The target maneuver decision-making system includes fixed, expert rules-based, and intelligent algorithm-based strategies, interacting with the UAV system within the environment.

Figure 4. Basic framework of the deep reinforcement learning-based UAV maneuver decision-making algorithm.

3.1 UAV model-driven state transition update mechanism

The maneuver action of the UAV is designed as a discrete set by NASA, which only considers several basic maneuvers, including uniform flight, accelerated flight, decelerated flight, left turn, right turn, forward climb, and forward dive (Huang et al., 2018). To further describe the maneuvering behavior of the UAV with continuous state in actual aerial confrontations, this paper establishes a state transition update mechanism based on the UAV motion model. The schematic diagram of the UAV motion model is shown in Figure 5 (Guo et al., 2023).

Figure 5
A diagram of an aircraft with velocity vectors in a three-dimensional coordinate system labeled X, Y, and Z. Angles ψ, γ, and φ are shown along with vectors v and nₓ.

Figure 5. Schematic diagram of the UAV motion model.

In Figure 5, v, ψ, γ, and ϕ represent the velocity, yaw angle, flight path angle, and roll angle, respectively.

The following particle motion model (Guo et al., 2023) of UAVs is considered in this paper.

ẋ=vcosγsinψẏ=vcosγcosψż=vsinγv̇=gnxsinγψ̇=gnzsinϕvcosγγ̇=gvnzcosϕcosγ,(5)

where x, y, and z denote the position of the UAV in the inertial coordinate system. ẋ, ẏ, and ż denote the projections of velocity v in the x-axis, y-axis, and z-axis, respectively. nx and nz denote the normal overload and tangential overload of the UAV, respectively. g is the acceleration of gravity.

According to Equation 5, the state of the UAV is affected by normal overload nx, tangential overload nz, and roll angle ϕ. Thus, the 3-DOF model mentioned above is established as a flight drive module in this paper. The new state of the UAV is calculated in real-time based on the current state and control input. The UAV motion model-driven state transition update mechanism is shown in Figure 6.

Figure 6
Flowchart depicting a flight control process. It has three main components:

Figure 6. Schematic diagram of UAV dynamic model.

3.2 Design of the reward function

As common knowledge dictates, there are four possible scenarios in an aerial confrontation, depending on the relative positions of the UAV and the target: the trailing side holds the advantage, the pursued side is at a disadvantage, and an equilibrium state is reached when both sides are flying either toward or away from each other. To enable UAVs to reach positions with more favorable environmental conditions, this paper considers the instantaneous aerial situation between the UAV and the target, along with the victory/defeat adjudication model, as the basis for reward and punishment signals.

The angle reward rα is defined as Equation 6 (He et al., 2020):

rα=1αB+αR2π,dDmaxa1αB+αR2πedDmaxa22Dmaxa2,d>Dmaxa,(6)

where αB and αR represent the relative azimuth angle of the UAV and the target, respectively.

When the target is within the attack range of the UAV, the distance reward of the UAV to the target rBR is defined as Equation 7 (He et al., 2020):

rBR=5,dDmaxaandαB<υmaxandαR<π/20,else.(7)

Similarly, when the UAV is within the attack range of the target, the distance reward of the target to the UAV rRB is defined as follows (He et al., 2020):

rRB=5,dDmaxaandαR<υmaxandαB<π/20,else.(8)

Invoking Equations 8, 9, the distance reward rd is defined as follows (He et al., 2020):

rd=rBRrRB.(9)

In addition, to ensure the UAV flight safety and avoid collisions, the height reward rh is defined as Equation 10 (Li et al., 2021):

rh=10,1km<h<12kmandd>200m0,else,(10)

where h represents the flight height of the UAV.

In summary, the total reward R is obtained by Equation 11:

R=rα+rd+rh.(11)

3.3 Uncertain information of maneuver decision-making

Due to the uncertainties in the UAV model and the complexity and flexibility of the actual aerial confrontation environment, the UAV and the target may not be able to reach the desired position after maneuvering during the aerial confrontation.

To describe the uncertainty of aerial confrontation, we assume that the state deviation Δs after a maneuver action follows a discrete empirical distribution Ω̂ constructed from observations. The nominal distribution Ω̂ can be indirectly observed through the historical data samples OHi, where i{1,2,,N}, and the current data samples OCj, where j{1,2,,M}.

Under the uniform distribution over the N historical data samples and M current data samples, we have

Ω̂=1N+λMi=1NξOHi+λN+λMi=1NξOCj,(12)

In Equation 12, ξOHi denotes the Dirac point mass at the ith historical data sample OHi, ξOCj denotes the Dirac point mass at the jth current data sample OCj, and λ>1 is the proportional parameter, representing that the current data samples are more valuable.

Moreover, to describe the true distribution of target state deviation Δs, the Wasserstein distance is considered to measure the distance between any two distributions. Lemma 1 defines the distance.

Lemma 1 (Wasserstein distance (Esfahani and Kuhn, 2018)): The Wasserstein distance between any probability distributions Θ and the nominal distribution Ω̂ is defined as Equation 13:

WpΘ,Ω̂=infπΠΘ,Ω̂Ξ×ΞΔsΔsπdΔs,dΔs,(13)

where is a norm and Π(Θ,Ω̂) is the set of all joint probability distributions of Δs and Δs with the marginals Θ and Ω̂, respectively.

According to Lemma 1, the true distribution of the state deviation Δs is defined in a set as Equation 14:

ΛΩ̂=ΔsPΞ:WpΘ,Ω̂ηk,(14)

where Ξ is a closed set containing all possible values of Δs, P(Ξ) is the family of all probability distributions supported on Ξ, and ηk is the maximum Wasserstein distance.

On the basis of this, the state deviation Δs is considered to ensure accuracy in this paper. The schematic diagram is shown in Figure 7.

Figure 7
Flowchart illustrating a maneuver decision process. It includes

Figure 7. Schematic diagram of maneuver decision-making uncertainty.

3.4 TD3–LSTM algorithm

The TD3 algorithm is a novel form of deep reinforcement learning algorithm, founded on the policy gradient algorithm and the actor–critic (AC) framework (Cheng et al., 2021; Duan et al., 2022). The proposed UAV model-driven state transition update mechanism in this paper is suitable and can be applied in continuous state and action spaces.

The TD3 algorithm consists of six networks, namely, the actor network πμ, the critic 1 network Qθ1, the critic 2 network Qθ2, the target actor network πμ, the target critic 1 network Qθ1, and the target critic 2 network Qθ2. The network parameters for each network are μ, θ1, θ2, μ, θ1, and θ2. Based on the two critic networks’ structure, the TD3 algorithm selects the smaller Q value to alleviate overestimation.

In each episode, TD3 selects an action aπϕ(s)+ϵ,ϵN(0,σ), using exploration noise, and observes reward r and new state s, where ϵ denotes the noise added to the output of the policy network to enhance the stability of the algorithm, as specified in Equation 15 (Fujimoto et al., 2018).

ãπφs+ε,εclipN0,δ̃,c,c.(15)

Based on the structure of the two critic networks, the TD3 algorithm selects the minimum between the two estimates of Qθ1 and Qθ2 to avoid overestimation. The objective function of the TD3 algorithm is defined as Equation 16 (Fujimoto et al., 2018):

yr+γmini=1,2Qθis,ã.(16)

The TD-error eTD of Qθ1 and Qθ2 is defined as Equation 17 (Fujimoto et al., 2018):

eTD=yQθis,aθi.(17)

The critics can be updated as Equation 18 (Fujimoto et al., 2018):

θiargminθiN1yQθis,a2.(18)

The actor network is updated via the deterministic policy gradient as Equation 19 (Fujimoto et al., 2018):

μJμ=N1aQθ1s,aa=πμsμπμs.(19)

The target networks are updated through a slow-moving update rate τ, following Equations 20, 21. Specifically,

θiτθi+1τθi,(20)
μτμ+1τμ.(21)

Due to the high dynamic and high-dimensional complexity in an actual aerial confrontation environment, the TD3 algorithm cannot effectively manage uncertainty and model the policy function and value function using a fully connected neural network. By adding a special gate structure to RNN, the LSTM network has a positive impact on processing time series data, thus enhancing the efficiency and effectiveness of the training algorithm (T. Ergen, 2018). It is considered that the state information of UAVs and targets in actual aerial confrontations exhibits time series characteristics. In this paper, an LSTM network is utilized to extract features from high-dimensional aerial confrontation situations with uncertainty. This structure aims to output valuable perceptual information and advance representation learning for sequential sample data. The policy and value functions are jointly approximated using a fully connected neural network. The TD3–LSTM algorithm architecture diagram is shown in Figure 8.

Figure 8
Diagram illustrating a reinforcement learning framework using LSTM networks. The process involves actors and critics, each with policy and value networks optimized by Adam optimizers. The system updates parameters through policy gradients, experiences uncertainty inputs, and uses an experience replay buffer for training. It includes TD error calculation and connections to a confrontation environment, depicting the flow of states, actions, rewards, and next states.

Figure 8. Architecture diagram of the TD3–LSTM algorithm.

The general structure of the LSTM policy network is shown in Figure 9.

Figure 9
Diagram of a neural network architecture showing an input layer, an LSTM network layer, a fully connected network layer, and an output layer. The input layer receives \( s_t + \Delta s_t \). The LSTM layer contains sequences \( X_0, X_1, X_2 \) connected to hidden states \( h_0, h_1, h_2 \). The output layer produces \( a_t \), with arrows indicating data flow between layers.

Figure 9. General structure of the LSTM policy network.

The structural parameters of the LSTM policy network are provided in Table 1.

Table 1
www.frontiersin.org

Table 1. Structural parameters of the LSTM policy network.

The general structure of the LSTM value network is presented in Figure 10.

Figure 10
Diagram of a neural network architecture. It includes an input layer with state \( s_t \) and action \( a_t \), followed by an LSTM network layer with states \( h_0, h_1, h_2 \) and hidden units \( X_0, X_1, X_2 \). This connects to a fully connected network layer, ending with an output \( Q \). Arrows indicate data flow direction.

Figure 10. General structure of the LSTM value network.

The structural parameters of the LSTM value network are shown in Table 2.

Table 2
www.frontiersin.org

Table 2. Structural parameters of the LSTM value network.

4 Simulation

To demonstrate the advantages of the proposed maneuver decision-making method, the simulation is verified in four different scenarios.

4.1 Scenario 1: target in straight flight

The initial position and attitude information of the UAV and the target are randomly initialized, and the target follows a strategy of uniform-speed straight flight. The confrontation trajectory of the UAV and the target in scenario 1 is depicted in Figure 11.

Figure 11
3D plot depicting two flight paths. The blue path, labeled

Figure 11. Confrontation trajectory in scenario 1.

The real-time reward curve of the UAV and the target in scenario 1 is depicted in Figure 12.

Figure 12
Line graph showing

Figure 12. Real-time reward curve in scenario 1.

On the basis of Figures 11, 12, the UAV gains a height advantage by climbing and then dives toward the target after reaching a specific altitude to gain velocity and angle advantages. This maneuver forces the target into the attack zone, leading to eventual triumph.

Figure 13 shows the cumulative reward curve of the UAV and the target.

Figure 13
Line graph depicting rewards against episodes ranging from 0 to 10,000. The light blue line represents individual rewards with noticeable fluctuations and drops. The dark blue line shows the average rewards over the last 100 episodes, remaining relatively stable near zero after initial volatility.

Figure 13. Cumulative reward curve in scenario 1.

In the early stages, the cumulative reward curve fluctuates because the UAV is unable to learn any effective strategies, leading to crashes or defeats in confrontations. With further training, the UAV learned effective maneuvers, developed an attack posture, locked onto the target, and gradually increased its cumulative reward value until convergence. It should be noted that the cumulative reward value may fluctuate slightly during the later stages of training due to uncertainty considerations. However, this variation will not impact the eventual acquisition of effective maneuvering strategies.

4.2 Scenario 2: target in circle flight

In scenario 2, the target employs the circle maneuver strategy. The trajectory of the UAV and the target during the confrontation is illustrated in Figure 14.

Figure 14
3D plot showing the paths of a UAV and a target. The UAV path is represented by a blue line with directional markers, while the target path is depicted with a red loop. Both paths are plotted on grid axes.

Figure 14. Confrontation trajectory in scenario 2.

The real-time reward curve of the UAV and the target in scenario 2 is presented in Figure 15.

Figure 15
Line graph showing reward versus step. The blue line represents UAV reward, increasing sharply after step 100. The red line represents target reward, remaining nearly constant. Legends are at the top left.

Figure 15. Real-time reward curve in scenario 2.

According to Figures 14, 15, the UAV climbs to gain a height advantage before diving toward the target to lock onto it for the first time. However, the target continues to circle due to the insufficient lock time. The UAV then ascends to gain a height advantage, locks onto the target again, and maintains the lock until it wins the confrontation.

Figure 16 illustrates the cumulative reward curve of the UAV and the target in scenario 2.

Figure 16
Line graph showing rewards over 20,000 episodes. The light blue line represents individual rewards fluctuating greatly, while the dark blue line shows the average of 100 rewards stabilizing around zero.

Figure 16. Cumulative reward curve in scenario 2.

Similarly, the UAV failed to learn any useful strategies at the beginning of the training. Once the UAV learns efficient maneuvers, the cumulative reward value gradually increases until convergence. The cumulative reward value also fluctuates slightly in the late training period due to the effect of uncertainty.

4.3 Scenario 3: expert rule-based target maneuvering strategy

In scenario 3, the target selects maneuvers according to expert rules. The confrontation trajectory of the UAV and the target in scenario 3 is illustrated in Figure 17.

Figure 17
A 3D plot depicting the trajectories of UAV_xyz in blue and Target_xyz in red. The paths intertwine with blue and orange arrowheads indicating direction. Axes display coordinates ranging from 0 to 40,000 on x and y, and 0 to 12,000 on z. A legend distinguishes the trajectories.

Figure 17. Confrontation trajectory in scenario 3.

The real-time reward curve of the UAV and the target in scenario 3 is shown in Figure 18.

Figure 18
Line graph depicting reward against steps. The blue line represents UAV reward, and the red line represents Target reward. Both remain near zero until step 70, after which both increase sharply, with UAV reward surpassing Target reward.

Figure 18. Real-time reward curve in scenario 3.

From Figures 17, 18, the UAV and the target climb simultaneously to gain a height advantage during the initial phase. Next, both the UAV and the target choose a turning strategy to prevent being locked onto each other. Finally, the UAV performs a somersault maneuver to circle behind the target, completes the lock, and triumphs over the confrontation.

Figure 19 shows the cumulative reward curve of the UAV and the target.

Figure 19
Line graph showing rewards over 200,000 episodes, with rewards fluctuating significantly in light blue and the average of the last 100 rewards stabilizing near zero in dark blue.

Figure 19. Cumulative reward curve in scenario 3.

Based on the results in Figure 19, the confrontation process is more intensive since the target has a specific maneuvering strategy. Because the UAV lacks environmental cognition, it is unable to develop effective strategies, leading to the UAV acting with high penalty values in the initial stages of training. The trend in the cumulative reward curve change indicates that the convergence speed is relatively slow, and the curve fluctuates sharply, reflecting the difficulty and complexity of aerial confrontation.

4.4 Scenario 4: genetic algorithm-based target maneuvering strategy

In scenario 4, the target selects the optimal maneuvers based on the current situation using a genetic algorithm. The confrontation trajectory of the UAV and the target in scenario 4 is shown in Figure 20.

Figure 20
3D plot showing two paths. The blue line represents UAV_xyz, and the red line represents Target_xyz. Both lines have arrowheads indicating direction. The axes range from 0 to 40,000 on the x-y plane, and 0 to 10,000 on the z-axis.

Figure 20. Confrontation trajectory in scenario 4.

The real-time reward curve of the UAV and the target in scenario 4 is shown in Figure 21.

Figure 21
Line graph showing two reward metrics over steps. UAV Reward in blue and Target Reward in red. Both rise sharply after step 100, but Target Reward drops suddenly after peaking.

Figure 21. Real-time reward curve in scenario 4.

As observed from Figures 20, 21, both the UAV and the target climb simultaneously, aiming to gain a height advantage at the start. Next, the target continues to climb in an attempt to gain a vantage point. The UAV takes the opportunity and maneuvers around the rear of the target during this process. Finally, the UAV completes the lock and wins the confrontation.

Figure 22 shows the cumulative reward curve of the UAV and the target.

Figure 22
Line graph showing rewards over episodes from zero to five hundred thousand. Light blue line depicts volatile individual rewards, while dark blue line shows a smoothed average of one hundred episode rewards, stabilizing around zero over time.

Figure 22. Cumulative reward curve in scenario 4.

As shown in Figure 22, the target selects the optimal maneuver based on the genetic algorithm optimization results under the current situation. The UAV has not yet learned the corresponding strategy and is unable to defeat the target in the early stages of training. The cumulative reward curve fluctuates sharply in the negative area within a certain number of iterations. As the UAV continues to learn the maneuvering strategy, it gradually becomes able to defeat the target. Compared with the three other scenarios, the convergence speed of the cumulative reward curve is slower, and the curve fluctuates sharply in scenario 4.

5 Conclusion

In this paper, a TD3–LSTM reinforcement learning-based intelligent algorithm is developed to address the maneuver decision-making problem of a UAV under uncertain information. To ensure the validity, robustness, and efficiency of maneuver decision-making in UAV aerial confrontation scenarios, four simulation experiments are considered in this manuscript: target in straight flight, target in circle flight, target in expert rule-based maneuvering strategy, and genetic algorithm-based strategy. The simulation results demonstrate that regardless of the maneuvering strategy the target adopts, the UAV can comprehend the environmental situation, execute appropriate maneuvering actions, and ultimately emerge victorious in an aerial confrontation. For future work, the implementation of multi-drone collaborative adversarial maneuvering decisions based on higher-fidelity models warrants consideration. Furthermore, achieving efficient sim-to-real policy transfer through transfer learning presents a significant research priority.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Author contributions

TZ: Conceptualization, Data curation, Formal Analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Software, Supervision, Validation, Visualization, Writing – original draft, Writing – review and editing. ZL: Data curation, Validation, Writing – review and editing. WJ: Data curation, Validation, Writing – review and editing. ZH: Conceptualization, Writing – review and editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This work was supported by grants from the National Natural Science Foundation of China (grant. 62203217), the Jiangsu Basic Research Program Natural Science Foundation (grant. BK20220885), and the Collaborative Innovation Fund Project of the Institute of Optoelectronics (grant. 613CXJJ2025TS04), China Postdoctoral Science Foundation under Grant Number GZC20242230.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

The handling editor ZJ is currently organizing a Research Topic with the author ZH.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Cheng, Y., Chen, L., Chen, C., and Wang, X. (2021). Off-policy deep reinforcement learning based on steffensen value iteration. IEEE Trans. Cognitive Dev. Syst. 13, 1023–1032. doi:10.1109/tcds.2020.3034452

CrossRef Full Text | Google Scholar

Duan, H., Lei, Y., Xia, J., Deng, Y., and Shi, Y. (2023). Autonomous maneuver decision for unmanned aerial vehicle via improved pigeon-inspired optimization. IEEE Trans. Aerosp. Electron. Syst. 59, 3156–3170. doi:10.1109/taes.2022.3221691

CrossRef Full Text | Google Scholar

Duan, H., Li, P., and Yu, Y. (2015). A predator-prey particle swarm optimization approach to multiple ucav air combat modeled by dynamic game theory. IEEE/CAA J. Automatica Sinica 2, 11–18. doi:10.1109/jas.2015.7032901

CrossRef Full Text | Google Scholar

Duan, J., Guan, Y., Li, S., Ren, Y., Sun, Q., and Cheng, B. (2022). Distributional soft actor-critic: off-policy reinforcement learning for addressing value estimation errors. IEEE Trans. Neural Netw. Learn. Syst. 33, 6584–6598. doi:10.1109/tnnls.2021.3082568

PubMed Abstract | CrossRef Full Text | Google Scholar

Ergen, T., and Kozat, S. S. (2018). Online training of lstm networks in distributed systems for variable length data sequences. IEEE Trans. Neural Netw. Learn. Syst. 29, 5159–5165. doi:10.1109/tnnls.2017.2770179

PubMed Abstract | CrossRef Full Text | Google Scholar

Esfahani, P., and Kuhn, D. (2018). Data-driven distributionally robust optimization using the wasserstein metric: performance guarantees and tractable reformulations. Math. Program. 171, 115–166. doi:10.1007/s10107-017-1172-1

CrossRef Full Text | Google Scholar

Fujimoto, S., Herke, H. V., and Meger, D. (2018). “Addressing function approximation error in actor-critic methods,” in International conference on machine learning (Sweden (Stockholm)), 1587–1596.

Google Scholar

Guo, H., Chen, M., Jiang, Y., and Lungu, M. (2023). Distributed adaptive human-in-the-loop event-triggered formation control for quavs with quantized communication. IEEE Trans. Industrial Inf. 19, 7572–7582. doi:10.1109/tii.2022.3211508

CrossRef Full Text | Google Scholar

He, J., Ding, Y., and Gao, Z. (2020). A stealthy engagement maneuvering strategy of uav based on double deep q network. Electron. Opt. Control 27, 52–57.

Google Scholar

Hong, D., Lee, Y. C. S., Cho, Y. H., Baek, D., Kim, J., and Chang, N. (2021). Energy-efficient online path planning of multiple drones using reinforcement learning. IEEE Trans. Veh. Technol. 70, 9725–9740. doi:10.1109/tvt.2021.3102589

CrossRef Full Text | Google Scholar

Huang, C., Dong, K., Huang, H., and Tang, S. (2018). Autonomous air combat maneuver decision using bayesian inference and moving horizon optimization. J. Syst. Eng. Electron. 29, 86–97. doi:10.21629/jsee.2018.01.09

CrossRef Full Text | Google Scholar

Kant, R., Saini, P., and Kumari, J. (2023). Long short-term memory auto-encoder-based position prediction model for fixed-wing uav during communication failure. IEEE Trans. Artif. Intell. 4, 173–181. doi:10.1109/tai.2022.3153763

CrossRef Full Text | Google Scholar

Li, B., Bai, S., and Meng, B. (2022). Autonomous air combat decision-making algorithm of uavs based on sac algorithm. Command Control Simul. 44, 24–30.

Google Scholar

Li, K., Zhang, K., Zhang, Z., Liu, Z., Hua, S., and He, J. (2021). A uav maneuver decision-making algorithm for autonomous airdrop based on deep reinforcement learning. Sensors 21, 2233. doi:10.3390/s21062233

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, Q., Wang, F., Yang, W., and Liu, Z. (2022). Air combat maneuver strategy algorithm based on two-layer game decision-making and distributed double game trees mcts under uncertain information. Electronics 11, 2608, doi:10.3390/electronics11162608

CrossRef Full Text | Google Scholar

Liu, F., Xie, F., and Meng, G. (2016). A uav air-combat decision expert system based on receding horizon control. J. Beijing Univ. Aeronautics Astronautics 41, 1994–1999.

Google Scholar

Luo, C., Wang, J., Huang, H., and Wang, P. (2022). Integrated guidance and control based air-to-air autonomous attack occupation of ucav. Math. Problems Eng. 2016, 1–18. doi:10.1155/2016/6431264

CrossRef Full Text | Google Scholar

Park, H., Lee, B., Tahk, M., and Yoo, D. W. (2016). Differential game based air combat maneuver generation using scoring function matrix. Int. J. Aeronautical Space Sci. 17, 204–213. doi:10.5139/ijass.2016.17.2.204

CrossRef Full Text | Google Scholar

Shin, H., Lee, J., Kim, H., and Hyunchul Shim, D. (2018). An autonomous aerial combat framework for two-on-two engagements based on basic fighter maneuvers. Aerosp. Sci. Technol. 72, 305–315. doi:10.1016/j.ast.2017.11.014

CrossRef Full Text | Google Scholar

Tan, M., Ding, D., and Xie, L. (2022). Ucav escape maneuvering decision based on fuzzy expert system and ide algorithm. Syst. Eng. Electron. 44, 1984–1993.

Google Scholar

Tu, Z., Fei, F., and Deng, X. (2021). Bio-inspired rapid escape and tight body flip on an at-scale flapping wing hummingbird robot via reinforcement learning. IEEE Trans. Robotics 37, 1742–1751. doi:10.1109/tro.2021.3064882

CrossRef Full Text | Google Scholar

Virtanen, K., Karelahti, J., and Raivio, T. (2006). Modeling air combat by a moving horizon influence diagram game. J. Guid. Control Dyn. 29, 1080–1091. doi:10.2514/1.17168

CrossRef Full Text | Google Scholar

Wang, M., Wang, X., Yue, T., and Liu, H. (2020). Influence of unmanned combat aerial vehicle agility on short-range aerial combat effectiveness. Aerosp. Sci. Technol. 96, 105534–12. doi:10.1016/j.ast.2019.105534

CrossRef Full Text | Google Scholar

Wang, Y., Zhang, W., and Zhou, R. (2022). Research on ucav maneuvering decision method based on heuristic reinforcement learning. Comput. Intell. Neurosci. 3, 1477078.

Google Scholar

Yang, Q., Zhang, J., and Shi, G. (2019). Modeling of uav path planning based on imm under pomdp framework. J. Syst. Eng. Electron. 30, 545–554. doi:10.21629/jsee.2019.03.12

CrossRef Full Text | Google Scholar

Zhang, H., and Huang, C. (2020). Maneuver decision-making of deep learning for ucav thorough azimuth angles. IEEE Access 8, 12976–12987. doi:10.1109/access.2020.2966237

CrossRef Full Text | Google Scholar

Zhang, H., Zhou, H., Wei, Y., and Huang, C. (2022). Autonomous maneuver decision-making method based on reinforcement learning and Monte Carlo tree search. Front. Neurorobotics 16, 996412. doi:10.3389/fnbot.2022.996412

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, X., Liu, G., Yang, C., and Wu, J. (2018). Research on air confrontation maneuver decision-making method based on reinforcement learning. Electronics 7, 279. doi:10.3390/electronics7110279

CrossRef Full Text | Google Scholar

Zhou, T., Chen, M., Wang, Y., He, J., and Yang, C. (2020a). Information entropy-based intention prediction of aerial targets under uncertain and incomplete information. Entropy 22, 279. doi:10.3390/e22030279

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhou, T., Chen, M., and Zou, J. (2020b). Reinforcement learning based data fusion method for multi-sensors. IEEE/CAA J. Automatica Sinica 7, 1489–1497. doi:10.1109/jas.2020.1003180

CrossRef Full Text | Google Scholar

Keywords: unmanned aerial vehicles, maneuver decision-making, reinforcement learning, twin delayed deep deterministic policy gradient, long short-time memory

Citation: Zhou T, Liu Z, Jin  W and Han Z (2025) Intelligent maneuver decision-making for UAVs using the TD3–LSTM reinforcement learning algorithm under uncertain information. Front. Robot. AI 12:1645927. doi: 10.3389/frobt.2025.1645927

Received: 12 June 2025; Accepted: 02 July 2025;
Published: 01 August 2025.

Edited by:

Zhehao Jin, Nanyang Technological University, Singapore

Reviewed by:

Xiang Wu, Zhejiang University of Technology, China
Li Yankai, Xi’an University of Technology, China
Siqi Wang, Zhengzhou Research Institute of Harbin Institute of Technology, China

Copyright © 2025 Zhou, Liu, Jin  and Han. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Tongle Zhou, emhvdXRvbmdsZUBudWFhLmVkdS5jbg==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.