ORIGINAL RESEARCH article

Front. Robot. AI, 14 February 2025

Sec. Robot Learning and Evolution

Volume 11 - 2024 | https://doi.org/10.3389/frobt.2024.1491907

Adaptive formation learning control for cooperative AUVs under complete uncertainty

  • 1. Department of Mechanical, Industrial and Systems Engineering, University of Rhode Island, Kingston, RI, United States

  • 2. Graduate School of Oceanography, University of Rhode Island, Kingston, RI, United States

  • 3. Department of Electrical, Computer, and Biomedical Engineering, University of Rhode Island, Kingston, RI, United States

Abstract

Introduction:

This paper addresses the critical need for adaptive formation control in Autonomous Underwater Vehicles (AUVs) without requiring knowledge of system dynamics or environmental data. Current methods, often assuming partial knowledge like known mass matrices, limit adaptability in varied settings.

Methods:

We proposed two-layer framework treats all system dynamics, including the mass matrix, as entirely unknown, achieving configuration-agnostic control applicable to multiple underwater scenarios. The first layer features a cooperative estimator for inter-agent communication independent of global data, while the second employs a decentralized deterministic learning (DDL) controller using local feedback for precise trajectory control. The framework's radial basis function neural networks (RBFNN) store dynamic information, eliminating the need for relearning after system restarts.

Results:

This robust approach addresses uncertainties from unknown parametric values and unmodeled interactions internally, as well as external disturbances such as varying water currents and pressures, enhancing adaptability across diverse environments.

Discussion:

Comprehensive and rigorous mathematical proofs are provided to confirm the stability of the proposed controller, while simulation results validate each agent’s control accuracy and signal boundedness, confirming the framework’s stability and resilience in complex scenarios.

1 Introduction

Robotics and autonomous systems have a wide range of applications, spanning from manufacturing and surgical procedures to exploration in challenging environments (Ghafoori et al., 2024; Jandaghi et al., 2023). However, controlling robots in such settings, especially in space and underwater, presents significant difficulties due to unpredictable dynamics. In the context of underwater exploration, AUVs have become essential tools, offering cost-effective, reliable, and versatile solutions for adapting to dynamic conditions. Effective use of AUVs is critical for unlocking the mysteries of marine environments, making advancements in their control and operation essential. As the demand for efficient underwater exploration increases and the complexity of tasks assigned to AUVs grows, there is a pressing need to enhance their operational capabilities. This includes developing sophisticated formation control strategies that allow multiple AUVs to operate in coordination, drawing inspiration from natural behaviors observed in fish schools and bird flocks (Zhou et al., 2023; Yang et al., 2021). By leveraging multi-agent systems, AUVs can work in coordinated groups, enhancing efficiency, stability, and coverage while navigating dynamic and complex underwater environments. These strategies are essential for ensuring precise operations in varied underwater tasks, ranging from pipeline inspections and seafloor mapping to environmental monitoring (Yan et al., 2023).

Despite challenges from intricate nonlinear dynamics, complex interactions among AUVs, and the uncertain dynamic nature of underwater environments, effective multi-AUV formation control is increasingly critical in modern ocean industries (Yan et al., 2018; Hou and Cheah, 2009). Historically, formation control research has predominantly utilized the behavioral approach (Balch and Arkin, 1998; Lawton, 2000), which divides the overall control design into subproblems, with each vehicle’s action determined by a weighted average of solutions, though selecting appropriate weighting parameters can be challenging. The leader-following approach (Cui et al., 2010; Rout and Subudhi, 2016) designates one vehicle as the leader while others follow, maintaining predefined geometric relationships, and controlling formation behavior by designing specific motions for the leader. Alternatively, the virtual structure approach (Millán et al., 2013).

Despite advancements in formation control and path planning for multi-AUV systems, challenges such as environmental disturbances, complex underwater dynamics, and communication limitations continue to pose difficulties (Hadi et al., 2021). To address these challenges, there is a critical need for controllers that are independent of both robot dynamics and environmental disturbances. Developing such controllers would enhance formation control by allowing for decentralized application, which increases flexibility in formation structures and improves robustness against communication constraints. Addressing these gaps is essential for advancing the capabilities and reliability of multi-AUV systems. On the other hand, communication constraints in underwater environments make decentralized control with a virtual leader-following topology ideal for AUVs, enabling coordination using local information despite communication delays or interruptions (Yan et al., 2023).

Reinforcement learning (RL) has also been extensively applied in robotic control (Christen et al., 2021; Cao et al., 2022). RL approaches, such as deep reinforcement learning (DRL), offer advantages in learning complex, non-linear control policies directly from data. However, RL methods generally lack the ability to provide mathematical stability proofs and guarantees for the controller’s behavior, making it challenging to ensure safety and reliability, especially in critical applications. Besides, while Zhang et al. (2018) developed various direct neural adaptive laws that lead to increased oscillations with higher adaptation gains, indirect neural adaptive laws using prediction error methods were proposed to mitigate this issue, though they could not guarantee parameter convergence. However, NN-based learning control methods, such as those utilizing adaptive neural networks or deterministic learning frameworks Jandaghi et al. (2024), can incorporate stability analysis and provide rigorous mathematical proofs for parameter convergence. These methods enable researchers to establish theoretical guarantees for the stability and robustness of the controller, which is essential for deploying controllers in real-world applications where safety and reliability are critical. Most recently, Tutsoy et al. (2024) proposed an optimization-based approach for path planning in Unmanned Air Vehicles (UAVs) with actuator failures using particle swarm optimization and genetic algorithms. Their method focuses on minimizing both time and distance by optimizing predefined cost functions through heuristic methods, while incorporating system constraints such as actuator limits, kinematic, and dynamic constraints, as well as parametric uncertainties.

Despite extensive literature in the field, to the best of our knowledge, existing researches assume homogeneous dynamics and certain system parameters for all AUV agents, which is unrealistic in unpredictable underwater environments. Factors such as buoyancy, drag, and varying water viscosity significantly alter system dynamics and behavior. Additionally, AUVs may change shape during tasks like underwater sampling or when equipped with robotic arms, further complicating control. Typically, designing multi-AUV formation control involves planning desired formation paths and developing tracking controllers for each AUV. However, accurately tracking these paths is challenging due to the complex nonlinear dynamics of AUVs, especially when precise models are unavailable. Implementing a fully distributed and decentralized formation control system is also difficult, as centralized control designs become exceedingly complex with larger AUV groups. To address these challenges previous work, such as Yuan et al. (2017) and Dong et al. (2019), developed adaptive learning controllers that relied on the assumption of a known mass matrix, which is not practical in real-world applications. These controllers relied on known system parameters that can fail due to varying internal forces caused by varying external environmental conditions. The solution is to develop environment-independent controllers that do not rely on any specific system dynamical parameters.

The framework’s control architecture is ingeniously divided into a first-layer Cooperative Estimator Observer and a lower-layer Decentralized Deterministic Learning (DDL) Controller. The first-layer observer is pivotal in enhancing inter-agent communication by sharing crucial system estimates, operating independently of any global information. Concurrently, the second-layer DDL controller utilizes local feedback to finely adjust each AUV’s trajectory, ensuring resilient operation under dynamic conditions heavily influenced by hydrodynamic forces and torques by considering system uncertainty completely unknown. This dual-layer setup not only facilitates acute adaptation to uncertain AUV dynamics but also leverages RBFNN for precise local learning and effective knowledge storage. Such capabilities enable AUVs to efficiently reapply previously learned dynamics after the system restarts. This tow-layer framework achieves a significant advancement by considering all system dynamics parameters as unknown, enabling a universal application across all AUVs, regardless of their operating environments. This universality is crucial for adapting to environmental variations such as water flow, which increases the AUV’s effective mass via the added mass phenomenon and affects the vehicle’s inertia. Additionally, buoyancy forces that vary with depth, along with hydrodynamic forces and torques, stemming from water flow variations, the AUV’s unique shape, its appendages, and drag forces due to water viscosity, significantly impact the damping matrix in the AUV’s dynamics. This framework not only improves operational efficiency but also significantly advances the field of autonomous underwater vehicle control by laying a robust foundation for future enhancements in distributed adaptive control systems and fostering enhanced collaborative intelligence among multi-agent networks in marine environments. Extensive simulations have underscored the effectiveness of the framework, demonstrating its potential to elevate the adaptability and resilience of AUV systems under the most demanding conditions. In summary, the contribution of this paper is as follows:

  • • The universal controller works in any environment and condition, such as currents or depth.

  • • Each AUV controller operates independently.

  • • The controller functions without needing information about the robot’s dynamic parameters, like mass, damping, or inertia. Each AUV can also have different dynamic parameters.

  • • The system learns the dynamics once and reuses the pre-trained weights, avoiding the need for retraining.

  • • The use of localized RBFNN reduces real-time computational demands.

  • • Providing rigorous stability analysis of the controller while providing mathematical proofs to ensure and guarantee the reliability of the controller.

The rest of the paper is organized as follows: Section 2 provides an initial overview of graph theory, RBFNN, and the problem statement. The design of the distributed cooperative estimator and the decentralized deterministic learning controller are discussed in Section 3. The formation adaptive control and formation control using pre-learned dynamics are explored in Section 4 and Section 5, respectively. Simulation studies are presented in Section 6, and Section 7 concludes the paper.

2 Preliminaries and problem statement

2.1 Notation and graph theory

Denoting the set of real numbers as , we define as the set of real matrices, and as the set of real vectors. The identity matrix is symbolized as . The vector with all elements being 1 in an -dimensional space is represented as . The sets and stand for real symmetric and positive definite matrices, respectively. A block diagonal matrix with matrices on its main diagonal is denoted by . signifies the Kronecker product of matrices and . For a matrix , is the vectorization of by stacking its columns on top of each other. For a series of column vectors , represents a column vector formed by stacking them together. Given two integers and with , . For a vector , its norm is defined as . For a square matrix , denotes its -th eigenvalue, while and represent its minimum and maximum eigenvalues, respectively.

A directed graph comprises nodes in the set and edges in . An edge from node to node is represented as , with as the parent node and as the child node. Node is also termed a neighbor of node . is considered as the subset of consisting of the neighbors of node . A sequence of edges in , , is called a path from node to node . Node is reachable from node . A directed tree is a graph where each node, except for a root node, has exactly one parent. The root node is reachable from all other nodes. A directed graph contains a directed spanning tree if at least one node can reach all other nodes. The weighted adjacency matrix of is a non-negative matrix , where and . The Laplacian of is denoted as , where and if . It is established that has at least one eigenvalue at the origin, and all nonzero eigenvalues of have positive real parts. From Ren and Beard (2005), has one zero eigenvalue and remaining eigenvalues with positive real parts if and only if has a directed spanning tree.

2.2 Radial basis function neural networks (RBFNN)

The RBFNN Networks can be described as , where and as input and weight vectors respectively (Park and Sandberg, 1991). indicates the number of NN nodes, with is a radial basis function, and is distinct points in the state space. The Gaussian function is generally used for radial basis function, where is the center and is the width of the receptive field. The Gaussian function categorized by localized radial basis function in the sense that as . Moreover, for any bounded trajectory within the compact set , can be approximated using a limited number of neurons located in a local region along the trajectory . denotes the indices of active RBFNN nodes where , based on the state . is the approximation error, with , , , , and the integers are defined by ( is a small positive constant) for some . This holds if for . The following lemma regarding the persistent excitation (PE) condition for RBFNN is recalled from Wang and Hill (2018).

Lemma 1Consider any continuous recurrent trajectory1. remains in a bounded compact set . Then for an RBFNN with centers placed on a regular lattice (large enough to cover the compact set ), the regressor subvector consisting of RBFNN with centers located in a small neighborhood of is persistently exciting.

2.3 Problem statement

A multi-agent system comprising AUVs with heterogeneous nonlinear uncertain dynamics is considered. The dynamics of each AUV can be expressed as Fossen (1999):

In this study, the subscript identifies each AUV within the multi-agent system. For every , the vector represents the -th AUV’s position coordinates and heading in the global coordinate frame, while denotes its linear velocities and angular rate of heading relative to a body-fixed frame. The positive definite inertial matrix , Coriolis and centripetal matrix , and damping matrix characterize the AUV’s dynamic response to motion. The vector accounts for the restoring forces and moments due to gravity and buoyancy. The term , with , describes the vector of generalized deterministic unmodeled uncertain dynamics for each AUV.

The vector represents the control inputs for each AUV. The associated rotation matrix is given by:

Unlike previous work Yuan et al. (2017), which assumed known values for the AUV’s inertia and rotation matrices, this study considers all matrix coefficients, including , , , and , as well as the inertia matrix, as completely unknown. The adaptive estimation process inherently addresses the effects of external forces and disturbances on system dynamics. This eliminates the need for explicit parameter estimation of these forces, as disturbances like water flow, varying currents, or depth variations are directly incorporated into the control input through adaptive estimation. This makes the controller universally applicable to any AUV, regardless of its design, weight, or environmental conditions by addressing both internal and external dynamic variation at the same time.

Internally, it handles unknown parameters such as mass and damping coefficients, as well as unmodeled nonlinear interactions and couplings. Externally, it accounts for unpredictable disturbances, including fluctuating water currents, depth-dependent pressures, and changes in hydrodynamic forces.

By avoiding reliance on predefined models, the proposed approach is robust and adaptable to diverse mission scenarios and unexpected environmental changes, ensuring reliable performance even in highly uncertain conditions.

In the context of leader-following formation tracking control, the following virtual leader dynamics generates the tracking reference signals:with “0” marking the leader node, the leader state with and , is a constant matrix available only to the leader’s neighboring AUV agents.

Considering the system dynamics of multiple AUVs (Equation 1) along with the leader dynamics (Equation 2), we establish a non-negative matrix , where such that for each , if and only if agent has access to the reference signals and . All remaining elements of are arbitrary non-negative values, such that for all . Correspondingly, we establish as a directed graph derived from , where designates node 0 as the leader, and the remaining nodes correspond to the AUV agents. We proceed under the following assumptions:

Assumption 1All the eigenvalues of in the leader’s dynamics (Equation 2) are located on the imaginary axis.

Assumption 2The directed graph contains a directed spanning tree with the node 0 as its root.

Assumption 1 is crucial for ensuring that the leader dynamics produce stable, meaningful reference trajectories for formation control. It ensures that all states of the leader, represented by , remain within the bounds of a compact set for all . The trajectory of the system, starting from and denoted by , generates periodic signal. This periodicity is essential for maintaining the Persistent Excitation (PE) condition, which is pivotal for achieving parameter convergence in Distributed Adaptive (DA) control systems. Modifications to the eigenvalue constraints on mentioned in Assumption 1 may be considered when focusing primarily on formation tracking control performance, as discussed later.

Additionally, Assumption 2 reveals key insights into the structure of the Laplacian matrix of the network graph . Let be an non-negative diagonal matrix where each -th diagonal element is for . The Laplacian is formulated as:where if and otherwise. This results in since . As cited in Su and Huang (2011), all nonzero eigenvalues of , if present, exhibit positive real parts, confirming as nonsingular under Assumption 2.

Problem 1

In the context of a multi-AUV system (

Equation 1

) integrated with virtual leader dynamics (

Equation 2

) and operating within a directed network topology

, the aim is to develop a distributed NN learning control protocol that leverages only local information. The specific goals are twofold:

  • 1) Formation Control: Each of the AUV agents will adhere to a predetermined formation pattern relative to the leader, maintaining a specified distance from the leader’s position .

  • 2) Decentralized Learning: The nonlinear uncertain dynamics of each AUV will be identified and learned autonomously during the formation control process. The insights gained from this learning process will be utilized to enhance the stability and performance of the formation control system.

Remark 1The leader dynamics described in Equation 2 are designed as a neutrally stable LTI system. This design choice facilitates the generation of sinusoidal reference trajectories at various frequencies which is essential for effective formation tracking control. This approach to leader dynamics is prevalent in the literature on multiagent leader-following distributed control systems like Yuan (2017) and Jandaghi et al. (2024).

Remark 2It is important to emphasize that the formulation assumes formation control is required only within the horizontal plane, suitable for AUVs operating at a constant depth, and that the vertical dynamics of the 6 degrees of freedom (DOF) AUV system, as detailed in Prestero (2001), are entirely decoupled from the horizontal dynamics.

As shown in Figure 1, a two-layer hierarchical design approach is proposed to address the aforementioned challenges. The first layer, the Cooperative Estimator, enables information exchange among neighboring agents. The second layer, known as the Decentralized Deterministic Learning (DDL) controller, processes only local data from each individual AUV. The development and formulation of the first layer are discussed in detail in Section 3.1, while the DDL control strategy, along with its corresponding controller design and analysis, is provided in Section 3.2.

FIGURE 1

3 Two-layer distributed controller architecture

3.1 First layer: cooperative estimator

In the context of leader-following formation control, not all AUV agents may have direct access to the leader’s information, including tracking reference signals and the system matrix . This necessitates collaborative interactions among the AUV agents to estimate the leader’s information effectively. Drawing on principles from multiagent consensus and graph theories Ren and Beard (2008), we propose to develop a distributed adaptive observer for the AUV systems as:

The observer states for each -th AUV, denoted by , aim to estimate the leader’s state, . As , these estimates are expected to converge, such that approaches and approaches , representing the leader’s position and velocity, respectively. Equation 3 accounts for the communication graph by including the adjacency matrix information through the term . Note that represents an estimate computed by agent i of the leader’s matrix dynamics which is also not available to the agents. Therefore, each agent estimates such matrix using a cooperative adaptation law:which borrowed from Ren and Beard (2008) as well. The constants and are all positive numbers and are subject to design.

Remark 3Each AUV agent in the group is equipped with an observer configured as specified in Equations 3, 4, comprising two state variables, and . For each , estimates the virtual leader’s state , while estimates the leader’s system matrix . The real-time data necessary for operating the -th observer includes: (1) the estimated state and matrix , obtained from the -th AUV itself, and (2) the estimated states and matrices for all , obtains from the -th AUV’s neighbors. Note that in Equations 3, 4, if , then , indicating that the -th observer does not utilize information from the -th AUV agent. This configuration ensures that the proposed distributed observer can be implemented in each local AUV agent using only locally estimated data from the agent itself and its immediate neighbors, without the need for global information such as the size of the AUV group or the network interconnection topology.

To verify the convergence properties, we need to compute the error dynamics. Now we define the estimation error for the state and the system matrix for agent as and , and then we derive the error dynamics:

Define the collective error states and adaptation matrices: for the state errors, for the adaptive parameter errors, representing the block diagonal of adaptive parameters, and for the diagonal matrices of design constants. With these definitions, the network-wide error dynamics can be expressed as:

Theorem 1

Consider the error system Equation 5. Under Assumptions 1, 2, and given that , it follows that for all and for any initial conditions , the error dynamics of the adaptive parameters and the states will converge to zero exponentially. Specifically, and .

This convergence is facilitated by the independent adaptation of each agent’s parameters within their respective error dynamics, represented by the block diagonal structure of and control gains and . These matrices ensure that each agent’s parameter updates are governed by local interactions and error feedback, consistent with the decentralized control framework.

Proof: We begin by examining the estimation error dynamics for as presented in Equation 5. This can be rewritten in the vector form:

Under Assumption 2, all eigenvalues of possess positive real parts according to Su and Huang (2011). Consequently, for any positive , the matrix is guaranteed to be Hurwitz, Which implies the exponential stability of system (Equation 6). Hence, it follows that exponentially, leading to exponentially for all . Now, we analyze the error dynamics for in Equation 5. Based on the previous discussions, We have exponentially, and the term will similarly decay to zero exponentially. Based on Cai et al. (2015), if the system defined byis exponentially stable, then exponentially. With Assumption 1, knowing that all eigenvalues of have zero real parts, and since as nonsingular with all eigenvalues in the right-half plane, system (Equation 7) is exponentially stable for any positive . Consequently, this ensures that , i.e., exponentially for all .

Now, each individual agent can accurately estimate both the state and the system matrix of the leader through cooperative observer estimation Equations 3, 4. This information will be utilized in the DDL controller design for each agent’s second layer, which will be discussed in the following subsection.

3.2 Second layer: decentralized deterministic learning controller

To fulfill the overall formation learning control objectives, in this section, we develop the DDL control law for the multi-AUV system defined in Equation 1. We use to denote the desired distance between the position of the -th AUV agent and the virtual leader’s position . Then, the formation control problem is framed as a position tracking control task, where each local AUV agent’s position is required to track the reference signal . Besides, due to the inaccessibility of the leader’s state information for all AUV agents, the tracking reference signal is employed instead of the reference signal . As established in Theorem 1, is autonomously generated by each local agent and will exponentially converge to . This ensures that the DDL controller is feasible and the formation control objectives are achievable for all using .

To design the DDL control law that addresses the formation tracking control and the precise learning of the AUVs’ complete nonlinear uncertain dynamics at the same time, we will integrate renowned backstepping adaptive control design method outlined in Krstic et al. (1995) along with techniques from Wang and Hill (2018) and Yuan et al. (2017) for deterministic learning using RBFNN. Specifically, for the -th AUV agent described in system (Equation 1), we define the position tracking error as for all . Considering for all , we proceed to:

To frame the problem in a more tractable way, we assume as a virtual control input and as a desired virtual control input in our control strategy design, and by implementing them in the above system we have:

A positive definite gain matrix is used for tuning the performance. Substituting into Equation 8 yields:

Now we derive the first derivatives of the virtual control input and the desired control input as follows:

As previously discussed, unlike earlier research that only identified the matrix coefficients , , , and as unknown system nonlinearities while assuming the mass matrix to be known, this work advances significantly by also considering as unknown. Consequently, all system dynamic parameters are treated as completely unknown, making the controller fully independent of the robot’s configuration such as its dimensions, mass, or any appendages and the uncertain environmental conditions it encounters, like depth, water flow, and viscosity. This independence is critical as it ensures that the controller does not rely on predefined assumptions about the dynamics, aligning with the main goal of this research. To address these challenges, we define a unique nonlinear function that encapsulates all nonlinear uncertainties as follows:where and , with being a bounded compact set. We then employ the following RBFNN to approximate the model dynamics in (Equation 11) expressed by nonlinear functions with for all and as follows:where is the ideal constant NN weights, and is the approximation error for all and , which satisfies . This error can be made arbitrarily small given a sufficient number of neurons in the network. A self-adaptation law is designed to estimate the unknown online. We aim to estimate with by constructing the DDL feedback control law as follows: is a feedback gain matrix that can be tuned to achieve the desired performance. To approximate the unknown nonlinear function vector in (Equation 11) along the trajectory within the compact set , we use:

Then, from Equations 1, 13 we have:

By subtracting from both sides and considering Equations 9, 11, we define , leading to:

For updating online, a robust self-adaptation law is constructed using the -modification technique Ioannou and Sun (1996) as follows:where , , and are free parameters to be designed for all and . Integrating Equations 9, 13, 14 yields the following closed-loop system:where, for all and , , and .

Remark 4Unlike the first-layer DA observer design, the second-layer control law is fully decentralized for each local agent. It utilizes only the local agent’s information for feedback control, including , , and , without involving any information exchange among neighboring AUVs.

Theorem 2

Consider the local closed-loop system (Equation 15). For each , if there exists a sufficiently large compact set such that for all , then for any bounded initial conditions, we have: 1) All signals in the closed-loop system remain uniformly ultimately bounded (UUB). 2) The position tracking error converges exponentially to a small neighborhood around zero in finite time by choosing the design parameters with sufficiently large and , and sufficiently small for all and .

  • Proof: 1) Consider the following Lyapunov function candidate for the closed-loop system (Equation 15):

Evaluating the derivative of along the trajectory of Equation 15 for all yields:

Choose such that . Using the completion of squares, we have:where . Then, we obtain:

It follows that is negative definite whenever:

For all

,

. This leads to the Uniformly Ultimately Bounded (UUF) behavior of the signals

,

, and

for all

and

. As a result, it can be easily verified that since

with

bounded (according to

Theorem 1

and

Assumption 1

),

is bounded for all

. Similarly, the boundedness of

can be confirmed by the fact that

in

Equation 9

is bounded. In addition,

is also bounded for all

and

because of the boundedness of

and

. Moreover, in light of

Equation 10

,

is bounded as all the terms on the right-hand side of

Equation 10

are bounded. This leads to the boundedness of the control signal

in

Equation 13

since the Gaussian function vector

is guaranteed to be bounded for any

. As such, all the signals in the closed-loop system remain UUB, which completes the proof of the first part.

  • 2) For the second part, it will be shown that will converge arbitrarily close to in some finite time for all . To this end, we consider the following Lyapunov function candidate for the dynamics of and in Equation 15:

The derivative of is:

Similar to the proof of part one, we let with . According to Wang and Hill (2018), the Gaussian RBFNN regressor is bounded by for any and for all with some positive number . Through completion of squares, we have:

Also . This leads to:where and , . Solving the inequality Equation 17 yields:which together with Equation 16 implies that:also

Consequently, it is straightforward that given , there exists a finite time for all such that for all , both and satisfy and , where can be made arbitrarily small by choosing sufficiently large and for all . This ends the proof.

By integrating the outcomes of Theorems 1, 2, the following theorem is established, which can be presented without additional proof:

Theorem 3

By Considering the multi-AUV system (Equation 1) and the virtual leader dynamics (Equation 2) with the network communication topology and under Assumptions 1 and 2, the objective 1 of Problem 1 (i.e., converges to exponentially for all ) can be achieved by using the cooperative observer Equations 3, 4 and the DDL control law Equations 13, 14 with all the design parameters satisfying the requirements in Theorems 1 and 2, respectively.

Remark 5With the proposed two-layer formation learning control architecture, inter-agent information exchange occurs solely in the first-layer DA observation. Only the observer’s estimated information, and not the physical plant state information, needs to be shared among neighboring agents. Additionally, since no global information is required for the design of each local AUV control system, the proposed formation learning control protocol can be designed and implemented in a fully distributed manner.

Remark 6It is important to note that the eigenvalue constraints on in Assumption 1 are not needed for cooperative observer estimation (as detailed in the Section 3 or for achieving formation tracking control performance (as discussed in this section). This indicates that formation tracking control can be attained for general reference trajectories, including both periodic paths and straight lines, provided they are bounded. However, these constraints will become necessary in the next section to ensure the accurate learning capability of the proposed method.

4 Accurate learning from formation control

It is necessary to demonstrate the convergence of the RBFNN weights in Equations 13, 14 to their optimal values for accurate learning and identification. The main result of this section is summarized in the following theorem.

Theorem 4

Consider the local closed-loop system (Equation 15) with Assumptions 1, 2. For each , if there exists a sufficiently large compact set such that for all , then for any bounded initial conditions and , the local estimated neural weights converge to small neighborhoods of their optimal values along the periodic reference tracking orbit (denoting the orbit of the NN input signal starting from time ). This leads to locally accurate approximations of the nonlinear uncertain dynamics in Equation 11 being obtained by , as well as by , where . Also, denotes the subset of neurons (or nodes) that are active when the system state is within a specific neighborhood of the state space.where represents a time segment after the transient process.

Proof: From Theorem 3, we have shown that for all , will closely track the periodic signal in finite time . In addition, (Equation 9) implies that will also closely track the signal since both and will converge to a small neighborhood around zero according to Theorem 2. Moreover, since will converge to according to Theorem 1, and is a bounded rotation matrix, will also be a periodic signal after finite time , because is periodic under Assumption 1. Consequently, since the RBFNN input becomes a periodic signal for all , the PE condition of some internal closed-loop signals, i.e., the RBFNN regression subvector , is satisfied according to Lemma 1. As mentioned in Section 2.2, represents the subset of RBFNN nodes and weights that are specifically utilized along the recurrent trajectory of the system state . This subset focuses on the active neural components required for approximating the system’s nonlinear dynamics locally, ensuring that the learning and adaptation processes are efficient and accurate within the compact region where the trajectory resides. It should be noted that the periodicity of leads to the PE of the regression subvector , but not necessarily the PE of the whole regression vector . Thus, we term this as a partial PE condition, and we will show the convergence of the associated local estimated neural weights , rather than .

Thus, to prove accurate convergence of local neural weights associated with the regression subvector under the satisfaction of the partial PE condition, we first rewrite the closed-loop dynamics of and along the periodic tracking orbit by using the localization property of the Gaussian RBFNN:where with and being the approximation error. Additionally, with subscripts and denoting the regions close to and far away from the periodic trajectory , respectively. According to Wang and Hill (2018), is small, and the NN local approximation error with is also a small number. Thus, the overall closed-loop adaptive learning system can be described by:andwherefor all . The exponential stability property of the nominal part of subsystem (Equation 19) has been well-studied in Wang and Hill (2018), Yuan and Wang (2011), and Yuan and Wang (2012), where it is stated that PE of will guarantee exponential convergence of for all and . Based on this, since , and can be made small by choosing sufficiently small for all , , both the state error signals and the local parameter error signals will converge exponentially to small neighborhoods of zero, with the sizes of the neighborhoods determined by the RBFNN ideal approximation error as in Equation 12 and . The convergence of implies that along the periodic trajectory , we havewhere for all , , due to the convergence of . The last equality is obtained according to the definition of (Equation 18) with being the corresponding subvector of along the periodic trajectory , and being an approximation error using . Apparently, after the transient process, we will have , , . Conversely, for the neurons whose centers are distant from the trajectory , the values of will be very small due to the localization property of Gaussian RBFNN. From the adaptation law (Equation 13) with , it can be observed that these small values of will only minimally activate the adaptation of the associated neural weights . As a result, both and , as well as and , will remain very small for all , along the periodic trajectory . This indicates that the entire RBFNN and can be used to accurately approximate the unknown function locally along the periodic trajectory , meaning that.with the approximation accuracy level of and for all , . This ends the proof.

Remark 7The key idea in the proof of Theorem 4 is inspired by Wang and Hill (2018). For more detailed analysis on the learning performance, including quantitative analysis on the learning accuracy levels and as well as the learning speed, please refer to Yuan and Wang (2011). Furthermore, the AUV nonlinear dynamics (Equation 9) to be identified do not contain any time-varying random disturbances. This is important to ensure accurate identification/learning performance under the deterministic learning framework. To understand the effects of time-varying external disturbances on deterministic learning performance, interested readers are referred to Yuan and Wang (2012) for more details.

Remark 8Based on Equation 18, to obtain the constant RBFNN weights for all , , one needs to implement the formation learning control law Equations 13, 14 first. Then, according to Theorem 4, after a finite-time transient process, the RBFNN weights will converge to constant steady-state values. Thus, one can select a time segment with for all to record and store the RBFNN weights for . Finally, based on these recorded data, can be calculated off-line using Equation 18.

Remark 9It is shown in Theorem 4 that locally accurate learning of each individual AUV’s nonlinear uncertain dynamics can be achieved using localized RBFNNs along the periodic trajectory . The learned knowledge can be further represented and stored in a time-invariant fashion using constant RBFNN, i.e., for all , . In contrast to many existing techniques (e.g., Peng et al., 2017; Peng et al., 2015), this is the first time, to the authors’ best knowledge, that locally accurate identification and knowledge representation using constant RBFNN are accomplished and rigorously analyzed for multi-AUV formation control under complete uncertain dynamics.

5 Formation control with pre-learned dynamics

In this section, we will further address objective 2 of Problem 1, which involves achieving formation control without readapting to the AUV’s nonlinear uncertain dynamics. To this end, consider the multiple AUV systems (Equation 1) and the virtual leader dynamics (Equation 2). We employ the estimator observer Equations 3, 4 to cooperatively estimate the leader’s state information. Instead of using the DDL feedback control law (Equation 13), and self-adaptation law (Equation 4), we introduce the following constant RBFNN controller, which does not require online adaptation of the NN weights:where is obtained from Equation 18. The term represents the locally accurate RBFNN approximation of the nonlinear uncertain function along the trajectory , and the associated constant neural weights are obtained from the formation learning control process as discussed in Remark 8.

Theorem 5

Consider the multi-AUV system (Equation 1) and the virtual leader dynamics (Equation 3) with the network communication topology . Under Assumptions 1, 2, the formation control performance (i.e., converges to exponentially with the same and defined in Theorem 3 for all ) can be achieved by using the DA observer Equations 3, 20 and the constant RBFNN control law (Equation 4) with the constant NN weights obtained from Equation 18.

Proof: The closed-loop system for each local AUV agent can be established by integrating the controller (Equation 20) with the AUV dynamics (Equation 1).where . Consider the Lyapunov function candidate , whose derivative along the closed-loop system described is given by:

Selecting where , we can utilize the method of completing squares to obtain:which implies that:where and . Using similar reasoning to that in the proof of Theorem 2, it is evident from the derived inequality that all signals within the closed-loop system remain bounded. Additionally, will converge to a small neighborhood around zero within a finite period. The magnitude of this neighborhood can be minimized by appropriately choosing large values for and across all . In line with Theorem 1, under Assumptions 1, 2, the implementation of the DA observer DA observer Equations 3, 4 facilitates the exponential convergence of towards . This conjunction of factors assures that rapidly aligns with , achieving the objectives set out for formation control.

Remark 10Building on the locally accurate learning outcomes discussed in Section 4, the newly developed distributed control protocol comprising Equations 3, 4, 20 facilitates stable formation control across a repeated formation pattern. Unlike the formation learning control approach outlined in Section 3.2, which involves Equations 3, 4 coupled with Equations 13, 14, the current method eliminates the need for online RBFNN adaptation for all AUV agents. This significantly reduces the computational demands, thereby enhancing the practicality of implementing the proposed distributed RBFNN formation control protocol. This innovation marks a significant advancement over many existing techniques in the field.

6 Simulation

We consider a multi-AUV heterogeneous system composed of 5 AUVs for the simulation. The dynamics of these AUVs are described in THE system (Equation 1). The system parameters for each AUV are specified as follows:where the mass and damping matrix components for each AUV are defined as:

According to the notations in Prestero (2001) and Skjetne et al. (2005) the coefficients are hydrodynamic parameters. For the associated system parameters are borrowed from Skjetne et al. (2005) (with slight modifications for different AUV agents) and simulation purposes and listed in Table 1. For all , we set and . Model uncertainties are given by:

TABLE 1

ParameterAUV 1AUV 2AUV 3AUV 4AUV 5
(kg)2325203035
(kg)1.82.01.52.22.5
(kg)−2.0−2.5−1.5−2.5−3.0
(kg)−10−10−10−15−15
(kg)−1.0−1.5−1.0−2.5−2.5
(kg/s)−0.8−1.0−1.0−1.5−2.0
(kg/s)−0.9−1.0−0.8−1.5−1.5
(kg/s)0.10.20.10.20.5
(kg/s)0.10.10.050.30.35
(kg/m)−1.3−1.3−1.0−0.85−1.5
(kg/m)−36−25−20−15−20

Parameters of AUVs.

Figure 2 illustrates the communication topology and the spanning tree where agent 0 is the virtual leader and is considered as the root, in accordance with Assumption 2. The desired formation pattern requires each AUV, , to track a periodic signal generated by the virtual leader . The dynamics of the leader are defined as follows:

FIGURE 2

The initial conditions and system matrix are structured to ensure all eigenvalues of lie on the imaginary axis, thus satisfying Assumption 1. The reference trajectory for is defined as . The predefined offsets , which determine the relative positions of the AUVs to the leader, are specified as follows:

Each AUV tracks its respective position in the formation by adjusting its location to .

6.1 DDL formation learning control simulation

The estimated virtual leader’s state, derived from the cooperative estimator in the first layer (see Equations 3, 4), is utilized to estimate each agent’s complete uncertain dynamics within the DDL controller (second layer) using Equations 13, 14. The uncertain nonlinear functions for each agent are approximated using RBFNN, as described in Equation 11. Specifically, for each agent , the nonlinear uncertain functions , dependent on , are modeled. The input to the NN, , allows the construction of Gaussian RBFNN, represented by , utilizing 4,096 neurons arranged in an grid. The centers of these neurons are evenly distributed over the state space , and each has a width , ensuring bounded and structured parameter optimization for all and .

The observer and controller parameters are chosen as , and the diagonal matrices and , with and for all and . The initial conditions for the agents are set as , , , , and . Zero initial conditions are assumed for all the distributed observer states and the DDL controller states for all and . Time-domain simulation is carried out using the DDL formation learning control laws as specified in Equations 13, 14, along with Equations 3, 4.

Figure 3 displays the simulation results of the cooperative estimator (first layer) for all five agents. It illustrates how each agent’s estimated states,, converge perfectly to the leader’s states thorough Equations 3, 4. Figure 4 presents the position tracking control responses of all agents. Figures 4A–C illustrate the tracking performance of AUVs along the x-axis, y-axis, and vehicle heading, respectively, demonstrating effective tracking of the leader’s position signal. While the first AUV exactly tracks the leader’s states, agents 2 through 5 are shown to successfully follow agent 1, maintaining prescribed distances and alignment along the x and y axes, and matching the same heading angle. These results underscore the robustness of the real-time tracking control system, which enforces a predefined formation pattern, initially depicted in Figure 2. Additionally, Figure 5 highlights the real-time control performance for all agents, showcasing the effectiveness of the tracking strategy in maintaining the formation pattern.

FIGURE 3

FIGURE 4

FIGURE 5

The sum of the absolute values of the neural network weights in Figure 6. This convergence reflects the network’s ability to maintain consistent performance, as further adjustments to the weights become minimal. Also, updating neural network weights and their convergence throughout the learning process into their optimal valies depicted in Figure 7. This convergence of all neural network weights to their optimal values during the training process, aligns with Theorem 4 as well. This leads to achieving accurate function approximation in the second layer. Figure 8 represents the successful function approximation results for the unknown system dynamics as defined in Equation 11 for the third AUV, using RBFNN. The approximations are plotted for both and for all which defined in Theorem 4. The results confirm that locally accurate approximations of the AUV’s nonlinear dynamics were achieved. Moreover, this learned knowledge about the dynamics is effectively stored and represented using localized constant RBFNN.

FIGURE 6

FIGURE 7

FIGURE 8

6.2 Simulation for formation control with pre-learned dynamics

To evaluate the distributed control performance of the multi-AUV system, we implemented the pre-learned distributed formation control law. This strategy integrates the estimator observer Equations 3, 4, this time coupled with the constant RBFNN controller (Equation 21). We employed the virtual leader dynamics described in Equation 21 to generate consistent position tracking reference signals, as previously discussed in Section 6.1. To ensure a fair comparison, identical initial conditions and control gains and inputs were used across all simulations. Figure 9 illustrates the comparison of the tracking control results from Equations 13, 14 with the results using pre-trained weights in Equation 20.

FIGURE 9

The control experiments and simulation results presented demonstrate that the constant RBFNN control law (Equation 4) can achieve satisfactory tracking control performance comparable to that of the adaptive control laws (Equations 13, 14), but with no computational demand. The elimination of online recalculations or readaptations of the NN weights under this control strategy significantly reduces the computational load whenever system restarts without needing to retrain again. This reduction is particularly advantageous in scenarios involving extensive neural networks with a large number of neurons, thereby conserving system energy and enhancing operational efficiency in real-time applications.

Before concluding the paper, a brief contribution of the paper is provided:

  • • Distributed Observer Results: Simulations showed that the distributed observer effectively estimated the leader’s state, allowing for accurate formation control without needing global information.

  • • Tracking Control Results: The controller demonstrated reliable tracking of reference signals, maintaining performance even under varying conditions and unknown system dynamics.

  • • Formation Control: The proposed controller maintained accurate formation control relative to a virtual leader in simulations, even when the system dynamics were unknown with different AUVs.

  • • Neural Network Weight Convergence: The simulation results demonstrated that the neural network weights converged effectively, ensuring accurate function approximation and reliable performance in controlling AUVs under uncertainties.

  • • Adaptability and Stability: The framework ensured stable tracking performance across various environmental conditions by relying on the RBFNN’s learning capabilities, allowing the AUVs to use prelearned information and maintain formation control without needing to relearn dynamics whenever system restarts.

  • • Reduction in Computational Load: The use of pre-trained neural network weights significantly reduced the computational burden during real-time operation, particularly when large neural networks were employed.

7 Conclusion

In conclusion, this paper has introduced a novel two-layer control framework designed for Autonomous Underwater Vehicles (AUVs), aimed at universal applicability across various AUV configurations and environmental conditions. This framework assumes all system dynamics to be unknown, thereby enabling the controller to operate independently of specific dynamic parameters and effectively handle any environmental challenges, including hydrodynamic forces and torques. The framework consists of a first-layer distributed observer estimator that captures the leader’s dynamics using information from adjacent agents, and a second-layer decentralized deterministic learning controller. Each AUV utilizes the estimated signals from the first layer to determine the desired trajectory, simultaneously training its own dynamics using Radial Basis Function Neural Networks (RBFNN). This innovative approach not only sustains stability and performance in dynamic and unpredictable environments but also allows AUVs to efficiently utilize previously learned dynamics after system restarts, facilitating rapid resumption of optimal operations. The robustness and versatility of this framework have been rigorously confirmed through comprehensive simulations, demonstrating its potential to significantly enhance the adaptability and resilience of AUV systems. By embracing total uncertainty in system dynamics, this framework establishes a new benchmark in autonomous underwater vehicle control and lays a solid groundwork for future developments aimed at minimizing energy use and maximizing system flexibility. We plan to expand this framework by accommodating more general leader dynamics and conducting experimental applications to validate its performance in real-world settings. Moreover, a more accurate model of some source of uncertainty could improve performance which we will address in our future research. The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Statements

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

EJ: Conceptualization, Writing–original draft, Writing–review and editing. MZ: Conceptualization, Funding acquisition, Supervision, Writing–review and editing. PS: Conceptualization, Supervision, Writing–review and editing. CY: Conceptualization, Formal Analysis, Funding acquisition, Supervision, Writing–review and editing.

Funding

The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This work is supported in part by the National Science Foundation under Grant CMMI-1952862 and CMMI-2154901.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Footnotes

1.^A recurrent trajectory represents a large set of periodic and periodic-like trajectories generated from linear/nonlinear dynamical systems. A detailed characterization of recurrent trajectories can be found in Wang and Hill (2018).

References

  • 1

    BalchT.ArkinR. (1998). Behavior-based formation control for multirobot teams. IEEE Trans. Robotics Automation14, 926939. 10.1109/70.736776

  • 2

    CaiH.LewisF. L.HuG.HuangJ. (2015). “Cooperative output regulation of linear multi-agent systems by the adaptive distributed observer,” in 2015 54th IEEE Conference on Decision and Control (CDC) (IEEE), 54325437.

  • 3

    CaoX.RenL.SunC. (2022). Dynamic target tracking control of autonomous underwater vehicle based on trajectory prediction. IEEE Trans. Cybern.53, 19681981. 10.1109/tcyb.2022.3189688

  • 4

    ChristenS.JendeleL.AksanE.HilligesO. (2021). Learning functionally decomposed hierarchies for continuous control tasks with path planning. IEEE Robotics Automation Lett.6, 36233630. 10.1109/lra.2021.3060403

  • 5

    CuiR.GeS. S.HowB. V. E.ChooY. S. (2010). Leader–follower formation control of underactuated autonomous underwater vehicles. Ocean. Eng.37, 14911502. 10.1016/j.oceaneng.2010.07.006

  • 6

    DongX.YuanC.StegagnoP.ZengW.WangC. (2019). Composite cooperative synchronization and decentralized learning of multi-robot manipulators with heterogeneous nonlinear uncertain dynamics. J. Frankl. Inst.356, 50495072. 10.1016/j.jfranklin.2019.04.028

  • 7

    FossenT. I. (1999). “Guidance and control of ocean vehicles,”. Norway: University of Trondheim. Doctors Thesis.

  • 8

    GhafooriS.RabieeA.CeteraA.AbiriR. (2024). Bispectrum analysis of noninvasive eeg signals discriminates complex and natural grasp types. arXiv Prepr. arXiv:2402.01026, 15. 10.1109/embc53108.2024.10782163

  • 9

    HadiB.KhosraviA.SarhadiP. (2021). A review of the path planning and formation control for multiple autonomous underwater vehicles. J. Intelligent and Robotic Syst.101, 6726. 10.1007/s10846-021-01330-4

  • 10

    HouS. P.CheahC. C. (2009). “Coordinated control of multiple autonomous underwater vehicles for pipeline inspection,” in Proceedings of the 48h IEEE Conference on Decision and Control (CDC) held jointly with 2009 28th Chinese Control Conference (IEEE), 31673172.

  • 11

    IoannouP. A.SunJ. (1996). Robust adaptive control, 1. Upper Saddle River, NJ: PTR Prentice-Hall.

  • 12

    JandaghiE.ChenX.YuanC. (2023). “Motion dynamics modeling and fault detection of a soft trunk robot,” in 2023 IEEE/ASME International Conference on Advanced Intelligent Mechatronics (AIM) (IEEE), 13241329.

  • 13

    JandaghiE.SteinD. L.HoburgA.StegagnoP.ZhouM.YuanC. (2024). “Composite distributed learning and synchronization of nonlinear multi-agent systems with complete uncertain dynamics,” in 2024 IEEE International Conference on Advanced Intelligent Mechatronics (AIM), 13671372. 10.1109/aim55361.2024.10637197

  • 14

    KrsticM.KokotovicP. V.KanellakopoulosI. (1995). Nonlinear and adaptive control design. John Wiley and Sons, Inc.

  • 15

    LawtonJ. R. T. (2000). A Behavior-Based Approach to Multiple Spacecraft Formation Flying (Ph.D. thesis). Brigham Young University, Provo, UT, United States.

  • 16

    MillánP.OrihuelaL.JuradoI.RubioF. R. (2013). Formation control of autonomous underwater vehicles subject to communication delays. IEEE Trans. Control Syst. Technol.22, 770777. 10.1109/tcst.2013.2262768

  • 17

    ParkJ.SandbergI. W. (1991). Universal approximation using radial-basis-function networks. Neural Comput.3, 246257. 10.1162/neco.1991.3.2.246

  • 18

    PengZ.WangD.ShiY.WangH.WangW. (2015). Containment control of networked autonomous underwater vehicles with model uncertainty and ocean disturbances guided by multiple leaders. Inf. Sci.316, 163179. 10.1016/j.ins.2015.04.025

  • 19

    PengZ.WangJ.WangD. (2017). Distributed maneuvering of autonomous surface vehicles based on neurodynamic optimization and fuzzy approximation. IEEE Trans. Control Syst. Technol.26, 10831090. 10.1109/tcst.2017.2699167

  • 20

    PresteroT. (2001). “Development of a six-degree of freedom simulation model for the remus autonomous underwater vehicle,” in MTS/IEEE Oceans 2001. An Ocean Odyssey. Conference Proceedings (IEEE Cat. No.01CH37295), Honolulu, HI, USA, 2001, pp.450455 vol.1, 10.1109/OCEANS.2001.968766

  • 21

    RenW.BeardR. W. (2005). Consensus seeking in multiagent systems under dynamically changing interaction topologies. IEEE Trans. automatic control50, 655661. 10.1109/tac.2005.846556

  • 22

    RenW.BeardR. W. (2008). Distributed consensus in multi-vehicle cooperative control, 27. Springer.

  • 23

    RoutR.SubudhiB. (2016). A backstepping approach for the formation control of multiple autonomous underwater vehicles using a leader–follower strategy. J. Mar. Eng. and Technol.15, 3846. 10.1080/20464177.2016.1173268

  • 24

    SkjetneR.FossenT. I.KokotovićP. V. (2005). Adaptive maneuvering, with experiments, for a model ship in a marine control laboratory. Automatica41, 289298. 10.1016/j.automatica.2004.10.006

  • 25

    SuY.HuangJ. (2011). Cooperative output regulation of linear multi-agent systems. IEEE Trans. Automatic Control57, 10621066. 10.1109/TAC.2011.2169618

  • 26

    TutsoyO.AsadiD.AhmadiK.Nabavi-ChashmiS. Y.IqbalJ. (2024). Minimum distance and minimum time optimal path planning with bioinspired machine learning algorithms for faulty unmanned air vehicles. IEEE Trans. Intelligent Transp. Syst.25, 90699077. 10.1109/tits.2024.3367769

  • 27

    WangC.HillD. J. (2018). Deterministic learning theory for identification, recognition, and control. CRC Press. 10.1201/9781315221755

  • 28

    YanT.XuZ.YangS. X.GadsdenS. A. (2023). Formation control of multiple autonomous underwater vehicles: a review. Intell. and Robotics3, 122. 10.20517/ir.2023.01

  • 29

    YanZ.LiuX.ZhouJ.WuD. (2018). Coordinated target tracking strategy for multiple unmanned underwater vehicles with time delays. IEEE Access6, 1034810357. 10.1109/access.2018.2793338

  • 30

    YangY.XiaoY.LiT. (2021). A survey of autonomous underwater vehicle formation: performance, formation control, and communication capability. IEEE Commun. Surv. and Tutorials23, 815841. 10.1109/comst.2021.3059998

  • 31

    YuanC. (2017). Leader-following consensus of parameter-dependent networks via distributed gain-scheduling control. Int. J. Syst. Sci.48, 20132022. 10.1080/00207721.2017.1309597

  • 32

    YuanC.LichtS.HeH. (2017). Formation learning control of multiple autonomous underwater vehicles with heterogeneous nonlinear uncertain dynamics. IEEE Trans. Cybern.48, 29202934. 10.1109/tcyb.2017.2752458

  • 33

    YuanC.WangC. (2011). Persistency of excitation and performance of deterministic learning. Syst. and control Lett.60, 952959. 10.1016/j.sysconle.2011.08.002

  • 34

    YuanC.WangC. (2012). Performance of deterministic learning in noisy environments. Neurocomputing78, 7282. 10.1016/j.neucom.2011.05.037

  • 35

    ZhangY.LiS.LiuX. (2018). Neural network-based model-free adaptive near-optimal tracking control for a class of nonlinear systems. IEEE Trans. neural Netw. Learn. Syst.29, 62276241. 10.1109/tnnls.2018.2828114

  • 36

    ZhouJ.SiY.ChenY. (2023). A review of subsea auv technology. J. Mar. Sci. Eng.11, 1119. 10.3390/jmse11061119

Summary

Keywords

environment-independent controller, autonomous underwater vehicles (AUV), dynamic learning, formation learning control, multi-agent systems, neural network control, adaptive control, robotics

Citation

Jandaghi E, Zhou M, Stegagno P and Yuan C (2025) Adaptive formation learning control for cooperative AUVs under complete uncertainty. Front. Robot. AI 11:1491907. doi: 10.3389/frobt.2024.1491907

Received

05 September 2024

Accepted

12 December 2024

Published

14 February 2025

Volume

11 - 2024

Edited by

Giovanni Iacca, University of Trento, Italy

Reviewed by

Önder Tutsoy, Adana Science and Technology University, Türkiye

Di Wu, Harbin University of Science and Technology, China

Updates

Copyright

*Correspondence: Chengzhi Yuan,

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics