Critic-actor reinforcement learning for optimized cooperative formation of multi-nonholonomic wheeled Mobile vehicles
View abstract on PubMed
Summary
This summary is machine-generated.This study introduces an optimized formation control strategy for multiple nonholonomic wheeled mobile vehicles using critic-actor reinforcement learning and adaptive neural networks. The method effectively manages complex dynamics for cooperative navigation.
Area Of Science
- Robotics
- Control Systems
- Artificial Intelligence
Background
- Formation control for multi-vehicle systems is complex due to under-actuation and nonholonomic constraints.
- Solving the Hamilton-Jacobi-Bellman equation for optimized formation protocols is challenging.
- Existing methods struggle with uncertainties in Lagrange dynamics.
Purpose Of The Study
- To develop an optimized leader-follower formation control scheme for multiple nonholonomic wheeled mobile vehicles (MNWMVs).
- To integrate critic-actor reinforcement learning (RL) with adaptive neural networks (NN) for robust formation tracking.
- To address nonlinear and coupled characteristics of the Hamilton-Jacobi-Bellman equation.
Main Methods
- A distributed cooperative formation scheme combining kinematic and dynamic torque controllers.
- Integration of an adaptive identifier within the critic-actor RL strategy to handle Lagrange dynamics uncertainties.
- Utilizing RL training laws derived from the negative gradient of a positive function for simplification.
Main Results
- A novel optimized formation tracking algorithm for MNWMVs was developed.
- The adaptive identifier effectively compensated for uncertainties in Lagrange dynamics.
- The RL training laws simplified the overall formation scheme.
Conclusions
- The proposed distributed cooperative formation scheme is effective for MNWMVs.
- The integration of RL and adaptive NNs provides a robust solution for complex formation control problems.
- Numerical simulations and physical experiments validated the theoretical findings.
Related Concept Videos
Cooperative allosteric transitions can occur in multimeric proteins, where each subunit of the protein has its own ligand-binding site. When a ligand binds to any of these subunits, it triggers a conformational change that affects the binding sites in the other subunits; this can change the affinity of the other sites for their respective ligands. The ability of the protein to change the shape of its binding site is attributed to the presence of a mix of flexible and stable segments in the...
Transcriptional regulators bind to specific cis-regulatory sequences in the DNA to regulate gene transcription. These cis-regulatory sequences are very short, usually less than ten nucleotide pairs in length. The short length means that there is a high probability of the exact same sequence randomly occurring throughout the genome. Since regulators can also bind to groups of similar sequences, this further increases the chances of random binding. Transcriptional regulators form...
The actor-observer effect, a cognitive bias closely linked to the fundamental attribution error, refers to the tendency for individuals to attribute their behavior to external, situational factors while explaining others’ behavior in terms of internal, dispositional traits. This asymmetry in attribution significantly influences social perception and judgment.Cognitive Mechanisms Behind the EffectTwo primary psychological mechanisms contribute to the actor-observer effect: differences in...
Positive and negative reinforcement are key concepts in operant conditioning, a learning process where the consequences of a behavior affect the likelihood of that behavior being repeated.
Positive reinforcement occurs when a behavior is followed by the presentation of a rewarding stimulus, increasing the frequency of that behavior. For example:
• If a person smiles at you after you greet them and you continue talking, the smile is a positive reinforcement for your greeting.
•...

