This study presents the design and implementation of a reinforcement learning (RL)-based framework for the control of an autonomous underwater vehicle (AUV) directly within Unreal Engine (UE). A high-fidelity aquatic environment was created using UE’s native Water System to simulate hydrodynamic forces and buoyancy. Unlike studies assuming continuous control, this research addresses the challenge of stabilizing an AUV subject to severe discrete 'bang-bang' hardware constraints. A parallelized Proximal Policy Optimization (PPO) algorithm was employed to synthesize adaptive control policies. Comparative analysis against tuned Proportional-Integral-Derivative (PID) baselines demonstrates that the RL agent outperforms classical methods in three key metrics: (1) in longitudinal navigation, the agent learned an emergent "pulsing" strategy—mimicking Pulse-Width Modulation (PWM)—to overcome these discrete actuation constraints, reducing steady-state error compared to the Proportional-Derivative (PD) baseline; (2) in vertical depth control, the agent autonomously learned gravity compensation, settling faster than integral-based controllers while avoiding buoyancy-induced stalling; and (3) in heading control, the agent demonstrated superior dynamic handling, completing stabilization maneuvers faster than the baseline. The key architectural innovation lies in the direct integration of UE’s Learning Agents plugin, eliminating the need for external middleware. This native integration enables real-time synchronization between simulation physics and learning processes, establishing a high-fidelity platform for developing intelligent underwater control systems.
Autonomous Underwater Vehicles Reinforcement Learning Unreal Engine Marine Robotics Proximal Policy Optimization Motion Planning
| Primary Language | English |
|---|---|
| Subjects | Autonomous Agents and Multiagent Systems |
| Journal Section | Research Article |
| Authors | |
| Submission Date | October 30, 2025 |
| Acceptance Date | March 24, 2026 |
| Publication Date | March 31, 2026 |
| DOI | https://doi.org/10.54287/gujsa.1813751 |
| IZ | https://izlik.org/JA58JP52YD |
| Published in Issue | Year 2026 Volume: 13 Issue: 1 |