kth.sePublications KTH
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Using Reinforcement Learning for Hydrobatic Maneuvering with Autonomous Underwater Vehicles
KTH, School of Engineering Sciences (SCI), Engineering Mechanics, Aerospace, moveability and naval architecture.
KTH, School of Engineering Sciences (SCI), Engineering Mechanics, Aerospace, moveability and naval architecture.ORCID iD: 0000-0002-5839-5573
KTH, School of Engineering Sciences (SCI), Engineering Mechanics, Aerospace, moveability and naval architecture.ORCID iD: 0000-0001-7542-3225
2024 (English)In: OCEANS 2024 - SINGAPORE, Institute of Electrical and Electronics Engineers (IEEE) , 2024Conference paper, Published paper (Refereed)
Abstract [en]

Hydrobatic autonomous underwater vehicles (AUVs) can be efficient in speed and range as well as agile in maneuvering, thereby enabling new use cases in ocean production, environmental sensing, and security. However, such robots are underactuated, have highly nonlinear dynamics at high angles of attack, and will be used in applications with high requirements for robustness. This paper explores the use of reinforcement learning (RL) to control hydrobatic AUVs, using the agile SAM AUV as a case study. The focus is on controlling the depth and pitch simultaneously, where there is a tight coupling between the states. This maneuver offers a simple, yet interesting test case to compare different control strategies. The twin-delay deep deterministic policy gradient (TD3) algorithm is applied to this AUV control problem. The resulting trained RL controller offers good robustness to noise and performs at a similar level as a Proportional-Integral-Derivative (PID) controller within the Stonefish simulation environment. The agent is also deployed and run on the robot hardware, with high overshoot. While the RL agent has good performance in simulation, the transfer from simulation to reality still leaves some open questions.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE) , 2024.
National Category
Control Engineering
Identifiers
URN: urn:nbn:se:kth:diva-357059DOI: 10.1109/OCEANS51537.2024.10682215ISI: 001332919300089Scopus ID: 2-s2.0-85206490676OAI: oai:DiVA.org:kth-357059DiVA, id: diva2:1918039
Conference
OCEANS Conference, APR 15-18, 2024, Singapore, Singapore
Note

Part of ISBN 979-8-3503-6207-7

QC 20241204

Available from: 2024-12-04 Created: 2024-12-04 Last updated: 2024-12-04Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Bhat, SriharshaStenius, Ivan

Search in DiVA

By author/editor
Woźniak, GrzegorzBhat, SriharshaStenius, Ivan
By organisation
Aerospace, moveability and naval architecture
Control Engineering

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 91 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf