kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Learning deep energy shaping policies for stability-guaranteed manipulation
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL.ORCID iD: 0000-0003-0443-7982
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL.ORCID iD: 0000-0002-3599-440x
ABB Corporate Research, Vasteras, 72178, Sweden.ORCID iD: 0000-0003-1133-0884
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL. ABB Corporate Research, Vasteras, 72178, Sweden.ORCID iD: 0000-0003-2965-2953
2021 (English)In: IEEE Robotics and Automation Letters, E-ISSN 2377-3766, Vol. 6, no 4, p. 8583-8590Article in journal (Refereed) Published
Abstract [en]

Deep reinforcement learning (DRL) has been successfully used to solve various robotic manipulation tasks. However, most of the existing works do not address the issue of control stability. This is in sharp contrast to the control theory community where the well-established norm is to prove stability whenever a control law is synthesized. What makes traditional stability analysis difficult for DRL are the uninterpretable nature of the neural network policies and unknown system dynamics. In this work, stability is obtained by deriving an interpretable deep policy structure based on the energy shaping control of Lagrangian systems. Then, stability during physical interaction with an unknown environment is established based on passivity. The result is a stability guaranteeing DRL in a model-free framework that is general enough for contact-rich manipulation tasks. With an experiment on a peg-in-hole task, we demonstrate, to the best of our knowledge, the first DRL with stability guarantee on a real robotic manipulator.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE) , 2021. Vol. 6, no 4, p. 8583-8590
Keywords [en]
Machine learning for robot control, reinforcement learning, Agricultural robots, Control theory, Industrial manipulators, Manipulators, Robotics, System stability, Control stability, Energy shaping control, Physical interactions, Robotic manipulation, Robotic manipulators, Stability analysis, Unconditional stability, Unknown environments, Deep learning
National Category
Robotics and automation
Identifiers
URN: urn:nbn:se:kth:diva-311752DOI: 10.1109/LRA.2021.3111962ISI: 000701239400004Scopus ID: 2-s2.0-85115187899OAI: oai:DiVA.org:kth-311752DiVA, id: diva2:1656000
Note

QC 20220504

Available from: 2022-05-04 Created: 2022-05-04 Last updated: 2025-02-09Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Abdul Khader, ShahbazYin, HangKragic, Danica

Search in DiVA

By author/editor
Abdul Khader, ShahbazYin, HangFalco, PietroKragic, Danica
By organisation
Robotics, Perception and Learning, RPL
In the same journal
IEEE Robotics and Automation Letters
Robotics and automation

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 307 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf