kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Enabling Visual Action Planning for Object Manipulation Through Latent Space Roadmap
Roma Tre University, Rome, Italy.
KTH, School of Electrical Engineering and Computer Science (EECS), Computer Science, Computational Science and Technology (CST).ORCID iD: 0000-0001-6920-5109
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL.ORCID iD: 0000-0003-3827-3824
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL.ORCID iD: 0000-0002-0900-1523
Show others and affiliations
2023 (English)In: IEEE Transactions on robotics, ISSN 1552-3098, E-ISSN 1941-0468, Vol. 39, no 1, p. 57-75Article in journal (Refereed) Published
Abstract [en]

In this article, we present a framework for visual action planning of complex manipulation tasks with high-dimensional state spaces, focusing on manipulation of deformable objects. We propose a latent space roadmap (LSR) for task planning, which is a graph-based structure globally capturing the system dynamics in a low-dimensional latent space. Our framework consists of the following three parts. First, a mapping module (MM) that maps observations is given in the form of images into a structured latent space extracting the respective states as well as generates observations from the latent states. Second, the LSR, which builds and connects clusters containing similar states in order to find the latent plans between start and goal states, extracted by MM. Third, the action proposal module that complements the latent plan found by the LSR with the corresponding actions. We present a thorough investigation of our framework on simulated box stacking and rope/box manipulation tasks, and a folding task executed on a real robot. 

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE) , 2023. Vol. 39, no 1, p. 57-75
Keywords [en]
Deep Learning in Robotics and Automation, Latent Space Planning, Manipulation Planning, Visual Learning, Deep learning, Graphic methods, Job analysis, Planning, Robot programming, Action planning, Deep learning in robotic and automation, Heuristics algorithm, Roadmap, Space planning, Stackings, Task analysis, Heuristic algorithms
National Category
Robotics and automation
Identifiers
URN: urn:nbn:se:kth:diva-326180DOI: 10.1109/TRO.2022.3188163ISI: 000829072000001Scopus ID: 2-s2.0-85135223386OAI: oai:DiVA.org:kth-326180DiVA, id: diva2:1753939
Note

QC 20230502

Available from: 2023-05-02 Created: 2023-05-02 Last updated: 2025-02-09Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Lippi, MartinaPoklukar, PetraWelle, Michael C.Varava, AnastasiiaYin, HangKragic, Danica

Search in DiVA

By author/editor
Lippi, MartinaPoklukar, PetraWelle, Michael C.Varava, AnastasiiaYin, HangKragic, Danica
By organisation
Computational Science and Technology (CST)Robotics, Perception and Learning, RPL
In the same journal
IEEE Transactions on robotics
Robotics and automation

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 238 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf