kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Latent Space Roadmap for Visual Action Planning of Deformable and Rigid Object Manipulation
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL. University of Salerno, Salerno, Italy.ORCID iD: 0000-0003-0470-9191
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL. KTH, School of Electrical Engineering and Computer Science (EECS), Centres, Centre for Autonomous Systems, CAS.ORCID iD: 0000-0001-6920-5109
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL.ORCID iD: 0000-0003-3827-3824
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL.ORCID iD: 0000-0002-0900-1523
Show others and affiliations
2020 (English)In: 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Institute of Electrical and Electronics Engineers (IEEE) , 2020, p. 5619-5626, article id 9340764Conference paper, Published paper (Refereed)
Abstract [en]

We present a framework for visual action planning of complex manipulation tasks with high-dimensional state spaces such as manipulation of deformable objects. Planning is performed in a low-dimensional latent state space that embeds images. We define and implement a Latent Space Roadmap (LSR) which is a graph-based structure that globally captures the latent system dynamics. Our framework consists of two main components: a Visual Foresight Module (VFM) that generates a visual plan as a sequence of images, and an Action Proposal Network (APN) that predicts the actions between them. We show the effectiveness of the method on a simulated box stacking task as well as a T-shirt folding task performed with a real robot.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE) , 2020. p. 5619-5626, article id 9340764
Series
IEEE International Conference on Intelligent Robots and Systems, ISSN 0302-9743
Keywords [en]
Representation Learning, Perception-Action Coupling, Novel Deep Learning Methods
National Category
Robotics and automation
Research subject
Computer Science
Identifiers
URN: urn:nbn:se:kth:diva-282853DOI: 10.1109/IROS45743.2020.9340764ISI: 000714033803051Scopus ID: 2-s2.0-85098616092OAI: oai:DiVA.org:kth-282853DiVA, id: diva2:1472207
Conference
2020 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2020 Las Vegas 24 October 2020 through 24 January 2021
Note

QC 20201008

Part of proceeding: ISBN 978-172816212-6

Available from: 2020-10-01 Created: 2020-10-01 Last updated: 2025-02-09Bibliographically approved

Open Access in DiVA

IROS2020(5344 kB)180 downloads
File information
File name FULLTEXT01.pdfFile size 5344 kBChecksum SHA-512
60f32c5a81888ce9b683f3b4860151a5973c471ba25a5e7452befcaa7dbb12ba2de5ef2fdfb11c2a5bbaba9660444bc60acb09c79959885b5e7feac7fc1f218a
Type fulltextMimetype application/pdf

Other links

Publisher's full textScopusConference

Authority records

Lippi, MartinaPoklukar, PetraWelle, Michael C.Varava, AnastasiiaYin, HangKragic, Danica

Search in DiVA

By author/editor
Lippi, MartinaPoklukar, PetraWelle, Michael C.Varava, AnastasiiaYin, HangKragic, Danica
By organisation
Robotics, Perception and Learning, RPLCentre for Autonomous Systems, CAS
Robotics and automation

Search outside of DiVA

GoogleGoogle Scholar
Total: 180 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 500 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf