kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Interactive Perception for Deformable Object Manipulation
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL. KTH, School of Electrical Engineering and Computer Science (EECS), Centres, Centre for Autonomous Systems, CAS.ORCID iD: 0000-0002-9486-9238
Hong Kong Polytech Univ PolyU, Kowloon, Hong Kong, Peoples R China..ORCID iD: 0000-0002-7020-0943
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL.ORCID iD: 0000-0002-3599-440x
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL. KTH, School of Engineering Sciences in Chemistry, Biotechnology and Health (CBH), Chemistry, Organic chemistry.ORCID iD: 0000-0002-9001-7708
Show others and affiliations
2024 (English)In: IEEE Robotics and Automation Letters, E-ISSN 2377-3766, Vol. 9, no 9, p. 7763-7770Article in journal (Refereed) Published
Abstract [en]

Interactive perception enables robots to manipulate the environment and objects to bring them into states that benefit the perception process. Deformable objects pose challenges to this due to manipulation difficulty and occlusion in vision-based perception. In this work, we address such a problem with a setup involving both an active camera and an object manipulator. Our approach is based on a sequential decision-making framework and explicitly considers the motion regularity and structure in coupling the camera and manipulator. We contribute a method for constructing and computing a subspace, called Dynamic Active Vision Space (DAVS), for effectively utilizing the regularity in motion exploration. The effectiveness of the framework and approach are validated in both a simulation and a real dual-arm robot setup. Our results confirm the necessity of an active camera and coordinative motion in interactive perception for deformable objects.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE) , 2024. Vol. 9, no 9, p. 7763-7770
Keywords [en]
Cameras, Manifolds, IP networks, End effectors, Task analysis, Couplings, Robot kinematics, Perception for grasping and manipulation, perception-action coupling, manipulation planning
National Category
Robotics and automation
Identifiers
URN: urn:nbn:se:kth:diva-352106DOI: 10.1109/LRA.2024.3431943ISI: 001283670800004Scopus ID: 2-s2.0-85199505576OAI: oai:DiVA.org:kth-352106DiVA, id: diva2:1891398
Note

QC 20240822

Available from: 2024-08-22 Created: 2024-08-22 Last updated: 2025-05-14Bibliographically approved
In thesis
1. Approach-constrained Grasp Synthesis and Interactive Perception for Rigid and Deformable Objects
Open this publication in new window or tab >>Approach-constrained Grasp Synthesis and Interactive Perception for Rigid and Deformable Objects
2025 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

This thesis introduces methods for two robotic tasks: grasp synthesis and deformable object manipulation. These tasks are connected by interactive perception, where robots actively manipulate objects to improve sensory feed-back and task performance. Achieving a collision-free, successful grasp is essential for subsequent interaction, while effective manipulation of deformable objects broadens real-world applications. For robotic grasp synthesis, we address the challenge of approach-constrained grasping. We introduce two methods: GoNet and CAPGrasp. GoNet learns a grasp sampler that generates grasp poses with approach directions that lie in a selected discretized bin. In contrast, CAPGrasp enables sampling in a continuous space without requiring explicit approach direction annotations in the learning phase, improving the grasp success rate and providing more flexibility for imposing approach constraint. For robotic deformable object manipulation, we focus on manipulating deformable bags with handles—a common daily human activity. We first propose a method that captures scene dynamics and predicts future states in environments containing both rigid spheres and a deformable bag. Our approach employs an object-centric graph representation and an encoder-decoder framework to forecast future graph states. Additionally, we integrate an active camera into the system, explicitly considering the regularity and structure of motion to couple the camera with the manipulator for effective exploration.

To address the common data scarcity issue in both domains, we also develop simulation environments and propose annotated datasets for extensive benchmarking. Experimental results on both simulated and real-world platforms demonstrate the effectiveness of our methods compared to established baselines.

Abstract [sv]

Denna avhandling introducerar metoder för två robotuppgifter: grepp-syntes och manipulering av deformerbara objekt. Dessa uppgifter är sam-mankopplade genom interaktiv perception, där robotar aktivt manipulerar objekt för att förbättra sensorisk feedback och uppgiftsutförande. Att uppnå ett kollisionsfritt, framgångsrikt grepp är avgörande för efterföljande interak-tion, medan effektiv manipulering av deformerbara objekt breddar verkliga tillämpningar. För robotisk greppsyntes tar vi oss an utmaningen med tillvägagångssätt-begränsat grepp. Vi introducerar två metoder: GoNet och CAPGrasp. GoNet lär sig en gripsamplare som genererar gripposer med inflygningsriktningar som ligger i en vald diskretiserad bin. CAPGrasp, däremot, möjliggör sampling i ett kontinuerligt utrymme utan att kräva explicita tillvägagångssättsanvisningar i inlärningsfasen, vilket förbättrar greppets framgångsfrekvens och ger mer flexibilitet för att införa begränsningar för tillvägagångssätt.

För robotmanipulering av deformerbara föremål fokuserar vi på att manipulera deformerbara påsar med handtag - en vanlig mänsklig aktivitet. Vi föreslår först en metod som fångar scenens dynamik och förutsäger framti-da tillstånd i miljöer som innehåller både stela sfärer och en deformerbar påse. Vårt tillvägagångssätt använder en objektcentrerad grafrepresentation och ett ramverk för kodare-avkodare för att förutsäga framtida graftillstånd. Dessutom integrerar vi en aktiv kamera i systemet, och tar uttryckligen hänsyn till rörelsens regelbundenhet och struktur för att koppla ihop kameran med manipulatorn för effektiv utforskning. För att ta itu med det vanliga problemet med databrist i båda domänerna utvecklar vi också simuleringsmiljöer och föreslår kommenterade datauppsättningar för omfattande benchmarking. Experimentella resultat på både simulerade och verkliga plattformar visar effektiviteten hos våra metoder jämfört med etablerade baslinjer.

Place, publisher, year, edition, pages
KTH Royal Institute of Technology, 2025. p. 52
Series
TRITA-EECS-AVL ; 2025:63
National Category
Robotics and automation
Research subject
Computer Science
Identifiers
urn:nbn:se:kth:diva-363359 (URN)978-91-8106-304-2 (ISBN)
Public defence
2025-06-10, https://kth-se.zoom.us/j/68663108750, D3, Lindstedtvägen 9, Stockholm, Stockholm, 14:30 (English)
Opponent
Supervisors
Note

QC 20250514

Available from: 2025-05-14 Created: 2025-05-14 Last updated: 2025-06-30Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Weng, ZehangYin, HangKravchenko, AlexanderVarava, AnastasiiaKragic, Danica

Search in DiVA

By author/editor
Weng, ZehangZhou, PengYin, HangKravchenko, AlexanderVarava, AnastasiiaKragic, Danica
By organisation
Robotics, Perception and Learning, RPLCentre for Autonomous Systems, CASOrganic chemistryCollaborative Autonomous Systems
In the same journal
IEEE Robotics and Automation Letters
Robotics and automation

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 175 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf