kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Visual object-action recognition: Inferring object affordances from human demonstration
KTH, School of Computer Science and Communication (CSC), Centres, Centre for Autonomous Systems, CAS.ORCID iD: 0000-0002-5750-9655
KTH, School of Computer Science and Communication (CSC), Centres, Centre for Autonomous Systems, CAS.
KTH, School of Computer Science and Communication (CSC), Centres, Centre for Autonomous Systems, CAS.ORCID iD: 0000-0003-2965-2953
2011 (English)In: Computer Vision and Image Understanding, ISSN 1077-3142, E-ISSN 1090-235X, Vol. 115, no 1, p. 81-90Article in journal (Refereed) Published
Abstract [en]

This paper investigates object categorization according to function, i.e., learning the affordances of objects from human demonstration. Object affordances (functionality) are inferred from observations of humans using the objects in different types of actions. The intended application is learning from demonstration, in which a robot learns to employ objects in household tasks, from observing a human performing the same tasks with the objects. We present a method for categorizing manipulated objects and human manipulation actions in context of each other. The method is able to simultaneously segment and classify human hand actions, and detect and classify the objects involved in the action. This can serve as an initial step in a learning from demonstration method. Experiments show that the contextual information improves the classification of both objects and actions.

Place, publisher, year, edition, pages
2011. Vol. 115, no 1, p. 81-90
Keywords [en]
Object recognition, Action recognition, Contextual recognition, Object affordances, Learning from demonstration
National Category
Information Systems
Identifiers
URN: urn:nbn:se:kth:diva-28603DOI: 10.1016/j.cviu.2010.08.002ISI: 000285275200008Scopus ID: 2-s2.0-78751574174OAI: oai:DiVA.org:kth-28603DiVA, id: diva2:388076
Funder
ICT - The Next Generation
Note
QC 20110117Available from: 2011-01-17 Created: 2011-01-17 Last updated: 2022-06-25Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Kjellström, HedvigKragic, Danica

Search in DiVA

By author/editor
Kjellström, HedvigRomero, JavierKragic, Danica
By organisation
Centre for Autonomous Systems, CAS
In the same journal
Computer Vision and Image Understanding
Information Systems

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 356 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf