Visual object-action recognition: Inferring object affordances from human demonstration
2011 (English)In: Computer Vision and Image Understanding, ISSN 1077-3142, E-ISSN 1090-235X, Vol. 115, no 1, 81-90 p.Article in journal (Refereed) Published
This paper investigates object categorization according to function, i.e., learning the affordances of objects from human demonstration. Object affordances (functionality) are inferred from observations of humans using the objects in different types of actions. The intended application is learning from demonstration, in which a robot learns to employ objects in household tasks, from observing a human performing the same tasks with the objects. We present a method for categorizing manipulated objects and human manipulation actions in context of each other. The method is able to simultaneously segment and classify human hand actions, and detect and classify the objects involved in the action. This can serve as an initial step in a learning from demonstration method. Experiments show that the contextual information improves the classification of both objects and actions.
Place, publisher, year, edition, pages
2011. Vol. 115, no 1, 81-90 p.
Object recognition, Action recognition, Contextual recognition, Object affordances, Learning from demonstration
IdentifiersURN: urn:nbn:se:kth:diva-28603DOI: 10.1016/j.cviu.2010.08.002ISI: 000285275200008ScopusID: 2-s2.0-78751574174OAI: oai:DiVA.org:kth-28603DiVA: diva2:388076
FunderICT - The Next Generation
QC 201101172011-01-172011-01-172012-01-19Bibliographically approved