kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Contextual Modeling with Labeled Multi-LDA
KTH, School of Computer Science and Communication (CSC), Computer Vision and Active Perception, CVAP. KTH, School of Computer Science and Communication (CSC), Centres, Centre for Autonomous Systems, CAS. (CVAP)ORCID iD: 0000-0002-8640-9370
KTH, School of Computer Science and Communication (CSC), Computer Vision and Active Perception, CVAP. KTH, School of Computer Science and Communication (CSC), Centres, Centre for Autonomous Systems, CAS.
KTH, School of Computer Science and Communication (CSC), Computer Vision and Active Perception, CVAP. KTH, School of Computer Science and Communication (CSC), Centres, Centre for Autonomous Systems, CAS.ORCID iD: 0000-0002-5750-9655
2013 (English)In: 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), IEEE , 2013, p. 2264-2271Conference paper, Published paper (Refereed)
Abstract [en]

Learning about activities and object affordances from human demonstration are important cognitive capabilities for robots functioning in human environments, for example, being able to classify objects and knowing how to grasp them for different tasks. To achieve such capabilities, we propose a Labeled Multi-modal Latent Dirichlet Allocation (LM-LDA), which is a generative classifier trained with two different data cues, for instance, one cue can be traditional visual observation and another cue can be contextual information. The novel aspects of the LM-LDA classifier, compared to other methods for encoding contextual information are that, I) even with only one of the cues present at execution time, the classification will be better than single cue classification since cue correlations are encoded in the model, II) one of the cues (e.g., common grasps for the observed object class) can be inferred from the other cue (e.g., the appearance of the observed object). This makes the method suitable for robot online and transfer learning; a capability highly desirable in cognitive robotic applications. Our experiments show a clear improvement for classification and a reasonable inference of the missing data.

Place, publisher, year, edition, pages
IEEE , 2013. p. 2264-2271
Series
IEEE International Conference on Intelligent Robots and Systems, ISSN 2153-0858
Keywords [en]
LDA, Topic Model, Contextual Modeling
National Category
Computer graphics and computer vision
Identifiers
URN: urn:nbn:se:kth:diva-134127DOI: 10.1109/IROS.2013.6696673ISI: 000331367402063Scopus ID: 2-s2.0-84893758693OAI: oai:DiVA.org:kth-134127DiVA, id: diva2:664981
Conference
IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), November 3-8, 2013 at Tokyo Big Sight, Japan
Note

QC 20131217

Available from: 2013-11-18 Created: 2013-11-18 Last updated: 2025-02-07Bibliographically approved

Open Access in DiVA

Contextual Modeling with Labeled Multi-LDA(1520 kB)499 downloads
File information
File name FULLTEXT01.pdfFile size 1520 kBChecksum SHA-512
b3b741e0a8a3f9cf05241728489310be20e797acfccc9b39bcaaf98ffa638d2530d2dc255254df71ba725816825db6e56143baae04c96c4cb344b254b421fc8b
Type fulltextMimetype application/pdf

Other links

Publisher's full textScopusConference Homepage

Authority records

Zhang, ChengKjellström, Hedvig

Search in DiVA

By author/editor
Zhang, ChengSong, DanKjellström, Hedvig
By organisation
Computer Vision and Active Perception, CVAPCentre for Autonomous Systems, CAS
Computer graphics and computer vision

Search outside of DiVA

GoogleGoogle Scholar
Total: 500 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 312 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf