kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Inter-battery topic representation learning
KTH, School of Computer Science and Communication (CSC), Robotics, perception and learning, RPL.
KTH, School of Computer Science and Communication (CSC), Robotics, perception and learning, RPL.ORCID iD: 0000-0002-5750-9655
2016 (English)In: Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), Springer, 2016, p. 210-226Conference paper, Published paper (Refereed)
Abstract [en]

In this paper, we present the Inter-Battery Topic Model (IBTM). Our approach extends traditional topic models by learning a factorized latent variable representation. The structured representation leads to a model that marries benefits traditionally associated with a discriminative approach, such as feature selection, with those of a generative model, such as principled regularization and ability to handle missing data. The factorization is provided by representing data in terms of aligned pairs of observations as different views. This provides means for selecting a representation that separately models topics that exist in both views from the topics that are unique to a single view. This structured consolidation allows for efficient and robust inference and provides a compact and efficient representation. Learning is performed in a Bayesian fashion by maximizing a rigorous bound on the log-likelihood. Firstly, we illustrate the benefits of the model on a synthetic dataset. The model is then evaluated in both uni- and multi-modality settings on two different classification tasks with off-the-shelf convolutional neural network (CNN) features which generate state-of-the-art results with extremely compact representations.

Place, publisher, year, edition, pages
Springer, 2016. p. 210-226
Keywords [en]
CNN feature, Factorized representation, Image classification, Multi-view model, Topic model, Computer vision, Electric batteries, Neural networks, Classification tasks, Compact representation, Convolutional neural network, Discriminative approach, Multi-view modeling, Topic Modeling
National Category
Computer graphics and computer vision
Identifiers
URN: urn:nbn:se:kth:diva-195545DOI: 10.1007/978-3-319-46484-8_13ISI: 000389500600013Scopus ID: 2-s2.0-84990026488ISBN: 9783319464831 (print)OAI: oai:DiVA.org:kth-195545DiVA, id: diva2:1048629
Conference
European Conference on Computer Vision (ECCV)
Note

QC 20161121

Available from: 2016-11-21 Created: 2016-11-03 Last updated: 2025-02-07Bibliographically approved

Open Access in DiVA

fulltext(1177 kB)309 downloads
File information
File name FULLTEXT01.pdfFile size 1177 kBChecksum SHA-512
866e5cb3ccfa09418d856a15cdeb5b197096c1713f9dfe5ae249a83691e22e09bfe0156c3ac10905f1ca480ef39eafc089607d1a969a636d0832a8c61c07710c
Type fulltextMimetype application/pdf

Other links

Publisher's full textScopus

Authority records

Zhang, ChengKjellström, Hedvig

Search in DiVA

By author/editor
Zhang, ChengKjellström, Hedvig
By organisation
Robotics, perception and learning, RPL
Computer graphics and computer vision

Search outside of DiVA

GoogleGoogle Scholar
Total: 309 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 213 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf