kth.sePublications
Change search
Link to record
Permanent link

Direct link
Publications (10 of 12) Show all publications
Wu, H. (2020). Compression-based Data Identification and Representation Learning. (Doctoral dissertation). KTH Royal Institute of Technology
Open this publication in new window or tab >>Compression-based Data Identification and Representation Learning
2020 (English)Doctoral thesis, monograph (Other academic)
Abstract [en]

Large-scale data generation, acquisition, and processing are happening at everymoment in our society. This thesis explores the opportunities for applying lossycompression methods and concepts for improving information engineering techniquesthat transform a large amount of collected data into useful applications. Two specificapplications are investigated: data identification and representation learning.The lossy compression methods, such as product quantization and hierarchicalvector quantization, can be used to build the data structures for efficient retrievaland identification. There exists a trade-off between the rate of compressed data andgeneral retrieval performance. This thesis focus on studying this trade-off under thesimilarity identification framework. The similarity identification task is to identifythe items in a database that are similar to a given query item for a given metric. Thisthesis studies the trade-off between the rate and identifiability of the compresseddata for correlated query and source signals. Signal processing methods such asKarhunen-Loève transform and linear prediction are applied to exploit the lineardependence of the source and query signals. In addition, practical schemes basedon tree-structured vector quantizers and transform-based models are proposed forsimilarity identification.Representation learning aims to transform the real-world observations into an-other feature space that is more amiable to particular applications. Ideally, thelearned representation should only contain essential information of the original data,and irrelevant information should be removed. This thesis focuses on integratingdeep learning models with lossy compression methods and concepts to improverepresentation learning. For learning representation for large-scale image retrieval,the product quantizer is incorporated into the bottleneck stage of autoencodermodels and trained in an end-to-end fashion for image retrieval tasks. The trainedencoder neural network concatenated with a product quantizer is then used toproduce short indices for each image for fast retrieval. For improving unsuper-vised representation learning, a quantization-based regularizer is introduced tothe autoencoder-based models for fostering a similarity-preserving mapping at theencoder. It is demonstrated that the proposed regularization method results inimproved latent representations for downstream tasks such as classification andclustering. Finally, a contrastive loss based on conditional mutual information (CMI)for learning representations of time series data is proposed. An encoder is firsttrained to maximize the mutual information between the latent variables and thetrend information conditioned on the encoded observed variables. Then the featuresextracted from the trained encoder are used to learn a subsequent logistic regressionmodel for predicting time series movements. The CMI maximization problem canbe transformed into a classification problem of determining whether two encodedivrepresentations are sampled from the same class or not. It is shown that the proposedmethod is effective for improving the generalization of deep learning models thatare trained on limited data.

Place, publisher, year, edition, pages
KTH Royal Institute of Technology, 2020
Series
TRITA-EECS-AVL ; 40
National Category
Electrical Engineering, Electronic Engineering, Information Engineering
Research subject
Electrical Engineering
Identifiers
urn:nbn:se:kth:diva-280319 (URN)
Public defence
2020-10-02, F3, Lindstedtsvägen 26, Stockholm, 14:30 (English)
Opponent
Supervisors
Note

QC 20200911

Available from: 2020-09-11 Created: 2020-09-11 Last updated: 2022-06-25Bibliographically approved
Wu, H., Gattami, A. & Flierl, M. (2020). Conditional mutual information-based contrastive loss for financial time series forecasting. In: Proceedings ICAIF '20: The First ACM International Conference on AI in Finance: . Paper presented at ICAIF '20: The First ACM International Conference on AI in Finance, New York, NY, USA, October 15-16, 2020. Association for Computing Machinery (ACM)
Open this publication in new window or tab >>Conditional mutual information-based contrastive loss for financial time series forecasting
2020 (English)In: Proceedings ICAIF '20: The First ACM International Conference on AI in Finance, Association for Computing Machinery (ACM) , 2020Conference paper, Published paper (Refereed)
Abstract [en]

We present a representation learning framework for financial time series forecasting. One challenge of using deep learning models for finance forecasting is the shortage of available training data when using small datasets. Direct trend classification using deep neural networks trained on small datasets is susceptible to the overfitting problem. In this paper, we propose to first learn compact representations from time series data, then use the learned representations to train a simpler model for predicting time series movements. We consider a class-conditioned latent variable model. We train an encoder network to maximize the mutual information between the latent variables and the trend information conditioned on the encoded observed variables. We show that conditional mutual information maximization can be approximated by a contrastive loss. Then, the problem is transformed into a classification task of determining whether two encoded representations are sampled from the same class or not. This is equivalent to performing pairwise comparisons of the training datapoints, and thus, improves the generalization ability of the encoder network. We use deep autoregressive models as our encoder to capture long-term dependencies of the sequence data. Empirical experiments indicate that our proposed method has the potential to advance state-of-the-art performance.

Place, publisher, year, edition, pages
Association for Computing Machinery (ACM), 2020
Keywords
Classification (of information), Deep neural networks, Equivalence classes, Finance, Signal encoding, Time series, Compact representation, Conditional mutual information, Financial time series forecasting, Learn+, Learning frameworks, Learning models, Over fitting problem, Small data set, Time-series data, Training data, Forecasting
National Category
Computer and Information Sciences
Identifiers
urn:nbn:se:kth:diva-313547 (URN)10.1145/3383455.3422550 (DOI)2-s2.0-85095337230 (Scopus ID)
Conference
ICAIF '20: The First ACM International Conference on AI in Finance, New York, NY, USA, October 15-16, 2020
Note

Part of ISBN 9781450375849

QC 20220614

Available from: 2022-06-14 Created: 2022-06-14 Last updated: 2022-06-25Bibliographically approved
Wu, H. & Flierl, M. (2020). Vector Quantization-Based Regularization for Autoencoders. In: Thirty-fourth AAAI conference on artificial intelligence, the thirty-second innovative applications of artificial intelligence conference and the tenth AAAI symposium on educational advances in artificial intelligence: . Paper presented at 34th AAAI Conference on Artificial Intelligence / 32nd Innovative Applications of Artificial Intelligence Conference / 10th AAAI Symposium on Educational Advances in Artificial Intelligence, FEB 07-12, 2020, New York, NY (pp. 6380-6387). ASSOC ADVANCEMENT ARTIFICIAL INTELLIGENCE
Open this publication in new window or tab >>Vector Quantization-Based Regularization for Autoencoders
2020 (English)In: Thirty-fourth AAAI conference on artificial intelligence, the thirty-second innovative applications of artificial intelligence conference and the tenth AAAI symposium on educational advances in artificial intelligence, ASSOC ADVANCEMENT ARTIFICIAL INTELLIGENCE , 2020, p. 6380-6387Conference paper, Published paper (Refereed)
Abstract [en]

Autoencoders and their variations provide unsupervised models for learning low-dimensional representations for downstream tasks. Without proper regularization, autoencoder models are susceptible to the overfitting problem and the so-called posterior collapse phenomenon. In this paper, we introduce a quantization-based regularizer in the bottleneck stage of autoencoder models to learn meaningful latent representations. We combine both perspectives of Vector Quantized-Variational AutoEncoders (VQ-VAE) and classical denoising regularization methods of neural networks. We interpret quantizers as regularizers that constrain latent representations while fostering a similarity-preserving mapping at the encoder. Before quantization, we impose noise on the latent codes and use a Bayesian estimator to optimize the quantizer-based representation. The introduced bottleneck Bayesian estimator outputs the posterior mean of the centroids to the decoder, and thus, is performing soft quantization of the noisy latent codes. We show that our proposed regularization method results in improved latent representations for both supervised learning and clustering downstream tasks when compared to autoencoders using other bottleneck structures.

Place, publisher, year, edition, pages
ASSOC ADVANCEMENT ARTIFICIAL INTELLIGENCE, 2020
Series
AAAI Conference on Artificial Intelligence, ISSN 2159-5399 ; 34
National Category
Probability Theory and Statistics
Identifiers
urn:nbn:se:kth:diva-299731 (URN)000667722806057 ()2-s2.0-85106408043 (Scopus ID)
Conference
34th AAAI Conference on Artificial Intelligence / 32nd Innovative Applications of Artificial Intelligence Conference / 10th AAAI Symposium on Educational Advances in Artificial Intelligence, FEB 07-12, 2020, New York, NY
Note

QC 20210816

Available from: 2021-08-16 Created: 2021-08-16 Last updated: 2023-04-05Bibliographically approved
Wu, H. & Flierl, M. (2019). Learning product codebooks using vector-quantized autoencoders for image retrieval. In: GlobalSIP 2019 - 7th IEEE Global Conference on Signal and Information Processing, Proceedings: . Paper presented at 7th IEEE Global Conference on Signal and Information Processing, GlobalSIP 2019, 11 November 2019 through 14 November 2019. Institute of Electrical and Electronics Engineers Inc.
Open this publication in new window or tab >>Learning product codebooks using vector-quantized autoencoders for image retrieval
2019 (English)In: GlobalSIP 2019 - 7th IEEE Global Conference on Signal and Information Processing, Proceedings, Institute of Electrical and Electronics Engineers Inc. , 2019Conference paper, Published paper (Refereed)
Abstract [en]

Vector-Quantized Variational Autoencoders (VQ-VAE)[1] provide an unsupervised model for learning discrete representations by combining vector quantization and autoencoders. In this paper, we study the use of VQ-VAE for representation learning of downstream tasks, such as image retrieval. First, we describe the VQ-VAE in the context of an information-theoretic framework. Then, we show that the regularization effect on the learned representation is determined by the size of the embedded codebook before the training. As a result, we introduce a hyperparameter to balance the strength of the vector quantizer and the reconstruction error. By tuning the hyperparameter, the embedded bottleneck quantizer is used as a regularizer that forces the output of the encoder to share a constrained coding space. With that, the learned latent features better preserve the similarity relations of the data space. Finally, we incorporate the product quantizer into the bottleneck stage of VQ-VAE and use it as an end-to-end unsupervised learning model for image retrieval tasks. The product quantizer has the advantage of generating large and structured codebooks. Fast retrieval can be achieved by using lookup tables that store the distance between any pair of sub-codewords. State-of-the-art retrieval results are achieved by the proposed codebooks. 

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers Inc., 2019
Keywords
Information theory, Learning systems, Table lookup, Vector quantization, Vectors, Constrained coding, Fast retrievals, Learning products, Reconstruction error, Similarity relations, State of the art, Structured codebooks, Vector quantizers, Image retrieval
National Category
Computer graphics and computer vision
Identifiers
urn:nbn:se:kth:diva-274151 (URN)10.1109/GlobalSIP45357.2019.8969272 (DOI)000555454800086 ()2-s2.0-85079284181 (Scopus ID)
Conference
7th IEEE Global Conference on Signal and Information Processing, GlobalSIP 2019, 11 November 2019 through 14 November 2019
Note

QC 20200622

Part of ISBN 9781728127231

Available from: 2020-06-22 Created: 2020-06-22 Last updated: 2025-02-07Bibliographically approved
Wu, H. & Flierl, M. (2018). Component-based quadratic similarity identification for multivariate Gaussian sources. In: Data Compression Conference Proceedings: . Paper presented at 2018 Data Compression Conference, DCC 2018, 27 March 2018 through 30 March 2018. Institute of Electrical and Electronics Engineers Inc.
Open this publication in new window or tab >>Component-based quadratic similarity identification for multivariate Gaussian sources
2018 (English)In: Data Compression Conference Proceedings, Institute of Electrical and Electronics Engineers Inc. , 2018Conference paper, Published paper (Refereed)
Abstract [en]

This paper considers the problem of compression for similarity identification. Unlike classic compression problems, the focus is not on reconstructing the original data. Instead, compression is determined by the reliability of answering given queries. The problem is characterized by the identification rate of a source which is the minimum compression rate which allows reliable answers for a given similarity threshold. In this work, we investigate the component-based quadratic similarity identification for multivariate Gaussian sources. The decorrelated original data is processed by a distinct D- A dmissible system for each component. For a special case, we characterize the component-based identification rate for a correlated Gaussian source. Furthermore, we derived the optimal bit allocation for a given total rate constraint.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers Inc., 2018
Keywords
Bit allocation, Similarity identification, Gaussian distribution, Component based, Compression rates, Gaussian sources, Identification rates, Optimal bit allocation, Rate constraints, Similarity threshold, Data compression
National Category
Computer and Information Sciences
Identifiers
urn:nbn:se:kth:diva-238074 (URN)10.1109/DCC.2018.00086 (DOI)000540644700079 ()2-s2.0-85050969530 (Scopus ID)9781538648834 (ISBN)
Conference
2018 Data Compression Conference, DCC 2018, 27 March 2018 through 30 March 2018
Note

Conference code: 138136; Export Date: 30 October 2018; Conference Paper; CODEN: DDCCF

QC 20180114

Available from: 2019-01-14 Created: 2019-01-14 Last updated: 2022-06-26Bibliographically approved
Wu, H., Wang, Q. & Flierl, M. (2018). Identification Rates for Block-correlated Gaussian Sources. In: Matthews, M B (Ed.), 2018 CONFERENCE RECORD OF 52ND ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS, AND COMPUTERS: . Paper presented at 52nd Asilomar Conference on Signals, Systems, and Computers, OCT 28-NOV 01, 2018, Pacific Grove, CA (pp. 2114-2118). IEEE
Open this publication in new window or tab >>Identification Rates for Block-correlated Gaussian Sources
2018 (English)In: 2018 CONFERENCE RECORD OF 52ND ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS, AND COMPUTERS / [ed] Matthews, M B, IEEE , 2018, p. 2114-2118Conference paper, Published paper (Refereed)
Abstract [en]

Among many current data processing systems, the objectives are often not the reproduction of data, but to compute some answers based on the data responding to sonic queries. The similarity identification task is to identify the items in a database which are similar to a given query item regarding to a certain metric. The problem of compression for similarity identification has been studied in [1]. Unlike classic compression problems, the focus is not on reconstructing the original data. Instead, the compression rate is determined by the desired reliability of the answers. Specifically, the information measure identification rate of a compression scheme characterizes the minimum compression rate that can be achieved which guarantees reliable answers with respect to a given similarity threshold. In this paper, we study the component-based quadratic similarity identification for correlated sources. The blocks are first decorrelated by Karhunen-Loeve transform. Then, the decorrelated data is processed by a distinct D-admissible system for each component. We derive the identification rate of component-based scheme for block correlated Gaussian sources. In addition, we characterize the identification rate of a special setting where any information regarding to the component similarity thresholds is unknown while only the similarity threshold of the whole scheme is given. Furthermore, we prove that block-correlated Gaussian sources are the "most difficult" to code under the special setting.

Place, publisher, year, edition, pages
IEEE, 2018
Series
Conference Record of the Asilomar Conference on Signals Systems and Computers, ISSN 1058-6393
National Category
Computer and Information Sciences
Identifiers
urn:nbn:se:kth:diva-252677 (URN)10.1109/ACSSC.2018.8645306 (DOI)000467845100373 ()2-s2.0-85062960125 (Scopus ID)978-1-5386-9218-9 (ISBN)
Conference
52nd Asilomar Conference on Signals, Systems, and Computers, OCT 28-NOV 01, 2018, Pacific Grove, CA
Note

QC 20190603

Available from: 2019-06-03 Created: 2019-06-03 Last updated: 2022-06-26Bibliographically approved
Wu, H., Wang, Q. & Flierl, M. (2018). PREDICTION-BASED SIMILARITY IDENTIFICATION FOR AUTOREGRESSIVE PROCESSES. In: 2018 IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (GLOBALSIP 2018): . Paper presented at IEEE Global Conference on Signal and Information Processing (GlobalSIP), NOV 26-29, 2018, Anaheim, CA (pp. 266-270). IEEE
Open this publication in new window or tab >>PREDICTION-BASED SIMILARITY IDENTIFICATION FOR AUTOREGRESSIVE PROCESSES
2018 (English)In: 2018 IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (GLOBALSIP 2018), IEEE , 2018, p. 266-270Conference paper, Published paper (Refereed)
Abstract [en]

The task of similarity identification is to identify items in a database which are similar to a given query item for a given metric. The identification rate of a compression scheme characterizes the minimum rate that can be achieved which guarantees reliable answers with respect to a given similarity threshold [1]. In this paper, we study a prediction-based quadratic similarity identification for autoregressive processes. We use an ideal linear predictor to remove linear dependencies in autoregressive processes. The similarity identification is conducted on the residuals. We show that the relation between the distortion of query and database processes and the distortion of their residuals is characterized by a sequence of eigenvalues. We derive the identification rate of our prediction-based approach for autoregressive Gaussian processes. We characterize the identification rate for the special case where only the smallest value in the sequence of eigenvalues is required to be known and derive its analytical upper bound by approximating a sequence of matrices with a sequence of Toeplitz matrices.

Place, publisher, year, edition, pages
IEEE, 2018
Series
IEEE Global Conference on Signal and Information Processing, ISSN 2376-4066
National Category
Computer and Information Sciences
Identifiers
urn:nbn:se:kth:diva-249832 (URN)10.1109/GlobalSIP.2018.8646407 (DOI)000462968100054 ()2-s2.0-85063103300 (Scopus ID)978-1-7281-1295-4 (ISBN)
Conference
IEEE Global Conference on Signal and Information Processing (GlobalSIP), NOV 26-29, 2018, Anaheim, CA
Note

QC 20190423

Available from: 2019-04-23 Created: 2019-04-23 Last updated: 2022-06-26Bibliographically approved
Wu, H. & Flierl, M. (2018). Transform-based compression for quadratic similarity queries. In: Conference Record of 51st Asilomar Conference on Signals, Systems and Computers, ACSSC 2017: . Paper presented at 51st Asilomar Conference on Signals, Systems and Computers, ACSSC 2017, Asilomar Hotel and Conference Grounds, Pacific Grove, United States, 29 October 2017 through 1 November 2017 (pp. 377-381). Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Transform-based compression for quadratic similarity queries
2018 (English)In: Conference Record of 51st Asilomar Conference on Signals, Systems and Computers, ACSSC 2017, Institute of Electrical and Electronics Engineers (IEEE), 2018, p. 377-381Conference paper, Published paper (Refereed)
Abstract [en]

This paper considers the problem of compression for similarity queries [1] and discusses transform-based compression schemes. Here, the focus is on the tradeoff between the rate of the compressed data and the reliability of the answers to a given query. We consider compression schemes that do not allow false negatives when answering queries. Hence, classical compression techniques need to be modified. We propose transform-based compression schemes which decorrelate original data and regard each transform component as a distinct D-admissible system. Both compression and retrieval will be performed in the transform domain. The transform-based schemes show advantages in terms of encoding speed and the ability of handling high-dimensional correlated data. In particular, we discuss component-based and vector-based schemes. We use P{maybe}, a probability that is related to the occurrence of false positives to assess our scheme. Our experiments show that component-based schemes offer both good performance and low search complexity.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2018
Series
Conference Record of the Asilomar Conference on Signals Systems and Computers, ISSN 1058-6393
National Category
Telecommunications
Identifiers
urn:nbn:se:kth:diva-233721 (URN)10.1109/ACSSC.2017.8335363 (DOI)000442659900065 ()2-s2.0-85050980426 (Scopus ID)9781538618233 (ISBN)
Conference
51st Asilomar Conference on Signals, Systems and Computers, ACSSC 2017, Asilomar Hotel and Conference Grounds, Pacific Grove, United States, 29 October 2017 through 1 November 2017
Note

QC 20180830

Available from: 2018-08-30 Created: 2018-08-30 Last updated: 2022-06-26Bibliographically approved
Wu, H., Wang, Q. & Flierl, M. (2017). Tree-Structured Vector Quantization for Similarity Queries. In: Bilgin, A Marcellin, MW SerraSagrista, J Storer, JA (Ed.), 2017 Data Compression Conference (DCC): . Paper presented at Data Compression Conference (DCC), APR 04-07, 2017, Snowbird, UT (pp. 467-467). IEEE Computer Society
Open this publication in new window or tab >>Tree-Structured Vector Quantization for Similarity Queries
2017 (English)In: 2017 Data Compression Conference (DCC) / [ed] Bilgin, A Marcellin, MW SerraSagrista, J Storer, JA, IEEE Computer Society, 2017, p. 467-467Conference paper, Published paper (Refereed)
Place, publisher, year, edition, pages
IEEE Computer Society, 2017
Series
IEEE Data Compression Conference, ISSN 1068-0314
National Category
Communication Systems
Identifiers
urn:nbn:se:kth:diva-243540 (URN)10.1109/DCC.2017.72 (DOI)000404240300077 ()978-1-5090-6721-3 (ISBN)
Conference
Data Compression Conference (DCC), APR 04-07, 2017, Snowbird, UT
Note

QC 20190208

Available from: 2019-02-08 Created: 2019-02-08 Last updated: 2022-06-26Bibliographically approved
Wu, H., Li, H. & Flierl, M. (2016). AN EMBEDDED 3D GEOMETRY SCORE FOR MOBILE 3D VISUAL SEARCH. In: 2016 IEEE 18TH INTERNATIONAL WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING (MMSP): . Paper presented at 18th IEEE International Workshop on Multimedia Signal Processing (MMSP), SEP 21-23, 2016, Montreal, CANADA. Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>AN EMBEDDED 3D GEOMETRY SCORE FOR MOBILE 3D VISUAL SEARCH
2016 (English)In: 2016 IEEE 18TH INTERNATIONAL WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING (MMSP), Institute of Electrical and Electronics Engineers (IEEE), 2016Conference paper, Published paper (Refereed)
Abstract [en]

The scoring function is a central component in mobile visual search. In this paper, we propose an embedded 3D geometry score for mobile 3D visual search (M3DVS). In contrast to conventional mobile visual search, M3DVS uses not only the visual appearance of query objects, but utilizes also the underlying 3D geometry. The proposed scoring function interprets visual search as a process that reduces uncertainty among candidate objects when observing a query. For M3DVS, the uncertainty is reduced by both appearance-based visual similarity and 3D geometric similarity. For the latter, we give an algorithm for estimating the query-dependent threshold for geometric similarity. In contrast to visual similarity, the threshold for geometric similarity is relative due to the constraints of image-based 3D reconstruction. The experimental results show that the embedded 3D geometry score improves the recall-datarate performance when compared to a conventional visual score or 3D geometry-based re-ranking.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2016
Series
IEEE International Workshop on Multimedia Signal Processing, ISSN 2163-3517
National Category
Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
urn:nbn:se:kth:diva-202670 (URN)10.1109/MMSP.2016.7813366 (DOI)000393302100034 ()2-s2.0-85013157698 (Scopus ID)978-1-5090-3724-7 (ISBN)
Conference
18th IEEE International Workshop on Multimedia Signal Processing (MMSP), SEP 21-23, 2016, Montreal, CANADA
Note

QC 20170306

Available from: 2017-03-06 Created: 2017-03-06 Last updated: 2022-06-27Bibliographically approved
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0003-2579-2107

Search in DiVA

Show all publications