kth.sePublications
Change search
Link to record
Permanent link

Direct link
Alternative names
Publications (10 of 14) Show all publications
Liang, X., Cumlin, F., Ungureanu, V., Reddy, C. K. A., Schuldt, C. & Chatterjee, S. (2024). DeePMOS-B: Deep Posterior Mean-Opinion-Score using Beta Distribution. In: 32ND EUROPEAN SIGNAL PROCESSING CONFERENCE, EUSIPCO 2024: . Paper presented at 32nd European Signal Processing Conference (EUSIPCO), AUG 26-30, 2024, Lyon, FRANCE (pp. 416-420). Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>DeePMOS-B: Deep Posterior Mean-Opinion-Score using Beta Distribution
Show others...
2024 (English)In: 32ND EUROPEAN SIGNAL PROCESSING CONFERENCE, EUSIPCO 2024, Institute of Electrical and Electronics Engineers (IEEE), 2024, p. 416-420Conference paper, Published paper (Refereed)
Abstract [en]

Mean opinion score (MOS) is a bounded speech quality measure, ranging between 1 and 5. We propose using a Beta distribution to model the posterior of the bounded MOS for a given speech clip. We use a deep neural network (DNN), trained using a maximum-likelihood principle, providing the parameters of the posterior Beta distribution. A self-teacher learning setup is used to achieve robustness against the inherent challenge of training on a noisy dataset. The dataset noise comes from the subjective nature of the MOS labels, and only a handful of quality score ratings are provided for each speech clip. To compare with existing state-of-the-art methods, we use the mean of Beta posterior as a point estimate of the MOS. The proposed method shows competitive performance vis-a-vis several existing DNN-based methods that provide MOS point estimates, and an ablation study shows the importance of various components of the proposed method.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2024
Series
European Signal Processing Conference, ISSN 2076-1465
Keywords
speech quality assessment, deep neural network, maximum-likelihood, Bayesian estimation
National Category
Signal Processing
Identifiers
urn:nbn:se:kth:diva-358710 (URN)001349787000083 ()
Conference
32nd European Signal Processing Conference (EUSIPCO), AUG 26-30, 2024, Lyon, FRANCE
Note

Part of ISBN 978-9-4645-9361-7, 979-8-3315-1977-3

QC 20250121

Available from: 2025-01-21 Created: 2025-01-21 Last updated: 2025-01-21Bibliographically approved
Liang, X. & Ma, X. (2023). AVIATOR: fAst Visual Perception and Analytics for Drone-Based Traffic Operations. In: 2023 IEEE 26th International Conference on Intelligent Transportation Systems, ITSC 2023: . Paper presented at 26th IEEE International Conference on Intelligent Transportation Systems, ITSC 2023, Bilbao, Spain, Sep 24 2023 - Sep 28 2023 (pp. 2959-2964). Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>AVIATOR: fAst Visual Perception and Analytics for Drone-Based Traffic Operations
2023 (English)In: 2023 IEEE 26th International Conference on Intelligent Transportation Systems, ITSC 2023, Institute of Electrical and Electronics Engineers (IEEE) , 2023, p. 2959-2964Conference paper, Published paper (Refereed)
Abstract [en]

Drone-based system is an emerging technology for advanced applications in Intelligent Transport Systems (ITS). This paper presents our latest developments of a visual perception and analysis system, called AVIATOR, for drone-based road traffic management. The system advances from the previous SeeFar system in several aspects. For visual perception, deep-learning based computer vision models still play the central role but the current system development focuses on fast and efficient detection and tracking performance during real-time image processing. To achieve that, YOLOv7 and ByteTrack models have replaced the previous perception modules to gain better computational performance. Meanwhile, a lane-based traffic steam detection module is added for recognizing detailed traffic flow per lane, enabling more detailed estimation of traffic flow patterns. The traffic analytics module has been modified to estimate traffic states using lane-based data collection. This includes detailed lane-based traffic flow counting as well as traffic density estimation according to vehicle arrival patterns per lane.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2023
Series
IEEE Conference on Intelligent Transportation Systems, Proceedings, ITSC, ISSN 2153-0009
National Category
Transport Systems and Logistics
Identifiers
urn:nbn:se:kth:diva-344359 (URN)10.1109/ITSC57777.2023.10422260 (DOI)2-s2.0-85186513153 (Scopus ID)
Conference
26th IEEE International Conference on Intelligent Transportation Systems, ITSC 2023, Bilbao, Spain, Sep 24 2023 - Sep 28 2023
Note

QC 20240314

Part of ISBN 979-835039946-2

Available from: 2024-03-13 Created: 2024-03-13 Last updated: 2024-03-14Bibliographically approved
Liang, X., Cumlin, F., Schüldt, C. & Chatterjee, S. (2023). DeePMOS: Deep Posterior Mean-Opinion-Score of Speech. In: Interspeech 2023: . Paper presented at 24th International Speech Communication Association, Interspeech 2023, Dublin, Ireland, Aug 20 2023 - Aug 24 2023 (pp. 526-530). International Speech Communication Association
Open this publication in new window or tab >>DeePMOS: Deep Posterior Mean-Opinion-Score of Speech
2023 (English)In: Interspeech 2023, International Speech Communication Association , 2023, p. 526-530Conference paper, Published paper (Refereed)
Abstract [en]

We propose a deep neural network (DNN) based method that provides a posterior distribution of mean-opinion-score (MOS) for an input speech signal. The DNN outputs parameters of the posterior, mainly the posterior's mean and variance. The proposed method is referred to as deep posterior MOS (DeePMOS). The relevant training data is inherently limited in size (limited number of labeled samples) and noisy due to the subjective nature of human listeners. For robust training of DeePMOS, we use a combination of maximum-likelihood learning, stochastic gradient noise, and a student-teacher learning setup. Using the mean of the posterior as a point estimate, we evaluate standard performance measures of the proposed DeePMOS. The results show comparable performance with existing DNN-based methods that only provide point estimates of the MOS. Then we provide an ablation study showing the importance of various components in DeePMOS.

Place, publisher, year, edition, pages
International Speech Communication Association, 2023
Keywords
deep neural network, maximum-likelihood, Speech quality assessment, voice conversion challenge
National Category
Signal Processing
Identifiers
urn:nbn:se:kth:diva-337876 (URN)10.21437/Interspeech.2023-1436 (DOI)001186650300107 ()2-s2.0-85171537160 (Scopus ID)
Conference
24th International Speech Communication Association, Interspeech 2023, Dublin, Ireland, Aug 20 2023 - Aug 24 2023
Note

QC 20231010

Available from: 2023-10-10 Created: 2023-10-10 Last updated: 2024-07-05Bibliographically approved
Liang, X., Javid, A. M., Skoglund, M. & Chatterjee, S. (2022). Decentralized learning of randomization-based neural networks with centralized equivalence. Applied Soft Computing, 115, Article ID 108030.
Open this publication in new window or tab >>Decentralized learning of randomization-based neural networks with centralized equivalence
2022 (English)In: Applied Soft Computing, ISSN 1568-4946, E-ISSN 1872-9681, Vol. 115, article id 108030Article in journal (Refereed) Published
Abstract [en]

We consider a decentralized learning problem where training data samples are distributed over agents (processing nodes) of an underlying communication network topology without any central (master) node. Due to information privacy and security issues in a decentralized setup, nodes are not allowed to share their training data and only parameters of the neural network are allowed to be shared. This article investigates decentralized learning of randomization-based neural networks that provides centralized equivalent performance as if the full training data are available at a single node. We consider five randomization-based neural networks that use convex optimization for learning. Two of the five neural networks are shallow, and the others are deep. The use of convex optimization is the key to apply alternating-direction-method-of-multipliers with decentralized average consensus. This helps us to establish decentralized learning with centralized equivalence. For the underlying communication network topology, we use a doubly-stochastic network policy matrix and synchronous communications. Experiments with nine benchmark datasets show that the five neural networks provide good performance while requiring low computational and communication complexity for decentralized learning. The performance rankings of five neural networks using Friedman rank are also enclosed in the results, which are ELM < RVFL< dRVFL < edRVFL < SSFN.

Place, publisher, year, edition, pages
Elsevier BV, 2022
Keywords
Randomized neural network, Distributed learning, Multi-layer feedforward neural network, Alternating direction method of multipliers
National Category
Telecommunications
Identifiers
urn:nbn:se:kth:diva-307316 (URN)10.1016/j.asoc.2021.108030 (DOI)000736977500005 ()2-s2.0-85120883070 (Scopus ID)
Note

QC 20220120

Available from: 2022-01-20 Created: 2022-01-20 Last updated: 2022-06-25Bibliographically approved
Ma, X., Liang, X., Ning, M. & Radu, A. (2022). METRIC: Toward a Drone-based Cyber-Physical Traffic Management System. In: Conference Proceedings - IEEE International Conference on Systems, Man and Cybernetics: . Paper presented at 2022 IEEE International Conference on Systems, Man, and Cybernetics, SMC 2022, Prague, Czech Republic, 9-12 October 2022 (pp. 3324-3329). Institute of Electrical and Electronics Engineers (IEEE), 2022-October
Open this publication in new window or tab >>METRIC: Toward a Drone-based Cyber-Physical Traffic Management System
2022 (English)In: Conference Proceedings - IEEE International Conference on Systems, Man and Cybernetics, Institute of Electrical and Electronics Engineers (IEEE) , 2022, Vol. 2022-October, p. 3324-3329Conference paper, Published paper (Refereed)
Abstract [en]

Drone-based system has a big potential to be applied for traffic monitoring and other advanced applications in Intelligent Transport Systems (ITS). This paper introduces our latest efforts of digitalising road traffic by various types of sensing systems, among which visual detection by drones provides a promising technical solution. A platform, called METRIC, is under recent development to carry out real-time traffic measurement and prediction using drone-based data collection. The current system is designed as a cyber-physical system (CPS) with essential functions aiming for visual traffic detection and analysis, real-time traffic estimation and prediction as well as decision supports based on simulation. In addition to the computer vision functions developed in the earlier stage, this paper also presents the CPS system architecture and the current implementation of the drone front-end system and a simulation-based system being used for further drone operations.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2022
National Category
Robotics and automation Transport Systems and Logistics
Identifiers
urn:nbn:se:kth:diva-329627 (URN)10.1109/SMC53654.2022.9945433 (DOI)2-s2.0-85142738221 (Scopus ID)
Conference
2022 IEEE International Conference on Systems, Man, and Cybernetics, SMC 2022, Prague, Czech Republic, 9-12 October 2022
Note

QC 20230622

Available from: 2023-06-22 Created: 2023-06-22 Last updated: 2025-02-05Bibliographically approved
Liang, X., Javid, A. M., Skoglund, M. & Chatterjee, S. (2021). Asynchronous Decentralized Learning of Randomization-based Neural Networks. In: : . Paper presented at International Joint Conference on Neural Networks (IJCNN).
Open this publication in new window or tab >>Asynchronous Decentralized Learning of Randomization-based Neural Networks
2021 (English)Conference paper, Published paper (Refereed)
Abstract [en]

In a communication network, decentralized learning refers to the knowledge collaboration between the different local agents (processing nodes) to improve the local estimation performance without sharing private data. The ideal case is that the decentralized solution approximates the centralized solution, as if all the data are available at a single node, and requires low computational power and communication overhead. In this work, we propose a decentralized learning of randomization-based neural networks with asynchronous communication and achieve centralized equivalent performance. We propose an ARock-based alternating-direction-method-of-multipliers (ADMM) algorithm that enables individual node activation and one-sided communication in an undirected connected network, characterized by a doubly-stochastic network policy matrix. Besides, the proposed algorithm reduces the computational cost and communication overhead due to its asynchronous nature. We study the proposed algorithm on different randomization-based neural networks, including ELM, SSFN, RVFL, and its variants, to achieve the centralized equivalent performance under efficient computation and communication costs. We also show that the proposed asynchronous decentralized learning algorithm can outperform a synchronous learning algorithm regarding computational complexity, especially when the network connections are sparse.

Keywords
decentralized learning, neural networks, asynchronous communication, ADMM
National Category
Other Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
urn:nbn:se:kth:diva-295431 (URN)10.1109/IJCNN52387.2021.9533574 (DOI)000722581702035 ()2-s2.0-85116479449 (Scopus ID)
Conference
International Joint Conference on Neural Networks (IJCNN)
Note

QC 20210520

Available from: 2021-05-20 Created: 2021-05-20 Last updated: 2022-09-23Bibliographically approved
Liang, X., Skoglund, M. & Chatterjee, S. (2021). Feature Reuse For A Randomization Based Neural Network. In: 2021 Ieee International Conference On Acoustics, Speech And Signal Processing (ICASSP 2021): . Paper presented at IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), JUN 06-11, 2021, ELECTR NETWORK (pp. 2805-2809). Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Feature Reuse For A Randomization Based Neural Network
2021 (English)In: 2021 Ieee International Conference On Acoustics, Speech And Signal Processing (ICASSP 2021), Institute of Electrical and Electronics Engineers (IEEE) , 2021, p. 2805-2809Conference paper, Published paper (Refereed)
Abstract [en]

We propose a feature reuse approach for an existing multi-layer randomization based feedforward neural network. The feature representation is directly linked among all the necessary hidden layers. For the feature reuse at a particular layer, we concatenate features from the previous layers to construct a large-dimensional feature for the layer. The large-dimensional concatenated feature is then efficiently used to learn a limited number of parameters by solving a convex optimization problem. Experiments show that the proposed model improves the performance in comparison with the original neural network without a significant increase in computational complexity.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2021
Keywords
Multi-layer neural network, randomization based neural network, convex optimization, feature reuse
National Category
Telecommunications
Identifiers
urn:nbn:se:kth:diva-305415 (URN)10.1109/ICASSP39728.2021.9413424 (DOI)000704288403012 ()2-s2.0-85114863008 (Scopus ID)
Conference
IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), JUN 06-11, 2021, ELECTR NETWORK
Note

Part of proceedings: ISBN 978-1-7281-7605-5, QC 20230118

Available from: 2021-12-01 Created: 2021-12-01 Last updated: 2023-01-18Bibliographically approved
Liang, X., Javid, A. M., Skoglund, M. & Chatterjee, S. (2021). Learning without Forgetting for Decentralized Neural Nets with Low Communication Overhead. In: 2020 28th European Signal Processing Conference (EUSIPCO): . Paper presented at 28th European Signal Processing Conference (EUSIPCO), Amsterdam (pp. 2185-2189). Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Learning without Forgetting for Decentralized Neural Nets with Low Communication Overhead
2021 (English)In: 2020 28th European Signal Processing Conference (EUSIPCO), Institute of Electrical and Electronics Engineers (IEEE) , 2021, p. 2185-2189Conference paper, Published paper (Refereed)
Abstract [en]

We consider the problem of training a neural net over a decentralized scenario with a low communication over-head. The problem is addressed by adapting a recently proposed incremental learning approach, called `learning without forgetting'. While an incremental learning approach assumes data availability in a sequence, nodes of the decentralized scenario can not share data between them and there is no master node. Nodes can communicate information about model parameters among neighbors. Communication of model parameters is the key to adapt the `learning without forgetting' approach to the decentralized scenario. We use random walk based communication to handle a highly limited communication resource.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2021
Keywords
Decentralized learning, feedforward neural net, learning without forgetting, low communication overhead
National Category
Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
urn:nbn:se:kth:diva-295432 (URN)10.23919/Eusipco47968.2020.9287777 (DOI)000632622300440 ()2-s2.0-85099303579 (Scopus ID)
Conference
28th European Signal Processing Conference (EUSIPCO), Amsterdam
Note

QC 20210621

Available from: 2021-05-20 Created: 2021-05-20 Last updated: 2022-06-25Bibliographically approved
Jurado, P. G., Liang, X., Javid, A. M. & Chatterjee, S. (2021). Use of Deterministic Transforms to Design Weight Matrices of a Neural Network. In: 29th European Signal Processing Conference (EUSIPCO 2021): . Paper presented at 29th European Signal Processing Conference, EUSIPCO 2021, Dublin, 23 August 2021 through 27 August 2021 (pp. 1366-1370). European Association for Signal, Speech and Image Processing (EURASIP)
Open this publication in new window or tab >>Use of Deterministic Transforms to Design Weight Matrices of a Neural Network
2021 (English)In: 29th European Signal Processing Conference (EUSIPCO 2021), European Association for Signal, Speech and Image Processing (EURASIP) , 2021, p. 1366-1370Conference paper, Published paper (Refereed)
Abstract [en]

Self size-estimating feedforward network (SSFN) is a feedforward multilayer network. For the existing SSFN, a part of each weight matrix is trained using a layer-wise convex optimization approach (a supervised training), while the other part is chosen as a random matrix instance (an unsupervised training). In this article, the use of deterministic transforms instead of random matrix instances for the SSFN weight matrices is explored. The use of deterministic transforms provides a reduction in computational complexity. The use of several deterministic transforms is investigated, such as discrete cosine transform, Hadamard transform, Hartley transform, and wavelet transforms. The choice of a deterministic transform among a set of transforms is made in an unsupervised manner. To this end, two methods based on features' statistical parameters are developed. The proposed methods help to design a neural net where deterministic transforms can vary across its layers' weight matrices. The effectiveness of the proposed approach vis-a-vis the SSFN is illustrated for object classification tasks using several benchmark datasets.

Place, publisher, year, edition, pages
European Association for Signal, Speech and Image Processing (EURASIP), 2021
Series
European Signal Processing Conference, ISSN 2076-1465
Keywords
Multilayer neural network, deterministic transforms, weight matrices
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-311283 (URN)10.23919/EUSIPCO54536.2021.9616182 (DOI)000764066600272 ()2-s2.0-85123210853 (Scopus ID)
Conference
29th European Signal Processing Conference, EUSIPCO 2021, Dublin, 23 August 2021 through 27 August 2021
Note

QC 20220422

Part of proceedings: ISBN 978-9-0827-9706-0

Available from: 2022-04-22 Created: 2022-04-22 Last updated: 2022-06-25Bibliographically approved
Liang, X., Javid, A. M., Skoglund, M. & Chatterjee, S. (2020). A Low Complexity Decentralized Neural Net with Centralized Equivalence using Layer-wise Learning. In: 2020 International joint conference on neural networks (IJCNN): . Paper presented at International Joint Conference on Neural Networks (IJCNN) held as part of the IEEE World Congress on Computational Intelligence (IEEE WCCI), JUL 19-24, 2020, ELECTR NETWORK. IEEE
Open this publication in new window or tab >>A Low Complexity Decentralized Neural Net with Centralized Equivalence using Layer-wise Learning
2020 (English)In: 2020 International joint conference on neural networks (IJCNN), IEEE , 2020Conference paper, Published paper (Refereed)
Abstract [en]

We design a low complexity decentralized learning algorithm to train a recently proposed large neural network in distributed processing nodes (workers). We assume the communication network between the workers is synchronized and can be modeled as a doubly-stochastic mixing matrix without having any master node. In our setup, the training data is distributed among the workers but is not shared in the training process due to privacy and security concerns. Using altemating-direction-method-of-multipliers (ADMM) along with a layer-wise convex optimization approach, we propose a decentralized learning algorithm which enjoys low computational complexity and communication cost among the workers. We show that it is possible to achieve equivalent learning performance as if the data is available in a single place. Finally, we experimentally illustrate the time complexity and convergence behavior of the algorithm.

Place, publisher, year, edition, pages
IEEE, 2020
Series
IEEE International Joint Conference on Neural Networks (IJCNN), ISSN 2161-4393
Keywords
decentralized learning, neural network, ADMM, communication network
National Category
Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
urn:nbn:se:kth:diva-292968 (URN)10.1109/IJCNN48605.2020.9206592 (DOI)000626021400002 ()2-s2.0-85093843749 (Scopus ID)
Conference
International Joint Conference on Neural Networks (IJCNN) held as part of the IEEE World Congress on Computational Intelligence (IEEE WCCI), JUL 19-24, 2020, ELECTR NETWORK
Note

QC 20210419

Available from: 2021-04-19 Created: 2021-04-19 Last updated: 2023-04-05Bibliographically approved
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0003-4406-536X

Search in DiVA

Show all publications