kth.sePublications
Change search
Link to record
Permanent link

Direct link
Alternative names
Publications (10 of 15) Show all publications
Liang, X., Javid, A. M., Skoglund, M. & Chatterjee, S. (2022). Decentralized learning of randomization-based neural networks with centralized equivalence. Applied Soft Computing, 115, Article ID 108030.
Open this publication in new window or tab >>Decentralized learning of randomization-based neural networks with centralized equivalence
2022 (English)In: Applied Soft Computing, ISSN 1568-4946, E-ISSN 1872-9681, Vol. 115, article id 108030Article in journal (Refereed) Published
Abstract [en]

We consider a decentralized learning problem where training data samples are distributed over agents (processing nodes) of an underlying communication network topology without any central (master) node. Due to information privacy and security issues in a decentralized setup, nodes are not allowed to share their training data and only parameters of the neural network are allowed to be shared. This article investigates decentralized learning of randomization-based neural networks that provides centralized equivalent performance as if the full training data are available at a single node. We consider five randomization-based neural networks that use convex optimization for learning. Two of the five neural networks are shallow, and the others are deep. The use of convex optimization is the key to apply alternating-direction-method-of-multipliers with decentralized average consensus. This helps us to establish decentralized learning with centralized equivalence. For the underlying communication network topology, we use a doubly-stochastic network policy matrix and synchronous communications. Experiments with nine benchmark datasets show that the five neural networks provide good performance while requiring low computational and communication complexity for decentralized learning. The performance rankings of five neural networks using Friedman rank are also enclosed in the results, which are ELM < RVFL< dRVFL < edRVFL < SSFN.

Place, publisher, year, edition, pages
Elsevier BV, 2022
Keywords
Randomized neural network, Distributed learning, Multi-layer feedforward neural network, Alternating direction method of multipliers
National Category
Telecommunications
Identifiers
urn:nbn:se:kth:diva-307316 (URN)10.1016/j.asoc.2021.108030 (DOI)000736977500005 ()2-s2.0-85120883070 (Scopus ID)
Note

QC 20220120

Available from: 2022-01-20 Created: 2022-01-20 Last updated: 2022-06-25Bibliographically approved
Das, S., Javid, A. M., Borpatra Gohain, P., Eldar, Y. C. & Chatterjee, S. (2022). Neural Greedy Pursuit for Feature Selection. In: 2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN): . Paper presented at IEEE International Conference on Fuzzy Systems (FUZZ-IEEE) / IEEE World Congress on Computational Intelligence (IEEE WCCI) / International Joint Conference on Neural Networks (IJCNN) / IEEE Congress on Evolutionary Computation (IEEE CEC), JUL 18-23, 2022, Padua, ITALY. Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Neural Greedy Pursuit for Feature Selection
Show others...
2022 (English)In: 2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), Institute of Electrical and Electronics Engineers (IEEE) , 2022Conference paper, Published paper (Refereed)
Abstract [en]

We propose a greedy algorithm to select N important features among P input features for a non-linear prediction problem. The features are selected one by one sequentially, in an iterative loss minimization procedure. We use neural networks as predictors in the algorithm to compute the loss and hence, we refer to our method as neural greedy pursuit (NGP). NGP is efficient in selecting N features when N << P, and it provides a notion of feature importance in a descending order following the sequential selection procedure. We experimentally show that NGP provides better performance than several feature selection methods such as DeepLIFT and Drop-one-out loss. In addition, we experimentally show a phase transition behavior in which perfect selection of all N features without false positives is possible when the training data size exceeds a threshold.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2022
Series
IEEE International Joint Conference on Neural Networks (IJCNN), ISSN 2161-4393
Keywords
Feature selection, Deep learning
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-323022 (URN)10.1109/IJCNN55064.2022.9892946 (DOI)000867070908056 ()2-s2.0-85140774694 (Scopus ID)
Conference
IEEE International Conference on Fuzzy Systems (FUZZ-IEEE) / IEEE World Congress on Computational Intelligence (IEEE WCCI) / International Joint Conference on Neural Networks (IJCNN) / IEEE Congress on Evolutionary Computation (IEEE CEC), JUL 18-23, 2022, Padua, ITALY
Note

Part of proceedings: ISBN 978-1-7281-8671-9

QC 20230112

Available from: 2023-01-12 Created: 2023-01-12 Last updated: 2023-01-12Bibliographically approved
Javid, A. M., Das, S., Skoglund, M. & Chatterjee, S. (2021). A Relu Dense Layer To Improve The Performance Of Neural Networks. In: 2021 IEEE International Conference On Acoustics, Speech And Signal Processing (ICASSP 2021): . Paper presented at IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), JUN 06-11, 2021, ELECTR NETWORK (pp. 2810-2814). Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>A Relu Dense Layer To Improve The Performance Of Neural Networks
2021 (English)In: 2021 IEEE International Conference On Acoustics, Speech And Signal Processing (ICASSP 2021), Institute of Electrical and Electronics Engineers (IEEE) , 2021, p. 2810-2814Conference paper, Published paper (Refereed)
Abstract [en]

We propose ReDense as a simple and low complexity way to improve the performance of trained neural networks. We use a combination of random weights and rectified linear unit (ReLU) activation function to add a ReLU dense (ReDense) layer to the trained neural network such that it can achieve a lower training loss. The lossless flow property (LFP) of ReLU is the key to achieve the lower training loss while keeping the generalization error small. ReDense does not suffer from vanishing gradient problem in the training due to having a shallow structure. We experimentally show that ReDense can improve the training and testing performance of various neural network architectures with different optimization loss and activation functions. Finally, we test ReDense on some of the state-of-the-art architectures and show the performance improvement on benchmark.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2021
Keywords
Rectified linear unit, random weights, deep neural network
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-305410 (URN)10.1109/ICASSP39728.2021.9414269 (DOI)000704288403013 ()2-s2.0-85115078893 (Scopus ID)
Conference
IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), JUN 06-11, 2021, ELECTR NETWORK
Note

Part of proceedings: ISBN 978-1-7281-7605-5, QC 20230118

Available from: 2021-12-01 Created: 2021-12-01 Last updated: 2023-01-18Bibliographically approved
Liang, X., Javid, A. M., Skoglund, M. & Chatterjee, S. (2021). Asynchronous Decentralized Learning of Randomization-based Neural Networks. In: : . Paper presented at International Joint Conference on Neural Networks (IJCNN).
Open this publication in new window or tab >>Asynchronous Decentralized Learning of Randomization-based Neural Networks
2021 (English)Conference paper, Published paper (Refereed)
Abstract [en]

In a communication network, decentralized learning refers to the knowledge collaboration between the different local agents (processing nodes) to improve the local estimation performance without sharing private data. The ideal case is that the decentralized solution approximates the centralized solution, as if all the data are available at a single node, and requires low computational power and communication overhead. In this work, we propose a decentralized learning of randomization-based neural networks with asynchronous communication and achieve centralized equivalent performance. We propose an ARock-based alternating-direction-method-of-multipliers (ADMM) algorithm that enables individual node activation and one-sided communication in an undirected connected network, characterized by a doubly-stochastic network policy matrix. Besides, the proposed algorithm reduces the computational cost and communication overhead due to its asynchronous nature. We study the proposed algorithm on different randomization-based neural networks, including ELM, SSFN, RVFL, and its variants, to achieve the centralized equivalent performance under efficient computation and communication costs. We also show that the proposed asynchronous decentralized learning algorithm can outperform a synchronous learning algorithm regarding computational complexity, especially when the network connections are sparse.

Keywords
decentralized learning, neural networks, asynchronous communication, ADMM
National Category
Other Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
urn:nbn:se:kth:diva-295431 (URN)10.1109/IJCNN52387.2021.9533574 (DOI)000722581702035 ()2-s2.0-85116479449 (Scopus ID)
Conference
International Joint Conference on Neural Networks (IJCNN)
Note

QC 20210520

Available from: 2021-05-20 Created: 2021-05-20 Last updated: 2022-09-23Bibliographically approved
Liang, X., Javid, A. M., Skoglund, M. & Chatterjee, S. (2021). Learning without Forgetting for Decentralized Neural Nets with Low Communication Overhead. In: 2020 28th European Signal Processing Conference (EUSIPCO): . Paper presented at 28th European Signal Processing Conference (EUSIPCO), Amsterdam (pp. 2185-2189). Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Learning without Forgetting for Decentralized Neural Nets with Low Communication Overhead
2021 (English)In: 2020 28th European Signal Processing Conference (EUSIPCO), Institute of Electrical and Electronics Engineers (IEEE) , 2021, p. 2185-2189Conference paper, Published paper (Refereed)
Abstract [en]

We consider the problem of training a neural net over a decentralized scenario with a low communication over-head. The problem is addressed by adapting a recently proposed incremental learning approach, called `learning without forgetting'. While an incremental learning approach assumes data availability in a sequence, nodes of the decentralized scenario can not share data between them and there is no master node. Nodes can communicate information about model parameters among neighbors. Communication of model parameters is the key to adapt the `learning without forgetting' approach to the decentralized scenario. We use random walk based communication to handle a highly limited communication resource.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2021
Keywords
Decentralized learning, feedforward neural net, learning without forgetting, low communication overhead
National Category
Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
urn:nbn:se:kth:diva-295432 (URN)10.23919/Eusipco47968.2020.9287777 (DOI)000632622300440 ()2-s2.0-85099303579 (Scopus ID)
Conference
28th European Signal Processing Conference (EUSIPCO), Amsterdam
Note

QC 20210621

Available from: 2021-05-20 Created: 2021-05-20 Last updated: 2022-06-25Bibliographically approved
Jurado, P. G., Liang, X., Javid, A. M. & Chatterjee, S. (2021). Use of Deterministic Transforms to Design Weight Matrices of a Neural Network. In: 29th European Signal Processing Conference (EUSIPCO 2021): . Paper presented at 29th European Signal Processing Conference, EUSIPCO 2021, Dublin, 23 August 2021 through 27 August 2021 (pp. 1366-1370). European Association for Signal, Speech and Image Processing (EURASIP)
Open this publication in new window or tab >>Use of Deterministic Transforms to Design Weight Matrices of a Neural Network
2021 (English)In: 29th European Signal Processing Conference (EUSIPCO 2021), European Association for Signal, Speech and Image Processing (EURASIP) , 2021, p. 1366-1370Conference paper, Published paper (Refereed)
Abstract [en]

Self size-estimating feedforward network (SSFN) is a feedforward multilayer network. For the existing SSFN, a part of each weight matrix is trained using a layer-wise convex optimization approach (a supervised training), while the other part is chosen as a random matrix instance (an unsupervised training). In this article, the use of deterministic transforms instead of random matrix instances for the SSFN weight matrices is explored. The use of deterministic transforms provides a reduction in computational complexity. The use of several deterministic transforms is investigated, such as discrete cosine transform, Hadamard transform, Hartley transform, and wavelet transforms. The choice of a deterministic transform among a set of transforms is made in an unsupervised manner. To this end, two methods based on features' statistical parameters are developed. The proposed methods help to design a neural net where deterministic transforms can vary across its layers' weight matrices. The effectiveness of the proposed approach vis-a-vis the SSFN is illustrated for object classification tasks using several benchmark datasets.

Place, publisher, year, edition, pages
European Association for Signal, Speech and Image Processing (EURASIP), 2021
Series
European Signal Processing Conference, ISSN 2076-1465
Keywords
Multilayer neural network, deterministic transforms, weight matrices
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-311283 (URN)10.23919/EUSIPCO54536.2021.9616182 (DOI)000764066600272 ()2-s2.0-85123210853 (Scopus ID)
Conference
29th European Signal Processing Conference, EUSIPCO 2021, Dublin, 23 August 2021 through 27 August 2021
Note

QC 20220422

Part of proceedings: ISBN 978-9-0827-9706-0

Available from: 2022-04-22 Created: 2022-04-22 Last updated: 2022-06-25Bibliographically approved
Liang, X., Javid, A. M., Skoglund, M. & Chatterjee, S. (2020). A Low Complexity Decentralized Neural Net with Centralized Equivalence using Layer-wise Learning. In: 2020 International joint conference on neural networks (IJCNN): . Paper presented at International Joint Conference on Neural Networks (IJCNN) held as part of the IEEE World Congress on Computational Intelligence (IEEE WCCI), JUL 19-24, 2020, ELECTR NETWORK. IEEE
Open this publication in new window or tab >>A Low Complexity Decentralized Neural Net with Centralized Equivalence using Layer-wise Learning
2020 (English)In: 2020 International joint conference on neural networks (IJCNN), IEEE , 2020Conference paper, Published paper (Refereed)
Abstract [en]

We design a low complexity decentralized learning algorithm to train a recently proposed large neural network in distributed processing nodes (workers). We assume the communication network between the workers is synchronized and can be modeled as a doubly-stochastic mixing matrix without having any master node. In our setup, the training data is distributed among the workers but is not shared in the training process due to privacy and security concerns. Using altemating-direction-method-of-multipliers (ADMM) along with a layer-wise convex optimization approach, we propose a decentralized learning algorithm which enjoys low computational complexity and communication cost among the workers. We show that it is possible to achieve equivalent learning performance as if the data is available in a single place. Finally, we experimentally illustrate the time complexity and convergence behavior of the algorithm.

Place, publisher, year, edition, pages
IEEE, 2020
Series
IEEE International Joint Conference on Neural Networks (IJCNN), ISSN 2161-4393
Keywords
decentralized learning, neural network, ADMM, communication network
National Category
Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
urn:nbn:se:kth:diva-292968 (URN)10.1109/IJCNN48605.2020.9206592 (DOI)000626021400002 ()2-s2.0-85093843749 (Scopus ID)
Conference
International Joint Conference on Neural Networks (IJCNN) held as part of the IEEE World Congress on Computational Intelligence (IEEE WCCI), JUL 19-24, 2020, ELECTR NETWORK
Note

QC 20210419

Available from: 2021-04-19 Created: 2021-04-19 Last updated: 2023-04-05Bibliographically approved
Javid, A. M., Liang, X., Skoglund, M. & Chatterjee, S. (2020). Adaptive Learning without Forgetting via Low-Complexity Convex Networks. In: 28th European Signal Processing Conference (EUSIPCO 2020): . Paper presented at 28th European Signal Processing Conference (EUSIPCO), JAN 18-22, 2021, ELECTR NETWORK (pp. 1623-1627). Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Adaptive Learning without Forgetting via Low-Complexity Convex Networks
2020 (English)In: 28th European Signal Processing Conference (EUSIPCO 2020), Institute of Electrical and Electronics Engineers (IEEE) , 2020, p. 1623-1627Conference paper, Published paper (Refereed)
Abstract [en]

We study the problem of learning without forgetting (LwF) in which a deep learning model learns new tasks without a significant drop in the classification performance on the previously learned tasks. We propose an LwF algorithm for multilayer feedforward neural networks in which we can adapt the number of layers of the network from the old task to the new task. To this end, we limit ourselves to convex loss functions in order to train the network in a layer-wise manner. Layer-wise convex optimization leads to low-computational complexity and provides a more interpretable understanding of the network. We compare the effectiveness of the proposed adaptive LwF algorithm with the standard LwF over image classification datasets.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2020
Series
European Signal Processing Conference, ISSN 2076-1465
Keywords
learning without forgetting, convex neural networks, size adaptive, low complexity
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-295259 (URN)10.23919/Eusipco47968.2020.9287632 (DOI)000632622300327 ()2-s2.0-85099319352 (Scopus ID)
Conference
28th European Signal Processing Conference (EUSIPCO), JAN 18-22, 2021, ELECTR NETWORK
Note

QC 20210621

Available from: 2021-06-03 Created: 2021-06-03 Last updated: 2023-04-05Bibliographically approved
Liang, X., Javid, A. M., Skoglund, M. & Chatterjee, S. (2020). Asynchrounous decentralized learning of a neural network. In: Proceedings IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2020: . Paper presented at 2020 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2020, Barcelona, Spain, May 4-8, 2020 (pp. 3947-3951). Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Asynchrounous decentralized learning of a neural network
2020 (English)In: Proceedings IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2020, Institute of Electrical and Electronics Engineers (IEEE) , 2020, p. 3947-3951Conference paper, Published paper (Refereed)
Abstract [en]

In this work, we exploit an asynchronous computing framework namely ARock to learn a deep neural network called self-size estimating feedforward neural network (SSFN) in a decentralized scenario. Using this algorithm namely asynchronous decentralized SSFN (dSSFN), we provide the centralized equivalent solution under certain technical assumptions. Asynchronous dSSFN relaxes the communication bottleneck by allowing one node activation and one side communication, which reduces the communication overhead significantly, consequently increasing the learning speed. We compare asynchronous dSSFN with traditional synchronous dSSFN in the experimental results, which shows the competitive performance of asynchronous dSSFN, especially when the communication network is sparse.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2020
Series
International Conference on Acoustics Speech and Signal Processing ICASSP, ISSN 1520-6149
Keywords
Asynchronous, decentralized learning, neural networks, convex optimization
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-292015 (URN)10.1109/ICASSP40776.2020.9053996 (DOI)000615970404039 ()2-s2.0-85089210003 (Scopus ID)
Conference
2020 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2020, Barcelona, Spain, May 4-8, 2020
Note

QC 20210324

Available from: 2021-03-24 Created: 2021-03-24 Last updated: 2022-06-25Bibliographically approved
Tsai, F., Javid, A. M. & Chatterjee, S. (2020). Design of a Non-negative Neural Network to Improve on NMF. In: 28thEuropean Signal Processing Conference (EUSIPCO 2020): . Paper presented at 28th European Signal Processing Conference (EUSIPCO), JAN 18-22, 2021, ELECTR NETWORK (pp. 461-465). Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Design of a Non-negative Neural Network to Improve on NMF
2020 (English)In: 28thEuropean Signal Processing Conference (EUSIPCO 2020), Institute of Electrical and Electronics Engineers (IEEE) , 2020, p. 461-465Conference paper, Published paper (Refereed)
Abstract [en]

For prediction of a non-negative target signal using a non-negative input, we design a feed-forward neural network to achieve a better performance than a non-negative matrix factorization (NMF) algorithm. We provide a mathematical relation between the neural network and NMF. The architecture of the neural network is built on a property of rectified-linearunit (ReLU) activation function and a convex optimization layerwise training approach. For an illustrative example, we choose a speech enhancement application where a clean speech spectrum is estimated from a noisy spectrum.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2020
Series
European Signal Processing Conference, ISSN 2076-1465
Keywords
Neural networks, non-negative matrix factorization, speech enhancement
National Category
Signal Processing
Identifiers
urn:nbn:se:kth:diva-295263 (URN)10.23919/Eusipco47968.2020.9287668 (DOI)000632622300093 ()2-s2.0-85099314228 (Scopus ID)
Conference
28th European Signal Processing Conference (EUSIPCO), JAN 18-22, 2021, ELECTR NETWORK
Note

QC 20210621

Available from: 2021-06-03 Created: 2021-06-03 Last updated: 2023-04-05Bibliographically approved
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0002-8534-7622

Search in DiVA

Show all publications