kth.sePublications
Change search
Link to record
Permanent link

Direct link
Publications (9 of 9) Show all publications
Ghosh, A., Honore, A. & Chatterjee, S. (2024). DANSE: Data-Driven Non-Linear State Estimation of Model-Free Process in Unsupervised Learning Setup. IEEE Transactions on Signal Processing, 72, 1824-1838
Open this publication in new window or tab >>DANSE: Data-Driven Non-Linear State Estimation of Model-Free Process in Unsupervised Learning Setup
2024 (English)In: IEEE Transactions on Signal Processing, ISSN 1053-587X, E-ISSN 1941-0476, Vol. 72, p. 1824-1838Article in journal (Refereed) Published
Abstract [en]

We address the tasks of Bayesian state estimation and forecasting for a model-free process in an unsupervised learning setup. For a model-free process, we do not have any a-priori knowledge of the process dynamics. In the article, we propose DANSE - a Data-driven Nonlinear State Estimation method. DANSE provides a closed-form posterior of the state of the model-free process, given linear measurements of the state. In addition, it provides a closed-form posterior for forecasting. A data-driven recurrent neural network (RNN) is used in DANSE to provide the parameters of a prior of the state. The prior depends on the past measurements as input, and then we find the closed-form posterior of the state using the current measurement as input. The data-driven RNN captures the underlying non-linear dynamics of the model-free process. The training of DANSE, mainly learning the parameters of the RNN, is executed using an unsupervised learning approach. In unsupervised learning, we have access to a training dataset comprising only a set of (noisy) measurement data trajectories, but we do not have any access to the state trajectories. Therefore, DANSE does not have access to state information in the training data and can not use supervised learning. Using simulated linear and non-linear process models (Lorenz attractor and Chen attractor), we evaluate the unsupervised learning-based DANSE. We show that the proposed DANSE, without knowledge of the process model and without supervised learning, provides a competitive performance against model-driven methods, such as the Kalman filter (KF), extended KF (EKF), unscented KF (UKF), a data-driven deep Markov model (DMM) and a recently proposed hybrid method called KalmanNet. In addition, we show that DANSE works for high-dimensional state estimation.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2024
Keywords
State estimation, Computational modeling, Training, Bayes methods, Noise measurement, Supervised learning, Unsupervised learning, Bayesian state estimation, forecasting, neural networks, recurrent neural networks
National Category
Computer Sciences
Identifiers
urn:nbn:se:kth:diva-346102 (URN)10.1109/TSP.2024.3383277 (DOI)001200035500017 ()2-s2.0-85189517590 (Scopus ID)
Note

QC 20240503

Available from: 2024-05-03 Created: 2024-05-03 Last updated: 2024-05-03Bibliographically approved
Ghosh, A., Abdalmoaty, M., Chatterjee, S. & Hjalmarsson, H. (2024). DeepBayes—An estimator for parameter estimation in stochastic nonlinear dynamical models. Automatica, 159, Article ID 111327.
Open this publication in new window or tab >>DeepBayes—An estimator for parameter estimation in stochastic nonlinear dynamical models
2024 (English)In: Automatica, ISSN 0005-1098, E-ISSN 1873-2836, Vol. 159, article id 111327Article in journal (Refereed) Published
Abstract [en]

Stochastic nonlinear dynamical systems are ubiquitous in modern, real-world applications. Yet, estimating the unknown parameters of stochastic, nonlinear dynamical models remains a challenging problem. The majority of existing methods employ maximum likelihood or Bayesian estimation. However, these methods suffer from some limitations, most notably the substantial computational time for inference coupled with limited flexibility in application. In this work, we propose DeepBayes estimators that leverage the power of deep recurrent neural networks. The method consists of first training a recurrent neural network to minimize the mean-squared estimation error over a set of synthetically generated data using models drawn from the model set of interest. The a priori trained estimator can then be used directly for inference by evaluating the network with the estimation data. The deep recurrent neural network architectures can be trained offline and ensure significant time savings during inference. We experiment with two popular recurrent neural networks — long short term memory network (LSTM) and gated recurrent unit (GRU). We demonstrate the applicability of our proposed method on different example models and perform detailed comparisons with state-of-the-art approaches. We also provide a study on a real-world nonlinear benchmark problem. The experimental evaluations show that the proposed approach is asymptotically as good as the Bayes estimator.

Place, publisher, year, edition, pages
Elsevier Ltd, 2024
Keywords
Deep learning, Dynamical systems, Nonlinear system identification, Parameter estimation, Recurrent neural networks
National Category
Control Engineering Other Electrical Engineering, Electronic Engineering, Information Engineering Probability Theory and Statistics
Identifiers
urn:nbn:se:kth:diva-339038 (URN)10.1016/j.automatica.2023.111327 (DOI)001161034600001 ()2-s2.0-85174673962 (Scopus ID)
Note

QC 20231129

Available from: 2023-11-29 Created: 2023-11-29 Last updated: 2024-05-02Bibliographically approved
Honore, A., Ghosh, A. & Chatterjee, S. (2023). Compressed Sensing of Generative Sparse-Latent (GSL) Signals. In: 31st European Signal Processing Conference, EUSIPCO 2023 - Proceedings: . Paper presented at 31st European Signal Processing Conference, EUSIPCO 2023, Helsinki, Finland, Sep 4 2023 - Sep 8 2023 (pp. 1918-1922). European Signal Processing Conference, EUSIPCO
Open this publication in new window or tab >>Compressed Sensing of Generative Sparse-Latent (GSL) Signals
2023 (English)In: 31st European Signal Processing Conference, EUSIPCO 2023 - Proceedings, European Signal Processing Conference, EUSIPCO , 2023, p. 1918-1922Conference paper, Published paper (Refereed)
Abstract [en]

We consider reconstruction of an ambient signal in a compressed sensing (CS) setup where the ambient signal has a neural network based generative model. The generative model has a sparse-latent input and we refer to the generated ambient signal as generative sparse-latent signal (GSL). The proposed sparsity inducing reconstruction algorithm is inherently non-convex, and we show that a gradient based search provides a good reconstruction performance. We evaluate our proposed algorithm using simulated data.

Place, publisher, year, edition, pages
European Signal Processing Conference, EUSIPCO, 2023
Keywords
Compressed sensing, generative models, inverse problems
National Category
Signal Processing
Identifiers
urn:nbn:se:kth:diva-340801 (URN)10.23919/EUSIPCO58844.2023.10289923 (DOI)2-s2.0-85178384307 (Scopus ID)
Conference
31st European Signal Processing Conference, EUSIPCO 2023, Helsinki, Finland, Sep 4 2023 - Sep 8 2023
Note

Part of ISBN 9789464593600

QC 20231214

Available from: 2023-12-14 Created: 2023-12-14 Last updated: 2023-12-14Bibliographically approved
Ghosh, A., Honore, A. & Chatterjee, S. (2023). DANSE: Data-driven Non-linear State Estimation of Model-free Process in Unsupervised Bayesian Setup. In: Proceedings 31st European Signal Processing Conference, EUSIPCO 2023: . Paper presented at 31st European Signal Processing Conference, EUSIPCO 2023, 4 September 2023 through 8 September 2023 (pp. 870-874). Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>DANSE: Data-driven Non-linear State Estimation of Model-free Process in Unsupervised Bayesian Setup
2023 (English)In: Proceedings 31st European Signal Processing Conference, EUSIPCO 2023, Institute of Electrical and Electronics Engineers (IEEE) , 2023, p. 870-874Conference paper, Published paper (Refereed)
Abstract [en]

We propose DANSE - a data-driven non-linear state estimation method. DANSE provides a closed-form posterior of the state of a model-free process, given linear measurements of the state in a Bayesian setup, like the celebrated Kalman filter (KF). Non-linear dynamics of the state are captured by data-driven recurrent neural networks (RNNs). The training of DANSE combines maximum-likelihood and gradient-descent in an unsupervised framework, i.e. only measurement data and no process data are required. Using simulated linear and non-linear process models, we demonstrate that DANSE - without knowledge of the process model - provides competitive performance against model-based approaches such as KF, unscented KF (UKF), extended KF (EKF), and a hybrid approach such as KalmanNet.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2023
Keywords
deep learning, neural networks, recurrent neural networks, state estimation, Bayesian networks, Deep neural networks, Gradient methods, Maximum likelihood estimation, Bayesian, Closed form, Data driven, Linear state estimation, Model free, Neural-networks, Nonlinear state, Process-models, State estimation methods
National Category
Control Engineering
Identifiers
urn:nbn:se:kth:diva-346076 (URN)10.23919/EUSIPCO58844.2023.10289946 (DOI)2-s2.0-85165166616 (Scopus ID)
Conference
31st European Signal Processing Conference, EUSIPCO 2023, 4 September 2023 through 8 September 2023
Note

Part of ISBN 9789464593600

QC 20240502

Available from: 2024-05-02 Created: 2024-05-02 Last updated: 2024-07-11Bibliographically approved
Fontcuberta, A. E., Ghosh, A., Chatterjee, S., Mitra, D. & Nandy, D. (2023). Forecasting Solar Cycle 25 with Physical Model-Validated Recurrent Neural Networks. Solar Physics, 298(1), Article ID 8.
Open this publication in new window or tab >>Forecasting Solar Cycle 25 with Physical Model-Validated Recurrent Neural Networks
Show others...
2023 (English)In: Solar Physics, ISSN 0038-0938, E-ISSN 1573-093X, Vol. 298, no 1, article id 8Article in journal (Refereed) Published
Abstract [en]

The Sun's activity, which is associated with the solar magnetic cycle, creates a dynamic environment in space known as space weather. Severe space weather can disrupt space-based and Earth-based technologies. Slow decadal-scale variations on solar-cycle timescales are important for radiative forcing of the Earth's atmosphere and impact satellite lifetimes and atmospheric dynamics. Predicting the solar magnetic cycle is therefore of critical importance for humanity. In this context, a novel development is the application of machine-learning algorithms for solar-cycle forecasting. Diverse approaches have been developed for this purpose; however, with no consensus across different techniques and physics-based approaches. Here, we first explore the performance of four different machine-learning algorithms - all of them belonging to a class called Recurrent Neural Networks (RNNs) - in predicting simulated sunspot cycles based on a widely studied, stochastically forced, nonlinear time-delay solar dynamo model. We conclude that the algorithm Echo State Network (ESN) performs the best, but predictability is limited to only one future sunspot cycle, in agreement with recent physical insights. Subsequently, we train the ESN algorithm and a modified version of it (MESN) with solar-cycle observations to forecast Cycles 22 - 25. We obtain accurate hindcasts for Solar Cycles 22 - 24. For Solar Cycle 25 the ESN algorithm forecasts a peak amplitude of 131 +/- 14 sunspots around July 2024 and indicates a cycle length of approximately 10 years. The MESN forecasts a peak of 137 +/- 2 sunspots around April 2024, with the same cycle length. Qualitatively, both forecasts indicate that Cycle 25 will be slightly stronger than Cycle 24 but weaker than Cycle 23. Our novel approach bridges physical model-based forecasts with machine-learning-based approaches, achieving consistency across these diverse techniques.

Place, publisher, year, edition, pages
Springer Nature, 2023
Keywords
Solar cycle, Sunspots, Statistics
National Category
Astronomy, Astrophysics and Cosmology
Identifiers
urn:nbn:se:kth:diva-323588 (URN)10.1007/s11207-022-02104-3 (DOI)000913507700001 ()2-s2.0-85146268485 (Scopus ID)
Note

QC 20230208

Available from: 2023-02-08 Created: 2023-02-08 Last updated: 2023-02-08Bibliographically approved
Ghosh, A., Fontcuberta, A. E., Abdalmoaty, M.-H. R. & Chatterjee, S. (2022). Time-varying Normalizing Flow for Generative Modeling of Dynamical Signals. In: 2022 30Th European Signal Processing Conference (EUSIPCO 2022): . Paper presented at 30th European Signal Processing Conference (EUSIPCO), AUG 29-SEP 02, 2022, Belgrade, SERBIA (pp. 1492-1496). IEEE
Open this publication in new window or tab >>Time-varying Normalizing Flow for Generative Modeling of Dynamical Signals
2022 (English)In: 2022 30Th European Signal Processing Conference (EUSIPCO 2022), IEEE , 2022, p. 1492-1496Conference paper, Published paper (Refereed)
Abstract [en]

We develop a time-varying normalizing flow (TVNF) for explicit generative modeling of dynamical signals. Being explicit, it can generate samples of dynamical signals, and compute the likelihood of a (given) dynamical signal sample. In the proposed model, signal flow in the layers of the normalizing flow is a function of time, which is realized using an encoded representation that is the output of a recurrent neural network (RNN). Given a set of dynamical signals, the parameters of TVNF are learned according to maximum-likelihood approach in conjunction with gradient descent (backpropagation). Use of the proposed model is illustrated for a toy application scenario - maximum-likelihood based speech-phone classification task.

Place, publisher, year, edition, pages
IEEE, 2022
Series
European Signal Processing Conference, ISSN 2076-1465
Keywords
Generative learning, recurrent neural networks, neural networks, normalizing flows
National Category
Signal Processing
Identifiers
urn:nbn:se:kth:diva-324330 (URN)000918827600293 ()2-s2.0-85141010789 (Scopus ID)
Conference
30th European Signal Processing Conference (EUSIPCO), AUG 29-SEP 02, 2022, Belgrade, SERBIA
Note

QC 20230228

Available from: 2023-02-28 Created: 2023-02-28 Last updated: 2023-06-19Bibliographically approved
Ghosh, A. (2020). Normalizing Flow based Hidden Markov Models for Phone Recognition. (Student paper). KTH
Open this publication in new window or tab >>Normalizing Flow based Hidden Markov Models for Phone Recognition
2020 (English)Student thesis
Alternative title[sv]
Normalisering av flödesbaserade dolda Markov-modeller för fonemigenkänning
Abstract [en]

The task of Phone recognition is a fundamental task in Speech recognition and often serves a critical role in bench-marking purposes. Researchers have used a variety of models used in the past to address this task, using both generative and discriminative learning approaches. Among them, generative approaches such as the use of Gaussian mixture model-based hidden Markov models are always favored because of their mathematical tractability. However, the use of generative models such as hidden Markov models and its hybrid varieties is no longer in fashion owing to a large inclination to discriminative learning approaches, which have been found to perform better. The only downside is that these approaches do not always ensure mathematical tractability or convergence guarantees as opposed to their generative counterparts. So, the research problem was to investigate whether there could be a process of augmenting the modeling capability of generative Models using a kind of neural network based architectures that could simultaneously prove mathematically tractable and expressive. Normalizing flows are a class of generative models that have been garnered a lot of attention recently in the field of density estimation and offer a method for exact likelihood computation and inference. In this project, a few varieties of Normalizing flow-based hidden Markov models were used for the task of Phone recognition on the TIMIT dataset. It was been found that these models and their mixture model varieties outperformed classical generative model varieties like Gaussian mixture models. A decision fusion approach using classical Gaussian and Normalizing flow-based mixtures showed competitive results compared to discriminative learning approaches. Further analysis based on classes of speech phones was carried out to compare the generative models used. Additionally, a study of the robustness of these algorithms to noisy speech conditions was also carried out.

Abstract [sv]

Uppgiften för fonemigenkänning är en grundläggande uppgift i taligenkänning och tjänar ofta en kritisk roll i benchmarkingändamål. Forskare har använt en mängd olika modeller som använts tidigare för att hantera denna uppgift genom att använda både generativa och diskriminerande inlärningssätt. Bland dem är generativa tillvägagångssätt som användning av Gaussian-blandnings modellbaserade dolda Markov-modeller alltid föredragna på grund av deras matematiska spårbarhet. Men användningen av generativa modeller som dolda Markov-modeller och dess hybridvarianter är inte längre på mode på grund av en stor lutning till diskriminerande inlärningsmetoder, som har visat sig fungera bättre. Den enda nackdelen är att dessa tillvägagångssätt inte alltid säkerställer matematisk spårbarhet eller konvergensgarantier i motsats till deras generativa motsvarigheter. Således var forskningsproblemet att undersöka om det kan finnas en process för att förstärka modelleringsförmågan hos generativa modeller med hjälp av ett slags neurala nätverksbaserade arkitekturer som samtidigt kunde visa sig matematiskt spårbart och uttrycksfullt. Normaliseringsflöden är en klass generativa modeller som nyligen har fått mycket uppmärksamhet inom området för densitetsberäkning och erbjuder en metod för exakt sannolikhetsberäkning och slutsats. I detta projekt användes några få varianter av Normaliserande flödesbaserade dolda Markov-modeller för uppgiften att fonemigenkänna i TIMIT-datasatsen. Det visade sig att dessa modeller och deras blandningsmodellvarianter överträffade klassiska generativa modellvarianter som Gaussiska blandningsmodeller. Ett beslutssmältningsstrategi med klassiska Gaussiska och Normaliserande flödesbaserade blandningar visade konkurrenskraftiga resultat jämfört med diskriminerande inlärningsmetoder. Ytterligare analys baserat på klasser av talsignaler utfördes för att jämföra de generativa modellerna som användes. Dessutom genomfördes en studie av robustheten hos dessa algoritmer till bullriga talförhållanden.

Publisher
p. 56
Series
TRITA-EECS-EX ; 2020:675
Keywords
Phone recognition, generative learning, Normalizing flows, Decision fusion, Speech recognition
National Category
Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
urn:nbn:se:kth:diva-286594 (URN)
Thesis level
Independent thesis Advanced level (degree of Master (Two Years)), 20 credits / 30 HE credits
Supervisors
Examiners
Available from: 2020-11-27 Created: 2020-11-26 Last updated: 2024-05-02Bibliographically approved
Ghosh, A., Honore, A., Liu, D., Henter, G. E. & Chatterjee, S. (2020). Robust classification using hidden markov models and mixtures of normalizing flows. In: 2020 IEEE 30th International Workshop on Machine Learning for Signal Processing (MLSP): . Paper presented at 30th IEEE International Workshop on Machine Learning for Signal Processing, MLSP 2020, 21 September 2020 through 24 September 2020, virtual, Espoo21 September 2020 through 24 September 2020. Institute of Electrical and Electronics Engineers (IEEE), Article ID 9231775.
Open this publication in new window or tab >>Robust classification using hidden markov models and mixtures of normalizing flows
Show others...
2020 (English)In: 2020 IEEE 30th International Workshop on Machine Learning for Signal Processing (MLSP), Institute of Electrical and Electronics Engineers (IEEE) , 2020, article id 9231775Conference paper, Published paper (Refereed)
Abstract [en]

We test the robustness of a maximum-likelihood (ML) based classifier where sequential data as observation is corrupted by noise. The hypothesis is that a generative model, that combines the state transitions of a hidden Markov model (HMM) and the neural network based probability distributions for the hidden states of the HMM, can provide a robust classification performance. The combined model is called normalizing-flow mixture model based HMM (NMM-HMM). It can be trained using a combination of expectation-maximization (EM) and backpropagation. We verify the improved robustness of NMM-HMM classifiers in an application to speech recognition.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2020
Series
IEEE International Workshop on Machine Learning for Signal Processing, ISSN 2161-0363
Keywords
Generative models, Hidden Markov models, Neural networks, Speech recognition, Backpropagation, Learning systems, Maximum likelihood, Maximum principle, Mixtures, Signal processing, Trellis codes, Combined model, Expectation Maximization, Generative model, Hidden state, Mixture model, Robust classification, Sequential data, State transitions
National Category
Probability Theory and Statistics Signal Processing
Identifiers
urn:nbn:se:kth:diva-291596 (URN)10.1109/MLSP49062.2020.9231775 (DOI)000630907800045 ()2-s2.0-85096485816 (Scopus ID)
Conference
30th IEEE International Workshop on Machine Learning for Signal Processing, MLSP 2020, 21 September 2020 through 24 September 2020, virtual, Espoo21 September 2020 through 24 September 2020
Note

QC 20210324

Part of conference proceedings: ISBN 9781728166629

Available from: 2021-03-24 Created: 2021-03-24 Last updated: 2024-05-02Bibliographically approved
Stridfeldt, F., Qin, H., Joelsson, S., Sahu, S. S., Ghosh, A., Hååg, P., . . . Dev, A.Machine Learning Reveals That Osimertinib Treatment Influences Surface Protein Profiles in Non-small Cell Lung Cancer Patients.
Open this publication in new window or tab >>Machine Learning Reveals That Osimertinib Treatment Influences Surface Protein Profiles in Non-small Cell Lung Cancer Patients
Show others...
(English)Manuscript (preprint) (Other academic)
National Category
Cancer and Oncology
Research subject
Biological Physics
Identifiers
urn:nbn:se:kth:diva-352957 (URN)
Note

QC 20240912

Available from: 2024-09-10 Created: 2024-09-10 Last updated: 2024-09-12Bibliographically approved
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0001-6612-6923

Search in DiVA

Show all publications