{"title":"Parameter estimation of hidden Markov models: comparison of EM and quasi-Newton methods with a new hybrid algorithm","authors":"Sidonie FoulonCESP, NeuroDiderot, Thérèse TruongCESP, Anne-Louise LeuteneggerNeuroDiderot, Hervé PerdryCESP","doi":"arxiv-2409.02477","DOIUrl":null,"url":null,"abstract":"Hidden Markov Models (HMM) model a sequence of observations that are\ndependent on a hidden (or latent) state that follow a Markov chain. These\nmodels are widely used in diverse fields including ecology, speech recognition,\nand genetics.Parameter estimation in HMM is typically performed using the\nBaum-Welch algorithm, a special case of the Expectation-Maximisation (EM)\nalgorithm. While this method guarantee the convergence to a local maximum, its\nconvergence rates is usually slow.Alternative methods, such as the direct\nmaximisation of the likelihood using quasi-Newton methods (such as L-BFGS-B)\ncan offer faster convergence but can be more complicated to implement due to\nchallenges to deal with the presence of bounds on the space of parameters.We\npropose a novel hybrid algorithm, QNEM, that combines the Baum-Welch and the\nquasi-Newton algorithms. QNEM aims to leverage the strength of both algorithms\nby switching from one method to the other based on the convexity of the\nlikelihood function.We conducted a comparative analysis between QNEM, the\nBaum-Welch algorithm, an EM acceleration algorithm called SQUAREM (Varadhan,\n2008, Scand J Statist), and the L-BFGS-B quasi-Newton method by applying these\nalgorithms to four examples built on different models. We estimated the\nparameters of each model using the different algorithms and evaluated their\nperformances.Our results show that the best-performing algorithm depends on the\nmodel considered. QNEM performs well overall, always being faster or equivalent\nto L-BFGS-B. The Baum-Welch and SQUAREM algorithms are faster than the\nquasi-Newton and QNEM algorithms in certain scenarios with multiple optimum. In\nconclusion, QNEM offers a promising alternative to existing algorithms.","PeriodicalId":501215,"journal":{"name":"arXiv - STAT - Computation","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - STAT - Computation","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.02477","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Hidden Markov Models (HMM) model a sequence of observations that are
dependent on a hidden (or latent) state that follow a Markov chain. These
models are widely used in diverse fields including ecology, speech recognition,
and genetics.Parameter estimation in HMM is typically performed using the
Baum-Welch algorithm, a special case of the Expectation-Maximisation (EM)
algorithm. While this method guarantee the convergence to a local maximum, its
convergence rates is usually slow.Alternative methods, such as the direct
maximisation of the likelihood using quasi-Newton methods (such as L-BFGS-B)
can offer faster convergence but can be more complicated to implement due to
challenges to deal with the presence of bounds on the space of parameters.We
propose a novel hybrid algorithm, QNEM, that combines the Baum-Welch and the
quasi-Newton algorithms. QNEM aims to leverage the strength of both algorithms
by switching from one method to the other based on the convexity of the
likelihood function.We conducted a comparative analysis between QNEM, the
Baum-Welch algorithm, an EM acceleration algorithm called SQUAREM (Varadhan,
2008, Scand J Statist), and the L-BFGS-B quasi-Newton method by applying these
algorithms to four examples built on different models. We estimated the
parameters of each model using the different algorithms and evaluated their
performances.Our results show that the best-performing algorithm depends on the
model considered. QNEM performs well overall, always being faster or equivalent
to L-BFGS-B. The Baum-Welch and SQUAREM algorithms are faster than the
quasi-Newton and QNEM algorithms in certain scenarios with multiple optimum. In
conclusion, QNEM offers a promising alternative to existing algorithms.