Hidden Markov Models: Difference between revisions

From MSc Voice Technology
Jump to navigation Jump to search
No edit summary
Line 3: Line 3:
== Introduction ==
== Introduction ==


A '''Hidden Markov Model (HMM)''' is a temporal probabilistic model in which some hidden or unobservable states are described by observable variables. <ref>Russell, S. J. (2010). ''Artificial intelligence a modern approach''. Pearson Education, Inc..</ref> These hidden states adhere to the Markov property, meaning that the current state is only dependent on the previous state. The observable variables, on the other hand, should be dependent on the hidden state such that a
A '''Hidden Markov Model (HMM)''' is a temporal probabilistic model in which some "hidden" or unobservable states are described by observable variables that are generated by these hidden states. <ref>Russell, S. J. (2010). ''Artificial intelligence a modern approach''. Pearson Education, Inc..</ref> These hidden states adhere to the Markov property, meaning that the future state is only dependent on the current state. Since one cannot observe the underlying states of a specific model, learning the transition function of this sequence of states involves aligning the HMM to the observable states.<ref>Eddy, S. R. (1996). Hidden markov models. ''Current opinion in structural biology'', ''6''(3), 361-365.</ref><ref>Rabiner, L. R. (1989). A tutorial on hidden Markov models and selected applications in speech recognition. ''Proceedings of the IEEE'', ''77''(2), 257-286.</ref>
 
Many real-world applications present hidden variables that are only observable through some emitted outcome, e.g. a speech signal of a word is observed rather than the specific phoneme states that are the underlying hidden states. To determine what the sequence of phonemes (states) would be that results in that specific word, the model learns the relation between the observed and unobservable variables.<ref>Juang, B. H., & Rabiner, L. R. (1991). Hidden Markov models for speech recognition. ''Technometrics'', ''33''(3), 251-272.</ref>


== Historical Context ==
== Historical Context ==

Revision as of 11:30, 16 September 2023

Claimed by Ömer and Jocomin

Introduction

A Hidden Markov Model (HMM) is a temporal probabilistic model in which some "hidden" or unobservable states are described by observable variables that are generated by these hidden states. [1] These hidden states adhere to the Markov property, meaning that the future state is only dependent on the current state. Since one cannot observe the underlying states of a specific model, learning the transition function of this sequence of states involves aligning the HMM to the observable states.[2][3]

Many real-world applications present hidden variables that are only observable through some emitted outcome, e.g. a speech signal of a word is observed rather than the specific phoneme states that are the underlying hidden states. To determine what the sequence of phonemes (states) would be that results in that specific word, the model learns the relation between the observed and unobservable variables.[4]

Historical Context

Key Innovations

Impact on the Field

Future Research

LLM Review

References

Here thus are the references: [5]

  1. Russell, S. J. (2010). Artificial intelligence a modern approach. Pearson Education, Inc..
  2. Eddy, S. R. (1996). Hidden markov models. Current opinion in structural biology, 6(3), 361-365.
  3. Rabiner, L. R. (1989). A tutorial on hidden Markov models and selected applications in speech recognition. Proceedings of the IEEE, 77(2), 257-286.
  4. Juang, B. H., & Rabiner, L. R. (1991). Hidden Markov models for speech recognition. Technometrics, 33(3), 251-272.
  5. Placeholder Reference