L54 Markov Switching Models
- MSM
- class of nonlinear TS models
- capture regime-switching behavior
- assuming transition between regime is governed by a hidden Markov process
- TAR and SETAR have regime changes depending on observable variables**
- VS MSM, where regime switching depends on unobservable state variable
Key Features of MSM¶
- Regime switching
- Markov process
- current regime depends only on previous regime1
- Hidden state
- regime is a latent (unobservable) variable
- Flexibility
- suitable for capturing abrupt or smooth transition, dependent on state transition probabilities
Model Formulation¶
- where
- \(s_{t}\) is the latent state (regime) taking values \(1,2,\dots,k\)
- \(\mu_{s_{t}}\) is a regime specific mean
- \(\phi_{s_{t},i}\) are AR coefficients
- \(e_{t}\) error terms with variance \(\sigma_{s_{t}}^{2}\)
Transition Probabilities¶
The regime-switching controlled by a state transition probability matrix (\(P\))
- \(p_{ij} = P(s_{t}= j | s_{t-1}=i)\)
- Each row of \(P\) sums to \(1\)
Estimation¶
- Likelihood function
- Expectation-Maximization (EM) Algorithm
- E-step = expected values of hidden states given params
- M-step = updates parameters to maximize expected likelihood
- Filter and Smoother
- Hamilton Filter = estimate probability of being in particular state at \(t\)
- Kim Filter = smooths the probability to improve state estimation
Advantages of MSM¶
- Flexibility
- State dependence → improve model accuracy
- Latent Regime Models → infers unobservable states that drive the system
Challenges in MSM¶
- Computational Complexity (high dimensional likelihood function)
- Model selection (number of regimes, lag structure)
- Interpretability: regime probability must be carefully interpreted
- Overfitting \(\impliedby\) too many parameters
Numerical Example¶
- \(n=300\)
- \(\mu = [1,-1]\)
- \(\phi = [0.8,0.5]\)
- \(\sigma= [0.5,1]\)
Estimation Interpretation¶
- mean \(\mu\), AR coefficients \(\phi\), variance \(\sigma^{2}\) for each regime
- transition probabilities \(p_{ij}\)
- Hidden states: smoothed probabilities indicate likelihood of TS being in each regime at a given time (Carefully address hidden states)
Output interpretation¶
- Parameter estimates
- Plot of smoothed probabilities
- Diagnostics
Other Possible Interesting Extensions¶
Bilinear Models¶
- Incorporates products of
- past observations and noise
- handles nonlinearities
- Uses: environmental data with interaction effects
Nonlinear ARX models¶
- Nonlinear function of
- past inputs and outputs
- \(f(\cdot)\) can be a polynomial, neural network or other nonlinear functions
- Applications: control systems, dynamic system modelling
Neutral-Network Based models¶
- Nonlinear Autoregressive Neural Network (NAR-NN)
- use neural network to model the autoregressive relationship
- Flexible, powerful for compl2ex nonlinear dependencies
- Long Short-Term Memory (LSTM) → handles sequential dependencies and long-term memory in nonlinear time series
Others¶
- ST-GARCH: Smooth Transition GARCH
- Extends GARCH by allowing regime switching in variance equation
- Hidden Markov Models (HMMs)
- switching dynamics in TS using discrete latent states
- Functional Time Series
- high (infinite) dimensional
- models temporal changes in curves/functions
- Wavelet Transform Models
- decompose TS into different frequency components
-
Markov Property: today's state is dependent only on yesterday's state ↩