Next Article in Journal
Enantiodivergent Aldol Condensation in the Presence of Aziridine/Acid/Water Systems
Next Article in Special Issue
An Improved Near-field Magnetic Probe Radiation Profile Boundaries Assessment for Optimal Radiated Susceptibility Pre-Mapping
Previous Article in Journal
A Novel Integrated PIPRECIA–Interval-Valued Triangular Fuzzy ARAS Model: E-Learning Course Selection
Previous Article in Special Issue
An Eigenvalues Approach for a Two-Dimensional Porous Medium Based Upon Weak, Normal and Strong Thermal Conductivities
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Bayesian Approach to Heavy-Tailed Finite Mixture Autoregressive Models

1
Institute of Research and Development, Duy Tan University, Da Nang 550000, Vietnam
2
Department of Statistics, Faculty of Science, Fasa University, Fasa 74616-86131, Iran
3
Department of Statistics, University of Isfahan, Isfahan 8174673441, Iran
4
Department of Mathematics, Faculty of Art and Sciences, Cankaya University, Balgat, 06530 Ankara, Turkey
5
Institute of Space Sciences, 077125 Magurele-Bucharest, Romania
6
University of Economics and Law, Ho Chi Minh City 700000, Vietnam
7
Fractional Calculus, Optimization and Algebra Research Group, Faculty of Mathematics and Statistics, Ton Duc Thang University, Ho Chi Minh City 758307, Vietnam
*
Author to whom correspondence should be addressed.
Symmetry 2020, 12(6), 929; https://doi.org/10.3390/sym12060929
Submission received: 10 March 2020 / Revised: 9 May 2020 / Accepted: 18 May 2020 / Published: 2 June 2020
(This article belongs to the Special Issue Composite Structures with Symmetry)

Abstract

:
In this paper, a Bayesian analysis of finite mixture autoregressive (MAR) models based on the assumption of scale mixtures of skew-normal (SMSN) innovations (called SMSN–MAR) is considered. This model is not simultaneously sensitive to outliers, as the celebrated SMSN distributions, because the proposed MAR model covers the lightly/heavily-tailed symmetric and asymmetric innovations. This model allows us to have robust inferences on some non-linear time series with skewness and heavy tails. Classical inferences about the mixture models have some problematic issues that can be solved using Bayesian approaches. The stochastic representation of the SMSN family allows us to develop a Bayesian analysis considering the informative prior distributions in the proposed model. Some simulations and real data are also presented to illustrate the usefulness of the proposed models.

1. Introduction

Data analysts apply computational models to describe and infer statements about complex datasets. Mixture models are a valuable class of these models. Finite mixture models are of great importance in statistical inferences. The multimodality, skewness, kurtosis, and unobserved heterogeneity are usually observed in many datasets, for example time series datasets. Importance of mixture distributions, which are the main tools for statistical mixture models, has been noted by many references [1,2,3,4,5,6,7,8]. Various statistical fields containing time series modeling and regression analysis frequently use the mixture modeling. In fact, in analyzing time series data, some events may affect and change the behavior of data over time, for example, finance crises in many financial or panel time series datasets. The finite mixture autoregressive (MAR) models were suggested by Wong and Li [7], to catch the multimodal phenomena, are flexible and applicable in many fields, such as electroencephalogram modeling in medicine [8], interest rates and bond pricing [9], Forex rate [10], and other various fields such as telecommunications, hydrology, biology, sociology, and medical sciences.
Most researchers considered the MAR models based on Gaussian distribution, which are called Gaussian mixture of autoregressive (GMAR) models [7,8,9,10,11]. The GMAR models are sensitive to the existence of outliers, heavy tailed distribution and asymmetry in datasets, and thus some authors considered MAR models based on non-Gaussian, heavy tailed and/or asymmetric distributions. Wong and Li [12] introduced logistic mixture autoregressive model. Wong and Chan [13] also introduced a student t–mixture autoregressive model and discussed its inferences and then applied this model to the heavy tailed financial data.
The SMSN family, introduced by Branco [1], is a rich family containing famous symmetric scale mixtures of normal (SMN; [14]) distributions and asymmetric distributions such as the skew-t, skew-slash and skew-contaminated-normal distribution. The MAR model based on the SMSN innovations, hereafter called SMSN–MAR models, has reasonable performance to model real time series data with outliers and skewness. The various MAR model members can determine weights to each observation, and therefore each single observation affects the estimation of the model’s parameters, and it leads to robust inferences. Bayesian inferences of the MAR models have some advantages compared to classical methods, that some of them are discussed below.
For maximum likelihood computations and classical inference of a MAR model, although the numerical optimization algorithms such as the EM–algorithm can be applied, converging to the major mode may fail, while this is not an issue in Bayesian inferences. Furthermore, in the MAR models there exist many situations with indirect information about the parameters or large changes in the results by small changes in the data. When we consider a sample from a MAR model, the sample includes no information about the parameters of one or more components with a positive probability. Therefore, the likelihood function can be unbounded and consequently the resultant likelihood estimates are only local maximum. Using suitable priors and some Bayesian techniques can solve such problems. Another problematic issue on the classical likelihood–based inference is that, in MAR model, the component parameters are not identifiable marginally, so distinguishing any component from other components in the likelihood is hard. The identifiability is not an important problem in Bayesian inferences. Some of the other advantages of the Bayesian inferences of the mixture models are given in [15,16].
In this paper, we consider a Bayesian technique to find estimates of the SMSN–MAR models by the Gibb sampling scheme and application of the Markov Chain Monte Carlo (MCMC) methods. The main characteristics of the SMSN distributions are reviewed in Section 2. In Section 3, we consider the MAR models and apply the Bayesian methodology for estimating the proposed MAR model’s parameters. To evaluate the Bayesian estimates of parameters and performance of the proposed model, numerical studies and a real practical example are reported in Section 4. The conclusion is provided in the last section.

2. The SMSN Distributions

A random variable Y is belonged to the scale mixtures of skew normal (SMSN) family, if it has the following representation:
Y = μ + k 1 / 2 ( U ) Z ,
where the random variable Z follows a skew-normal distribution in [17], where σ 2 and λ are respectively scale and skewness parameters, the function k ( u ) is positive respect to u , U is the scale random variable with density function or pmf h ( ; ν ) and distribution function H ( ; ν ) , and the vector or scalar ν indexes the distribution of U . The random variable Y is denoted by Y   ~   S M S N ( μ , σ 2 , λ , ν ) , and in this work, because of its suitable mathematical properties, we let k ( u ) = 1 / u , so Y | U = u ~ S N ( μ , u 1 σ 2 , λ ) . Therefore, the density of   Y   is given by
f ( y ; θ ) = 2 0 ϕ ( y ; μ , u 1 σ 2 ) Φ ( u 1 / 2 λ ( y μ ) σ ) d H ( u | ν ) ,
where   θ = ( μ , σ 2 , λ , ν ) , and ϕ ( ) and   Φ ( ) are density and distribution functions of the standard normal distribution, respectively.
Proposition 1.
Let    Y   ~   S M S N ( μ , σ 2 , λ ; H ) ,
(a)
If  E [ U 1 / 2 ] < then  μ Y = E [ Y ] = μ b Δ ,
(b)
If  E [ U 1 ] < then  σ Y 2 = V a r [ Y ] = σ 2 k 2 b 2 Δ 2 ,
(c)
Y  has a stochastic representation given by Y = μ + Δ W + γ U 1 / 2 W 1 ,
where   b = 2 / π E [ U 1 / 2 ] ,  δ = λ / 1 + λ 2 ,  Δ = σ δ , γ 2 = σ 2 Δ 2 , W = U 1 / 2 | w 0 | , and W 0  and  W 1  are independent standard normal random variables.
The famous members of the SMSN family are the scale mixtures of normal (SMN) family [8,13], as the symmetric members of the SMSN family, and skew-normal (SN), skew-t (ST), skew-slash (SSL) and skew-contaminated-normal (SCN) distributions. More details are given in [18,19].

3. The SMSN–MAR Model and Bayesian Estimates

3.1. The SMSN–MAR Model

In this paper, we study the MAR model, which has been studied by Wong and Li [7], and Wong and Chan [20], due to its flexibility in use for non-linear time series analysis. This model is represented by
X t = φ i , 0 + j = 1 p i φ i , j X t j + ε i , t ,   with   probability   π i ;   i = 1 , 2 , , g ;   t = 0 , ± 1 , ± 2 , ,
or equivalently,
X t = { φ 1 , 0 + j = 1 p 1 φ 1 , j X t j + ε 1 , t ;   with   probability   π 1 φ 2 , 0 + j = 1 p 2 φ 2 , j X t j + ε 2 , t ;   with   probability   π 2 ,     t = 0 , ± 1 , ± 2 , , φ g , 0 + j = 1 p g φ g , j X t j + ε g , t ;   with   probability   π g
where
  • g is a known positive integer which indicates the number of components in the model;
  • each component occurs with probabilities   π i > 0 ,   i = 1 , , g ,   i = 1 g π i = 1 , which obeys a discrete distribution π ;
  • for each i = 1 , , g ,
    the ith autoregressive component is of order p i 1 ;
    φ i , j ,   j = 0 , , p i , are the autoregressive coefficients of the ith components;
    each ith innovation’s component   ε i , t distributed as following:
    ε i , t       i . i . d . ~       S M S N ( b i Δ i ,   σ i 2 ,   λ i ,   ν i ) ;   i = 1 , , g ;   t = 0 , ± 1 , ± 2 ,
    where b i = 2 / π E [ U i r / 2 ] , δ i = λ i / 1 + λ i 2 , Δ i = σ i δ i and γ i 2 = σ i 2 + Δ i 2 . Also in each ith component, { ε i , t } are jointly independent and are independent of past X t s and also independent of other component’s innovations.
Hereafter, this model will be called SMSN–MAR(p,g). Note that in each component, when   b i exists,   E ( ε i , t ) = 0 , and   Var ( ε i , t ) = σ i 2 k i 2 b i 2 Δ i 2   ;   i = 1 , , g . Note that we can assume that the vector of coefficients in MAR models, distributed as   π , therefore, this model can be also interpreted as random coefficients autoregressive (RCA) model.
Without loss of generality, it is convenient to set p = max i = 1 , , g   p i and   φ i , j = = φ i , p = 0 for   j > p i . Also, let   { Z t } be an i.i.d. sequence of random variables which are distributed as   π (Multinomial), such that   P ( Z t = i ) = π i , i = 1 , , g , that determines the components, and be independent of innovations and random history of model   X = ( X 1 , , X n ) . Therefore, X t   obeys from ith component of MAR model, if   Z t = i . Thus, Equation (3) can be written as
X t = φ Z t , 0 + j = 1 p φ Z t , j X t j + ε Z t , t ,       t = 0 , ± 1 , ± 2 ,
where   Z t ~   Multinomial ( 1 , π i , , π g ) .
It is convenient to set vector and matrix notations of sample   X = ( X 1 , , X n ) in the form of X t 1 = ( 1 , X t 1 , , X t p ) ;   t = 1 , , n ,   ϕ Z t = ( φ 0 , Z t , , φ p , Z t ) , ε Z t = ( ε Z t , 1 , , ε Z t , n ) and A be the n × p matrix whose tth row is X t 1 , in the rest of the paper. Thus, the following matrix representation
X = A ϕ Z t + ε Z t ,
Note that in Equation (7), X t = ϕ Z t X t 1 + ε Z t , t where Z t ,   t = 1 , , g and distributed as   π , so by Equation (5) we have
X t | X t 1 , Z t ~ S M S N ( ϕ Z t X t 1 + b Z t Δ Z t ,   σ Z t 2 ,   λ Z t ,   ν Z t ) ;   Z t = 1 , , g ,       t = 1 , , n .
Therefore, the density of N t is given by
f X t | X t 1 , ,   X t p ( x t ; θ ) = i = 1 g π i f i ( x t | X t 1 , θ i ) ,
where f i ( . | X t 1 , θ i ) is the pdf of S M S N ( ϕ i X t 1 + b i Δ i ,   σ i 2 ,   λ i ,   ν i ) defined in Equation (2), hereafter note that θ = ( θ 1 , , θ g ) , where   θ i = ( π i ,   ϕ i , σ i 2 ,   λ i ,   ν i ) ,   i = 1 , , g be the vector of parameters.
By using Markovian property of this model, we have that   L ( θ | X ) = f X ( X 1 ,   ,   X n ; θ ) = t = 1 n f X t | X t p , ,   X t 1 ( X t ; θ ) , where   f X ( . ) be the density of   X . An auxiliary determiner of components in MAR models (6), can be expressed in the random vector Z t = ( Z t 1 , , Z t g ) , where
Z t i = { 1 , if   t th   observation   of   MAR   process   obeys   the   i th   compoment 0   , otherwise
Therefore, under the above approach, the latent random vector Z t ,   t = 1 , , n has the following multinomial Z t   ~   M u l t i n o m i a l ( 1 , π 1 , , π g ) with the following probability mass function:
P ( Z i = z i ) = π 1 z t 1 π 2 z t 2 π g z t g ;   t = 1 , , n ,
such that Σ i = 1 g π i = 1 , π i > 0 ,   i = 1 , , g . Therefore,
X t | X t 1 , Z t i = 1 ~ S M S N ( ϕ i X t 1 + b i Δ i ,   σ i 2 ,   λ i ,   ν i ) ,   i = 1 , , g ;   t = 1 , , n .
Also, to provide a MCMC method, we have:
X t | X t 1 , W t i = w t i ,   U t i = u t i ,   Z t i = 1   ~   N ( ϕ i X t 1 + Δ i w t i ,   u t i 1 γ i 2   ) ,
W t i | U t i = u t i ,   Z t i = 1   ~   T N 1 ( b i ,   u t i 1 ) I ( b i , ) ,
U t i | Z t i = 1   ~   H ( u t i ;   ν i ) ,
Z t   ~   M u l t i n o m i a l ( 1 , π 1 , , π g ) ,
for t = 1 , , n . Let D = ( X , M ) as the complete data, where   X is the observed part of data, M = ( ( W t i , U t i , Z t i ) t = 1 n ) i = 1 g is the missing parts of data, and θ is the vector of unknown parameters. The conditional likelihood function based on the complete data is provided by
L ( θ | D ) = t = 1 n i = 1 g ( π i f X t | X t 1 , W t i ,   U t i ,   Z t i ( X t ; θ i ) f W t i | U t i ,   Z t i ( W t i ; θ i ) f U t i | Z t i ( U t i ; θ i ) ) Z t i .

3.2. Bayesian Approach

In this part, we use the Bayesian approach by applying MCMC algorithms for the model in Equation (3). We consider informative priors under squared error loss function (SEL). To consider the prior distributions for the parameters of the considered model, we assign prior distributions to the parameters as follow:
ϕ i   ~   N p + 1 ( φ i ,   Σ i ) ,       Δ i ~   N ( μ Δ i , σ Δ i 2 ) ,       γ i 2   ~   IG ( ρ i / 2 , ϱ i / 2 ) ,   π = ( π 1 , , π g )   ~   Dir ( δ 1 , , δ g ) ,
for i = 1 , , g ,   where Dir and IG denote the Dirichlet distribution and the Inverse-Gamma distribution, respectively.
The prior distribution assigned to ν i ,   i = 1 , g , varies with the particular SMSN distributions. For skew–t model, we can consider ν i ~   exp ( ς i / 2 ) I ( 2 , ) , exponential distribution with mean   2 / ς i before truncation, for skew–slash model, we can consider ν i ~   Gamma ( a i , b i ) with small positive values a i and b i such that b i a i , and for skew–contaminated–normal model, the independent and non-informative U ( 0 , 1 )   prior distributions are considered for each component of ν i = ( ν i , γ i ) . We also assume that priors are independent and all hyper-parameters (parameters of priors) are known, which guarantees that all posteriors are proper. The joint posterior of θ and missing variables is in the form of π ( θ | D ) L ( θ | D ) π ( θ ) , which is not analytically tractable, but the MCMC methods such as the Gibbs sampling, will be applied to draw samples from the marginal of conditional posteriors which are shown as follows, when θ ( m ) , means the mth element of θ has been removed (Algorithm 1).
Algorithm 1: MCMC
For i = 1 , , g , and t = 1 , , n ,
  • Z t | θ ,   U , W ; X ~ M u l t i n o m i a l ( 1 , p 1 , , p g ) ,
  • π | θ ( π ) ,   D ~ D i r ( δ 1 + t = 1 n Z t 1 , , δ g + t = 1 n Z t g ) ,
  • W t i | θ ,   U , Z t i = 1 ; X ~ T N 1 ( μ W t i + b i , u t i 1 M i 2 ) I ( b i , ) ,
  • ϕ i | θ ( ϕ i ) , Z t i = 1 , D ~ N p + 1 ( μ ϕ i ,   Σ ϕ i ) ,
  • γ i 2 | θ ( γ i ) ,   Z t i = 1 , D ~ I G ( ρ i , τ i ) ,      
  • Δ i | θ ( Δ i ) , Z t i = 1 ,   D ~ N ( β i σ Δ i 2 + γ i 2 μ Δ i α i σ Δ i 2 + γ i 2 , α i σ Δ i 2 + γ i 2 γ i 2 σ Δ i 2 ) ,
  • U t i | θ , Z t i = 1 , D ~ H p o s t ( u ; θ , ν ) ,
  • ν | θ ( ν ) , U , Z t i = 1 , D ~ f p o s t ( ν   ; θ ) ,
where Z t = ( Z t 1 , , Z t g ) , p i = π i f ( X t | X t 1 ; θ i ) i = 1 g π i f ( X t | X t 1 ; θ i ) , M i 2 = γ i 2 γ i 2 + Δ i 2 , μ W t i = M i 2   Δ i r i 2 ( X t ϕ i X t 1 b i Δ i ) , μ ϕ i = ( γ i 2 ( Z Λ i Z ) + M i 0 1 ) 1 ( γ i 2 ( Z Λ i Z ) M i + M i 0 1 ϕ i 0 ) , Σ ϕ i = ( γ i 2 ( Z Λ i Z ) + M i 0 1 ) 1 , ρ i = n p + p i + i 2 , τ i = 1 2 ( ϱ i + ( ϕ i M i ) ( Z Λ i Z ) ( ϕ i M i ) ) , α i = W i Λ i   W and β i = W i Λ i ( X t ϕ i X t 1 b i Δ i ) .
Finally, in order to complete the specifications of the sampling via MCMC methods, the posterior distribution of the latent variable U t i , t = 1 , , n ,   i = 1 , , g , H p o s t ( u ; θ , ν ) , and its indexed parameter   ν i , f p o s t ( ν ; θ ) , should be determined. Set c t i = ( X t ϕ i X t 1 Δ i W t i ) 2 / σ i 2 + ( W t i b i ) 2 , we then have, for t = 1 , , n and the several models given as:
ST–MAR:
U t i | θ , Z t i = 1 , D ~ Gamma ( ν i + 2 2 , ν i + c t i 2 ) ,   i = 1 , , g ,   t = 1 , , n ,
and
π ( ν i | θ ( ν i ) , Z t i = 1 , D ) ( 2 ν i / 2 Γ ( ν i / 2 ) ) n × Gamma ( n ν i 2 1 , 1 2 [ t = 1 n ( U t i log U t i ) + ς i ] ) I ( 2 , ) ( ν i ) .
SSL–MAR:
U t i | θ , Z t i = 1 , D   ~   Gamma ( ν i + 1 , c t i 2 ) I ( 0 , 1 ) ( U t i ) ,   i = 1 , , g ,
ν i | θ ( ν i ) , U , Z t i = 1 , D   ~   Gamma ( n + a i , b i t = 1 n log U t i ) ,   i = 1 , , g .
SCN–MAR:
π ( u t i | θ , Z t i = 1 , D ) = A t i A t i + B t i I ( U t i = γ i ) + B t i A t i + B t i I ( U t i = 1 ) ,   i = 1 , , g ,
where A t i = ν i γ i exp ( γ i c t i / 2 ) , B t = ( 1 ν i ) exp ( c t i / 2 ) , and
ν i | θ ( ν i ) , Z t i = 1 , D ~ Beta ( n t = 1 n U t i 1 γ i + 1 , t = 1 n U t i n γ i 1 γ i + 1 ) ,   i = 1 , , g ,
π ( γ i | θ ( γ i ) , Z t i = 1 , D ) ν i n t = 1 n U t i 1 γ i ( 1 ν i ) t = 1 n U t i n γ i 1 γ i   ,   i = 1 , , g .
Note that Equations for v i and γ i do not have closed forms, but a Metropolis–Hasting algorithm which has used in [21] can be employed to obtain draws of them.
In the Bayesian technique, some model selection criteria such as the expected Bayesian information criterion (EBIC) and the expected Akaike information criterion (EAIC) [22,23] can be considered. Let θ 1 , , θ m as a sample from π ( θ | D ) . The posterior mean of the deviance is estimated by   D ¯ = i = 1 m D ( θ i ) / m , such that D ( θ ) = 2 t = 1 n log f ( X t | X t 1 , θ ) , and   f ( | X t 1 , θ ) is the density given in Equation (8). These criteria can be respectively estimated by EAIC ^ = D ¯ + 2 k and EBIC ^ = D ¯ + k log n , where   k is the number of parameters,   n is the sample size and D ^ = D ( θ ¯ ) , where   θ ¯ = i = 1 m θ i / m .

4. Numerical Studies

In this section, numerical studies and a real data example are given to study the ability of the considered SMSN–MAR(p,g) model. The statistical R software version 3.6.1 is used. A sample size of n = 200 , and the following prior distributions are selected: ϕ i   ~   N p + 1 ( 0 , 10 3 I p ) , Δ i ~   N 1 ( 0 , 10 3 ) and γ i 2   ~   I G ( 0.001 ,   0.001 ) , with ν i   ~   exp ( 0.1 ) I ( 2 , ) for the skew-t model, ν i   ~ gamma ( 1 ,   0.01 ) for the skew-slash model, and ν i   ~   U ( 0 ,   1 ) and independent of γ i   ~   U ( 0 ,   1 ) for the skew-contaminated normal model. Also, a Gibbs sampling with iterations equal to 60,000 and a burn-in of 10,000 cycles is applied.
In the simulation parts, we considered two schemes. In the first scheme, simulations are based on the SMSN–MAR models to show the performances of the considered model and Bayesian estimates, and in the second scheme, simulations are mixture autoregressive model based on Generalized–Hyperbolic distribution (GH–MAR model). The GH distribution is well known because of it’s asymmetry and heavy-tailed nature [24,25]. The second scheme will show the performance of the proposed Bayesian estimates of the heavy-tailed SMSN–MAR models to the time series with heavy-tailed and asymmetric innovations. Finally, the SMSN–MAR model is applied on the closing price of the Iran Telecommunication Company stock via the proposed Bayesian methodology.

4.1. First Scheme

The first simulation study is devoted to the following SMSN–MAR(2,2) model with 1000 samples simulations from the SN, ST and SSL distributions for the innovations with parameters ϕ 1 = ( φ 1 , 0 , φ 1 , 1 , φ 1 , 2 ) = ( 0.1 , 0.5 , 0.3 ) , ϕ 2 = ( φ 2 , 0 , φ 2 , 1 , φ 2 , 2 ) = ( 0.1 , 0.6 , 0.2 ) ,   π 1 = 0.4 , σ 1 2 = 1 , σ 2 2 = 2 , λ 1 = 2 and λ 2 = 4 , and ν 1 = ν 2 = 3 for ST and SSL, and ( ν 1 , γ 1 ) = ( ν 2 , γ 2 ) = ( 0.5 , 0.5 ) for the SCN. Means and standard deviations of the proposed Bayesian estimates are shown in Table 1.

4.2. Second Scheme

The second simulation study is devoted to the following GH–MAR(2,2) model with 200 samples as follows:
X t = { ϕ 1 X t 1 + ε t 1 ;   w . p .   π 1 = 0.4 , ϕ 2 X t 1 + ε t 2 ;   w . p .   π 2 = 0.6 ,
with the autoregressive coefficients in the first scheme, for which { ε t 1 } be an i.i.d. sequence of G H ( μ = 0 ,   σ = 1 , γ = 0.1 ,   α ¯ = 2 , λ = 1 ) with weak skewness, and { ε t 2 } be an i.i.d. sequence of G H ( μ = 0 ,   σ = 1 , γ = + 3 ,   α ¯ = 2 , λ = 1 ) with strong skewness for innovations. The EAIC criteria of all Bayesian fitted SMSN–MAR models given in Figure 1, satisfies that the asymmetry and heavy-tailed members (i.e., ST–MAR, SSL–MAR and SCN–MAR models), especially ST–MAR, are the best fitted models in comparison with the light-tailed SN–MAR member.

4.3. Real Data

In this section, we consider the 446 daily observations of the closing price of Iran Telecommunication Company (I.T.C.) stock based on the Iranian Rial from 2011-07-02 up to 2013-07-02 (see data at Tehran securities exchange technology management company site: www.tsetmc.com). Original time series plot of the I.T.C. stock data { x t } , where x t is the closing price at time   t , are given in Figure 2. It can be observed that the time series { x t } is not stationary. A histogram of the data is plotted in Figure 3.
Using the Box-Cox transformation with λ = 0 , and the difference with order of one, we have
y t = x t * x t 1 * = log ( x t / x t 1 ) ,
where
x t * = log x t .
The time series { y t } has been plotted in in Figure 4. The Dickey–Fuller test verifies that the transformed time series data are stationary (p-value = 0.01).
The EAIC and EBIC criteria demonstrate that the best SMSN–MAR model has two components with order p1 = 3 for the first AR component and p2 = 1 for the second AR component. Using the proposed Bayesian approach, we fitted the N–MAR (GMAR), SN–MAR, ST–MAR, SSL–MAR and SCN–MAR models with g = 1,2,3 components. The Bayesian model selection criteria for them are given in Table 2. The Bayesian estimates (posterior mean) of the ST–MAR (3,2) (the best fitted) model parameters for the return series of the I.T.C. stock data are given in Table 3.

5. Conclusions

In this paper, we have considered the robust MAR models based on the scale mixtures of skew normal distributions, called SMSN–MAR, and a Bayesian approach to estimate the models’ parameters. These models are used for modeling non-linear time series data, which involve GMAR models, and offer greater flexibility than normal distribution. Numerical studies verified the good performance of the considered models and suitability of Bayesian estimates. A real data example demonstrated that SMSN–MAR models can be useful tools for non-linear and non-stationary time series modeling. In future studies, the researchers can consider cyclostationary or almost cyclostationary processes [26,27,28,29,30,31,32,33] with SMSN errors or mixture models based on these processes.

Author Contributions

M.R.M.: Conceptualization, Methodology, Software, Supervision; M.M.: Data curation, Validation, Writing—Original draft preparation; D.B.: Visualization, Writing—Reviewing and Editing; V.-T.N.: Visualization, Writing—Reviewing and Editing; K.-H.P.: Visualization, Investigation, Writing—Reviewing and Editing. All authors have read and agreed to the published version of the manuscript

Funding

This research received no external funding.

Acknowledgments

We would like to thank the anonymous referees for their careful reading of this manuscript and also for their constructive suggestions which considerably improved the article.

Conflicts of Interest

The authors declare that they have no conflict of interest.

References

  1. Branco, M.D.; Dey, D.K. A general class of multivariate skew-elliptical distributions. J. Multivar. Anal. 2001, 79, 99–113. [Google Scholar] [CrossRef] [Green Version]
  2. Lindsay, B.G. Mixture models: Theory geometry and applications. In NSF-CBMS Regional Conference Series in Probability and Statistics; Institute of Mathematical Statistics: Hayward, CA, USA, 1995; Volume 51. [Google Scholar]
  3. Böhning, D. Computer-assisted analysis of mixtures and applications. In Meta-Analysis, Disease Mapping and Others; Chapman&Hall/CRC: Boca Raton, FL, USA, 2000. [Google Scholar]
  4. McLachlan, G.J.; Peel, D. Finite Mixture Models; Wiley: New York, NY, USA, 2000. [Google Scholar]
  5. Frühwirth-Schnatter, S. Finite Mixture and Markov Switching Models; Springer: New York, NY, USA, 2006. [Google Scholar]
  6. Mengersen, K.; Robert, C.P.; Titterington, D.M. Mixtures: Estimation and Applications; John Wiley and Sons: New York, NY, USA, 2011. [Google Scholar]
  7. Wong, C.S.; Li, W.K. On a mixture autoregressive model. J. R. Stat. Soc. Ser. B 2000, 62, 95–115. [Google Scholar] [CrossRef]
  8. Maleki, M.; Nematollahi, A.R. Autoregressive Models with Mixture of Scale Mixtures of Gaussian Innovations. Iran. J. Sci. Technol. Trans. Sci. 2017, 41, 1099–1107. [Google Scholar] [CrossRef]
  9. Ni, H.; Yin, H. A self-organizing mixture autoregressive network for time series modeling and prediction. Neurocomputing 2008, 72, 3529–3537. [Google Scholar] [CrossRef]
  10. McCulloch, R.E.; Tsay, R.S. Statistical inference of macroeconomic time series via markov switching models. J. Time Ser. Anal. 1994, 15, 523–539. [Google Scholar] [CrossRef]
  11. Glasbey, C.A. Non-linear autoregressive time series with multivariate Gaussian mixtures as marginal distributions. J. R. Stat. Soc. Ser. C 2001, 50, 143–154. [Google Scholar] [CrossRef]
  12. Wong, C.S.; Li, W.K. On a logistic mixture autoregressive model. Biometrika 2001, 88, 833–846. [Google Scholar] [CrossRef]
  13. Wong, C.S.; Chan, W.S. A student t-mixture autoregressive model with applications to heavy tailed financial data. Biometrika 2009, 96, 751–760. [Google Scholar] [CrossRef] [Green Version]
  14. Andrews, D.F.; Mallows, C.L. Scale mixtures of normal distributions. J. R. Stat. Soc. Ser. B 1974, 36, 99–102. [Google Scholar] [CrossRef]
  15. Maleki, M.; Wraith, D.; Arellano-Valle, R.B. Robust finite mixture modeling of multivariate unrestricted skew-normal generalized hyperbolic distributions. Stat. Comput. 2019, 29, 415–428. [Google Scholar] [CrossRef]
  16. Maleki, M.; Wraith, D. Mixtures of multivariate restricted skew-normal factor analyzer models in a Bayesian framework. Comput. Stat. 2019, 34, 1039–1053. [Google Scholar] [CrossRef]
  17. Zarrin, P.; Maleki, M.; Khodadadi, Z.; Arellano-Valle, R.B. Time series models based on the unrestricted skew-normal process. J. Stat. Comput. Sim. 2019, 89, 38–51. [Google Scholar] [CrossRef]
  18. Zeller, C.B.; Cabral CR, B.; Lachos, V.H. Robust mixture regression modeling based on scale mixtures of skew-normal distributions. Comput. Stat. Data Anal. 2010, 54, 2926–2941. [Google Scholar] [CrossRef]
  19. Maleki, M.; Wraith, D.; Arellano-Valle, R.B. A flexible class of parametric distributions for Bayesian linear mixed models. TEST 2019, 28, 543–564. [Google Scholar] [CrossRef]
  20. Wong, C.S.; Chan, W.S. Statistical Inference for Non-Linear Time Series Models. Ph.D. Thesis, University of Hong Kong, Hong Kong, China, 1998. [Google Scholar]
  21. Moravveji, B.; Khodadadi, Z.; Maleki, M. A Bayesian Analysis of Two-Piece Distributions Based on the Scale Mixtures of Normal Family. Iran. J. Sci. Technol. Trans. Sci. 2019, 43, 991–1001. [Google Scholar] [CrossRef]
  22. Brooks, S.P. Discussion on the paper by Spiegelhalter, Best, Carlin, and van der Linde. J. R. Stat. Soc. B 2002, 64, 616–618. [Google Scholar]
  23. Carlin, B.P.; Louis, T.A. Bayes and Empirical Bayes Methods for Data Analysis, 2nd ed.; Chapman & Hall/CRC: Boca Raton, FL, USA, 2001. [Google Scholar]
  24. Barndorff-Nielsen, O. Normal inverse Gaussian distributions and stochastic volatility modelling. Scand. J. Stat. 1997, 24, 1–13. [Google Scholar] [CrossRef]
  25. McNeil, A.J.; Frey, R.; Embrechts, P. Quantitative Risk Management: Concepts, Techniques and Tools; Princeton University Press: Princeton, NJ, USA, 2005. [Google Scholar]
  26. Mahmoudi, M.R.; Nematollahi, A.R.; Soltani, A.R. On the detection and estimation of the simple harmonizable processes. Iran. J. Sci. Technol. A (Sci.) 2015, 39, 239–242. [Google Scholar]
  27. Mahmoudi, M.R.; Maleki, M. A new method to detect periodically correlated structure. Comput. Stat. 2017, 32, 1569–1581. [Google Scholar] [CrossRef]
  28. Nematollahi, A.R.; Soltani, A.R.; Mahmoudi, M.R. Periodically correlated modeling by means of the periodograms asymptotic distributions. Stat. Pap. 2017, 58, 1267–1278. [Google Scholar] [CrossRef]
  29. Mahmoudi, M.R.; Heydari, M.H.; Avazzadeh, Z.; Pho, K.H. Goodness of fit test for almost cyclostationary processes. Digit. Signal Process. 2020, 96, 102597. [Google Scholar] [CrossRef]
  30. Mahmoudi, M.R.; Heydari, M.H.; Roohi, R. A new method to compare the spectral densities of two independent periodically correlated time series. Math. Comput. Simulat. 2019, 160, 103–110. [Google Scholar] [CrossRef]
  31. Mahmoudi, M.R.; Heydari, M.H.; Avazzadeh, Z. Testing the difference between spectral densities of two independent periodically correlated (cyclostationary) time series models. Commun. Stat. Theory Methods 2019, 48, 2320–2328. [Google Scholar] [CrossRef]
  32. Mahmoudi, M.R.; Maleki, M.; Pak, A. Testing the Difference between Two Independent Time Series Models. Iran. J. Sci. Technol. A (Sci.) 2017, 41, 665–669. [Google Scholar] [CrossRef]
  33. Mahmoudi, M.R.; Heydari, M.H.; Avazzadeh, Z. On the asymptotic distribution for the periodograms of almost periodically correlated (cyclostationary) processes. Digit. Signal Process. 2018, 81, 186–197. [Google Scholar] [CrossRef]
Figure 1. EAIC criterion for 200 MAR samples based on the GH distributions.
Figure 1. EAIC criterion for 200 MAR samples based on the GH distributions.
Symmetry 12 00929 g001
Figure 2. Time series plot of the I.T.C. stock from 2011-07-02 up to 2013-07-02.
Figure 2. Time series plot of the I.T.C. stock from 2011-07-02 up to 2013-07-02.
Symmetry 12 00929 g002
Figure 3. Histogram of the I.T.C. stock from 2011-07-02 up to 2013-07-02.
Figure 3. Histogram of the I.T.C. stock from 2011-07-02 up to 2013-07-02.
Symmetry 12 00929 g003
Figure 4. Plot of returned I.C.T stock data.
Figure 4. Plot of returned I.C.T stock data.
Symmetry 12 00929 g004
Table 1. Means and standard deviations of the Bayesian estimates of the simulated SMSN–MAR(2,2) models.
Table 1. Means and standard deviations of the Bayesian estimates of the simulated SMSN–MAR(2,2) models.
ModelSN–MARST–MARSSL–MARSCN–MAR
Parameters (Values)MeanSDMeanSDMeanSDMeanSD
φ 1 , 0 ( 0.1 ) 0.10030.021310.10010.027430.10040.023090.10100.02301
φ 2 , 0 ( 0.1 ) 0.10020.022420.10070.024360.10050.024190.10120.02502
φ 1 , 1 ( 0.5 ) 0.50170.014830.50110.017380.50100.013330.50330.01501
φ 1 , 2 ( 0.3 ) 0.30110.017460.30090.018460.30240.019630.30320.01758
φ 2 , 1 ( 0.6 ) 0.60320.015920.60480.019770.60240.018460.61010.01610
φ 2 , 2 ( 0.2 ) 0.21010.020320.20970.020060.20080.019260.21120.02094
σ 1 2 ( 1 ) 1.10230.020641.09690.317411.10210.274931.10260.02080
σ 2 2 ( 2 ) 2.19780.030272.20190.289502.19920.311712.20110.03311
λ 1 ( 2 ) 2.01840.816512.00250.945611.93670.828462.02210.82331
λ 2 ( 4 ) 4.00381.098173.90140.959283.99300.905634.01121.09831
ν 1 ( 3 ) ----3.89570.568423.94731.145870.52110.01918
ν 2 ( 3 ) ----3.89570.468773.94731.245870.51920.02111
γ 1 ( 0.5 ) ------------0.51030.02013
γ 2 ( 0.5 ) ------------0.52010.02321
π 1 ( 0.4 ) 0.40110.041130.40080.027950.39570.017580.40110.04113
Table 2. Bayesian model selection criteria for the proposed SMSN–MAR(3,2) with various number of components g = 1 , 2 , 3 .
Table 2. Bayesian model selection criteria for the proposed SMSN–MAR(3,2) with various number of components g = 1 , 2 , 3 .
ModelNumber of Component (g)EAICEBIC
N-MAR13290.63523297.8210
23067.34563291.9571
33175.47583210.9882
SN-MAR13262.73643271.7441
23038.74643165.0641
33175.73873183.7251
ST-MAR12979.37482986.2951
22816.83742887.3601
32910.63642977.6781
SSL-MAR13176.36473240.2990
22997.06943099.8811
33146.75643157.6441
SCN-MAR13164.75643191.3452
23009.37453097.3383
33165.75643182.6013
Note: The best values (between all possible models) are indicated in bold.
Table 3. Bayesian estimates with standard deviations of the parameters of fitted ST–MAR (3,2) model for the return series of the I.T.C. stock data.
Table 3. Bayesian estimates with standard deviations of the parameters of fitted ST–MAR (3,2) model for the return series of the I.T.C. stock data.
ComponentParametersBayesian EstimatesS.E.
First AR Component π 1 0.5802 0.0156
φ 1 , 0 0.0046 0.0100
φ 1 , 1 0.1478 0.0304
φ 1 , 2 0.0487 0.0102
φ 1 , 3 0.0238 0.0011
σ 1 2 0.0036 0.0032
λ 1 1.0312 0.0193
ν 2.0034 0.0722
Second AR Component π 2 0.4198 0.0121
φ 2 , 0 0.0030 0.0013
φ 2 , 1 0.2256 0.0345
σ 2 2 1.0679 0.0094
λ 2 0.9017 0.0435
ν 2.0042 0.0847

Share and Cite

MDPI and ACS Style

Mahmoudi, M.R.; Maleki, M.; Baleanu, D.; Nguyen, V.-T.; Pho, K.-H. A Bayesian Approach to Heavy-Tailed Finite Mixture Autoregressive Models. Symmetry 2020, 12, 929. https://doi.org/10.3390/sym12060929

AMA Style

Mahmoudi MR, Maleki M, Baleanu D, Nguyen V-T, Pho K-H. A Bayesian Approach to Heavy-Tailed Finite Mixture Autoregressive Models. Symmetry. 2020; 12(6):929. https://doi.org/10.3390/sym12060929

Chicago/Turabian Style

Mahmoudi, Mohammad Reza, Mohsen Maleki, Dumitru Baleanu, Vu-Thanh Nguyen, and Kim-Hung Pho. 2020. "A Bayesian Approach to Heavy-Tailed Finite Mixture Autoregressive Models" Symmetry 12, no. 6: 929. https://doi.org/10.3390/sym12060929

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop