EEG Signal Processing and Machine Learning. Saeid Sanei

Читать онлайн книгу.

EEG Signal Processing and Machine Learning - Saeid Sanei


Скачать книгу
rel="nofollow" href="#fb3_img_img_9e96dfb3-f642-55ed-b0eb-141a293943bf.png" alt="equation"/>

      and

      (3.54)equation

      (3.55)equation

      where αj and βj are the nonlinear model coefficients. The second term (first sum) in the right side corresponds to a qth order moving average (MA) dynamical noise term and the third term (second sum) corresponds to an AR model of order p. It is seen that the current conditional variance of the residual at time sample n depends on both its previous sample values and previous variances.

      Although in many practical applications such as forecasting of stock prices the orders p and q are set to small fixed values such as (p,q) = (1,1); for a more accurate modelling of natural signals such as EEGs the orders have to be determined mathematically. The prediction coefficients for various GARCH models or even the nonlinear functions g and h are estimated iteratively as for the linear ARMA models [36, 37].

      Such simple GARCH models are only suitable for multiplicative nonlinear dependence. In addition, additive dependencies can be captured by extending the modelling approach to the class of GARCH‐M models [40].

      Another limitation of the above simple GARCH model is failing to accommodate sign asymmetries. This is because the squared residual is used in the update equations. Moreover, the model cannot cope with rapid transitions such as spikes. Considering these shortcomings, numerous extensions to the GARCH model have been proposed. For example, the model has been extended and refined to include the asymmetric effects of positive and negative jumps such as the exponential GARCH model EGARCH [41], the GJR‐GARCH model [42], the threshold GARCH model (TGARCH) [43], the asymmetric power GARCH model APGARCH [44], and quadratic GARCH model QGARCH [45].

      (3.56)equation

      where b, α 1, α 2, and κ are constants and ηn is an indicator function that is zero when un is negative and one otherwise.

      Despite modelling the signals, the GARCH approach has many other applications. In some recent works [46] the concept of GARCH modelling of covariance is combined with Kalman filtering to provide a more flexible model with respect to space and time for solving the inverse problem. There are several alternatives for solution to the inverse problem. Many approaches fall into the category of constrained least‐squares methods employing Tikhonov regularization [47]. Among numerous possible choices for the GARCH dynamics, the EGARCH model [41] has been used to estimate the variance parameter of the Kalman filter iteratively.

      Nonlinear models have not been used for EEG processing. To enable use of these models the parameters and even the order should be adapted to the EEG properties. Also, such a model should incorporate the changes in the brain signals due to abnormalities and onset of diseases.

      3.4.3 Gaussian Mixture Model

      In this very popular modelling approach the signals are characterized using the parameters of their distributions. The distributions in terms of probability density functions are sum of a number of Gaussian functions with different variances which are weighted and delayed differently [48]. The overall distribution subject to a set of K Gaussian components is defined as:

      (3.57)equation

      The vector of unknown parameters θk = [wk , μk , σk ] for k = 1,2, …, K. wk is equivalent to the probability (weighting) that the data sample is generated by the kth mixture component density subject to:

      (3.58)equation

      μk , and σk are mean and variances of the kth Gaussian distribution and p(x| μk , σk ) is a Gaussian function of x with parameters μk , and σk . Expectation maximization (EM) [49] is often used to estimate the above parameters by maximizing the log‐likelihood of the mixture of Gaussian (MOG) for an N‐sample data defined as:

      (3.59)equation

      The EM algorithm alternates between updating the posterior probabilities used for generating each data sample by the kth mixture component (in a so‐called E‐step) as:

      (3.60)equation

      and weighted maximum likelihood updates of the parameters of each mixture component (in a so‐called M‐step) as:

      (3.61)equation

      (3.62)equation

      (3.63)equation

Schematic illustration of mixture of Gaussian (dotted curves) models of a multimodal unknown distribution (bold curve).
Скачать книгу