Artificial Intelligence and Quantum Computing for Advanced Wireless Networks. Savo G. Glisic
Читать онлайн книгу.(3.48) is simplified to yield an autoregressive (AR(p)) filter
which is also an IIR filter. The filter described by Eq. (3.50) is the basis for modeling the speech generating process. The presence of feedback within the AR(p) and ARMA (p, q) filters implies that selection of the ai, i = 1, 2, … , p, coefficients must be such that the filters are bounded input bounded output (BIBO) stable. The most straightforward way to test stability is to exploit the
(3.51)
To guarantee stability, the p roots of the denominator polynomial of (z), that is, the values of z for which D(z) = 0, the poles of the transfer function, must lie within the unit circle in the z‐plane, ∣z ∣ < 1.
Nonlinear predictors: If a measurement is assumed to be generated by an ARMA (p, q) model, the optimal conditional mean predictor of the discrete time random signal {y(k)}
(3.52)
is given by
where the residuals ê
where Θ(·) is an unknown differentiable zero‐memory nonlinear function. Notice e(k) is not included within Θ(·) as it is unobservable. The term NARMA (p, q) is adopted to define Eq. (3.54), since except for the (k), the output of an ARMA (p, q) model is simply passed through the zero‐memory nonlinearity Θ(·).
The corresponding NARMA (p, q) predictor is given by
(3.55)
where the residuals ê
(3.56)
and its associated predictor is
(3.57)
The two predictors are shown together in Figure 3.10, where it is clearly indicated which parts are included in a particular scheme. In other words, feedback is included within the NARMA (p, q) predictor, whereas the NAR(p) predictor is an entirely feedforward structure. In control applications, most generally, NARMA (p, q) models also include also external (exogeneous) inputs, (k − s), s = 1, 2, … , r, giving
Figure 3.10 Nonlinear AR/ARMA predictors.
(3.58)
and referred to as a NARMA