jagomart
digital resources
picture1_Matrix Pdf 174307 | Matform3


 127x       Filetype PDF       File size 0.27 MB       Source: www.census.gov


File: Matrix Pdf 174307 | Matform3
matrix formulas for nonstationary arima signal extraction tucker mcelroy u s census bureau abstract the paper provides general matrix formulas for minimum mean squared error signal extrac tion for a ...

icon picture PDF Filetype PDF | Posted on 27 Jan 2023 | 2 years ago
Partial capture of text on file.
             Matrix Formulas for Nonstationary ARIMA Signal Extraction
                                    Tucker McElroy
                                  U.S. Census Bureau
                                      Abstract
                The paper provides general matrix formulas for minimum mean squared error signal extrac-
              tion, for a finitely sampled time series whose signal and noise components are nonstationary
              ARIMAprocesses. These formulas are quite practical; as well as being simple to implement on
              a computer, they make it possible to easily derive important general properties of the signal
              extraction filters. We also extend these formulas to estimates of future values of the unobserved
              signal, and show how this result combines signal extraction and forecasting.
          Keywords. ARIMA model, forecasting, linear filter, nonstationary time series, seasonal adjust-
          ment.
          1 Introduction
            We consider signal extraction for finitely-sampled nonstationary time series data that can be
          transformed to a mean zero, covariance stationary process by differencing operators. We suppose
          that the signal and noise components are nonstationary, and that differencing operators exist which
          transform the signal and noise components into weakly stationary time series. Signal extraction
          for infinitely large samples of such series has a long history, including Hannan (1967), Sobel (1967),
          Cleveland and Tiao (1976), Bell (1984), and Bell and Martin (2004). The unpublished report Bell
          and Hillmer (1988) – whose basic results are summarized in Bell (2004) – treats the finite sample
          case, presenting matrix formulas for the mean square optimal time-varying filters. One drawback
          of Bell and Hillmer’s approach is its need for the separate estimation of initial values of nonstation-
          ary signal and noise components, resulting in formulas that are awkward to implement. Pollock
          (2001) also relies upon estimation of these initial values (and assumes that the noise component
          is stationary). McElroy and Sutcliffe (2006) furnishes an improvement over the Bell and Hillmer
          (1988) formulas, but only for a specific type of unobserved components model. Assuming that the
          signal extraction problem is formulated as ARIMA signal plus ARIMA noise (we later describe how
          a multiple unobserved ARIMA components model can be recast into a two-component model), we
          show in this paper an especially simple formula that does not involve the estimation of initial val-
          ues. This novel result is the main content of the paper at hand. This and other formulas discussed
                                         1
          below are quite practical, in that they can be used to derive important general properties of the
          filters, while also being simple to implement on a computer.
           In the context of finite sample model-based signal extraction, one popular approach – that utilized
          in SEATS (see the TSW Manual by Maravall and Caparello (2004) available at www.bde.es) – has
          been to apply the bi-infinite filters of Bell (1984) (which are the generalizations of the Wiener-
          Kolmogorov signal extraction filters to the nonstationary case) to the finite sample extended with
          infinitely many forecasts and backcasts. The result can be calculated exactly with the aid of an
          algorithm of Tunnicliffe-Wilson (see Burman, 1980). While being easy to implement, this method
          does not readily reveal properties of finite sample filters. Moreover, it cannot produce correct finite-
          sample Mean Squared Errors (MSEs) for the signal estimates. Another approach is to construct the
          signal plus noise model in State Space Form (Durbin and Koopman, 2001) andapplytheappropriate
          state space smoother (Kailath, Sayed, and Hassibi, 2000) to extract the signal. Efficient algorithms
          exist to obtain the time-varying signal extraction filters from the state space smoother, if desired
          (Koopman and Harvey, 2003); of course, these methods provide recursive formulas rather than
          explicit algebraic formulas. Thus, the state space approach cannot reveal certain fundamental
          properties of the signal extraction filters that are obvious from the matrix formulas; see Section
          4.1 and 4.2. In addition, the matrix approach readily provides the full covariance matrix of the
          signal error, a quantity that is useful in certain applications; see Findley, McElroy, and Wills
          (2004). Hence, there is both need and appeal for having explicit, readily implemented matrix
          formulas for nonstationary signal extraction. Note also that one of the original motivations for
          the matrix approach to signal extraction of Bell and Hillmer (1988), was to provide a method of
          signal extraction for models that could not be put into state space form, e.g., long memory models.
          Although our results are presented in the ARIMA-model based framework, generalizations to long
          memory or heteroscedastic models are discussed as well.
           We first present background material on signal extraction. The main theoretical results are in
          Section 2. Section 3 extends these results to estimating an unobserved signal at future times. Some
          applications of the matrix formulas are provided in Section 4. Examples of optimal finite sample
          seasonal adjustment and trend filters, along with their gain functions, are discussed in Section 5.
          Derivations of the basic formulas are contained in the appendix.
          2  Matrix Formulas
           Consider a nonstationary time series Y that can be written as the sum of two possibly nonsta-
                                   t
          tionary components S and N , the signal and the noise:
                       t    t
                                    Y =S +N                         (1)
                                     t  t  t
                                       2
               Following Bell (1984), we let Y be an integrated process such that W = δ(B)Y is weakly stationary.
                                             t                                   t         t
               Here B is the backshift operator and δ(z) is a polynomial with all roots located on the unit circle
               of the complex plane (also, δ(0) = 1 by convention). This δ(z) is referred to as the differencing
               operator of the series, and we assume it can be factored into relatively prime polynomials δS(z)
               and δN(z) (i.e., polynomials with no common zeroes), such that the series
                                                U =δS(B)S         V =δN(B)N                                  (2)
                                                 t          t      t           t
               are mean zero weakly stationary time series, which are uncorrelated with one another. Note that
               δS = 1 and/or δN = 1 are included as special cases. (In these cases either the signal or the noise
               or both are stationary.) We let d be the order of δ, and d   and d   are the orders of δS and δN;
                                                                         S       N
               since the latter operators are relatively prime, δ = δS · δN and d = d + d .
                                                                                   S    N
                  There are many examples from econometrics and engineering that conform to this scheme. In the
               context of component estimation for seasonal time series, for example, the data typically consist of
               seasonal, trend, and irregular components (G´omez and Maravall, 2001):
                                                       Y =S +T +I.                                           (3)
                                                        t     t   t    t
               Alternatively, a cycle component (Durbin and Koopman, 2001) is included as well:
                                                     Y =S +T +C +I.                                          (4)
                                                      t    t    t    t   t
               Now although there are three or four unobserved components, we can always rewrite the models
               in terms of two components (signal and noise). For example, if we are interested in seasonally
               adjusting the series, we identify the seasonal component with the noise and the sum of the remaining
               components becomes the signal of interest. That is, N = S and S = T + I or S = T + C +
                                                                      t    t       t    t    t     t    t    t
               I , depending on whether (3) or (4) holds respectively.    Typically, the seasonal component is
                 t
               nonstationary with δN(z) = 1+z+z2+···z11 for monthly data, and the trend is nonstationary as
               well. For a twice-integrated trend (and assuming that the irregular and cycle are stationary, which
                                    S             2
               is usually the case) δ (z) = (1 − z) . If instead we assume (4) (with the same ARIMA component
               models) and are interested in estimating the business cycle, then S = C and N = S + T + I ;
                                                                                  t     t       t    t    t    t
               in this case δ (z) = 1 and δ (z) = (1 − z)(1 − z12). In this fashion, any number of multiple
                             S               N
               unobserved components can be reduced to two components, where the signal is defined as the sum
               of those components that are of interest, and the noise consists of the rest. We will see below in
               Section 5.2 that one can apply the matrix formulas for signal extraction without having to derive
               newARIMAmodelsforeachseparate combination of components. It shares this property with the
               Wiener-Kolmogorov approach of SEATS, which only requires a partitioning of the autocovariance
               generating function for the observed series.
                                                               3
                    As in Bell and Hillmer (1988), we assume Assumption A of Bell (1984) holds for the component
                  decomposition, and we treat the case of a finite sample with t = 1,2,··· ,n for n > d. Assumption
                  Astates that the initial d values of Y , i.e., the variables Y = (Y ,Y ,··· ,Y ), are independent of
                                                            t                       ∗      1   2        d
                  {U } and {V }. For a discussion of the implications of this assumption, see Bell (1984) and Bell
                     t          t
                  and Hillmer (1988); in particular Bell (1984) discusses how the initial values for both the signal
                  and noise components may be generated such that Assumption A holds. Note that mean square
                  optimal signal extraction filters derived under Assumption A agree exactly with the filters implicitly
                  used by a properly initialized state space smoother, see Bell and Hillmer (1991, 1992). A further
                  assumption made is that {U } and {V } are uncorrelated with one another.
                                                 t          t
                    Now we can write (2) in a matrix form, as follows. Let ∆ be a (n − d) × n matrix with entries
                                =          (with the convention that       =0if        0 or       ). The analogously defined
                  given by ∆       δ                                    δ         k <       k > d
                              ij    i−j+d                                k
                  matrices ∆ and ∆ have entries given by the coefficients of δS(z) and δN(z), but are (n−d )×n
                              S        N                                                                                  S
                  and (n−d )×n dimensional respectively. This means that each row of these matrices consists of
                              N
                  the coefficients of the corresponding differencing polynomial, horizontally shifted in an appropriate
                  fashion. Hence
                                                   W=∆Y            U =∆ S          V =∆ N
                                                                           S               N
                  where Y is the transpose of (Y ,Y ,··· ,Y ), and W, U, V, S, and N are analogously defined. We
                                                    1   2        n
                  will denote the transpose of Y by Y′. To express
                                                           W =δN(B)U +δS(B)V                                                 (5)
                                                             t            t            t
                  in matrix form we need to define further differencing matrices ∆            and ∆ with row entries δN
                                                                                         N         S                     i−j+dN
                  and δS        given by the coefficients of δN(z) and δS(z) respectively, which are (n−d)×(n−d )
                        i−j+dS                                                                                                S
                  and (n−d)×(n−dN) dimensional. It follows from Lemma 1 of McElroy and Sutcliffe (2006) that
                                                            ∆=∆ ∆ =∆ ∆ .                                                     (6)
                                                                    N S        S N
                  Then we can write down the matrix version of (5):
                                                              W=∆ U+∆ V                                                      (7)
                                                                       N        S
                                                                                                                  ˆ
                  For each 1 ≤ t ≤ n, the minimum mean squared error signal extraction estimate is S = E[S |Y].
                                                                                                                   t       t
                  This can be expressed as a certain linear function of the data vector Y when the data are Gaussian.
                  This estimate is also the minimum mean squared error linear estimate when the data is non-
                  Gaussian. For the remainder of the paper, we do not assume Gaussianity, and by optimality we
                                                                                                        ˆ      ˆ ˆ          ˆ ′
                  always refer to the minimum mean squared error linear estimate. Writing S = (S ,S ,··· ,S ) ,
                                                                                                                1   2        n
                  the coefficients of these linear functions form the rows of a matrix F:
                                                                     ˆ
                                                                    S =FY
                                                                        4
The words contained in this file might help you see if this file matches what you are looking for:

...Matrix formulas for nonstationary arima signal extraction tucker mcelroy u s census bureau abstract the paper provides general minimum mean squared error extrac tion a nitely sampled time series whose and noise components are arimaprocesses these quite practical as well being simple to implement on computer they make it possible easily derive important properties of lters we also extend estimates future values unobserved show how this result combines forecasting keywords model linear lter seasonal adjust ment introduction consider data that can be transformed zero covariance stationary process by dierencing operators suppose exist which transform into weakly innitely large samples such has long history including hannan sobel cleveland tiao bell martin unpublished report hillmer basic results summarized in treats nite sample case presenting square optimal varying one drawback approach is its need separate estimation initial nonstation ary resulting awkward pollock relies upon assumes co...

no reviews yet
Please Login to review.