As indicated in Chapter 1, although the ordering of a time series is usually through time, it may also be ordered through space. If distances of spaces between observations are equal, the time series models and methods discussed earlier are equally applicable. However, in practice, we often see the phenomenon where regular time series are observed in different regions, and the point of interest is the study of the spatial patterns of these series. For example, in studying a traffic pattern of a certain location, traffic flow data is collected at constant intervals of time not only at that location but also at neighboring locations. In studying the prices of a certain commodity in a city, we not only want to study the price series of that commodity in that city but also the prices of the same commodity in nearby cities. In this chapter, we will present methods and models that can be used to describe spatial time series that are related to both time and space.
One way to model the space–time series is to treat them as a vector time series where the ith component corresponds to the ith location. Thus,
One can use the methods that were introduced in Chapter 2 to construct a vector autoregressive moving average VARMA(p,q) model or a vector autoregressive VAR(p) model for the series and use it to forecast and make inferences.
Alternatively, by classifying the spaces (locations) into different groups (regions), these spatial series are often measurements of the same phenomenon from several locations. For example, in studying the prices of a certain commodity in a city, we may divide the city into different regions where each region includes certain city blocks and the measurements are prices of the commodity under study. Hence, we can treat the spatial series at different locations in these groups (regions) at time t as repeated measurements at time t.
More specifically, we can present the spatial series for time from 1 to n as (n × 1) vector time series Zi,j, where i refers to the ith group (region), and j refers to the jth location. If the purpose of the study is to find out whether there are any differences between groups (regions), we can use the methods discussed in the repeated measurements in Chapter 7 to analyze the data set, where locations in each group (region) constitute a random sample.
In this chapter, we will introduce another modeling method specifically designed for space–time series and illustrate the method with detailed examples.
In this section, we will discuss an approach where we build a model that deals not only the serial dependence but also spatial dependence between observations at each time point. Specifically, we will introduce a class of time series models where the structure of spatial dependence is incorporated into the commonly used ARIMA(p,d,q) models, which are found to be useful in many applications.
To study space and time dependence, first we need to specify the order of spatial neighbors that reflects their distances to a particular location. Conventionally, the first order neighbors are closer than the second order ones, and the second order neighbors are closer than the third order neighbors, and so on. A standard definition of spatial order is sometimes available, but other times a model builder may need to define these orders so that they are suitable to the particular application.
Suppose that there are a total of m locations and we let Zt = [Z1,t, Z2,t, … , Zm,t]′ be the vector of times series of these m locations. Based on the spatial orders, with respect to the time series at location i, we will assign weights related to this location, such that its value is nonzero only when the location j is the ℓth order neighbors of i, and the sum of these weights is equal to 1. In other words, with respect to location i, we have where
Combining these weights, for all m locations, we have the spatial weight matrix for the neighborhood, which is a m × m matrix with being nonzero if and only if locations i and j are in the same ℓth order neighbor and each row sums to 1. The weight can be chosen to reflect physical properties such as border length or distance of neighboring locations. One can also assign equal weights to all the locations of the same spatial order. Clearly, W(0) = I, an identity matrix, because each location is its own zeroth order neighbor.
If we define the first order neighborhoods of a region to be those sharing a border with it, the second order neighborhoods of a region to be those sharing a border with the first order neighborhoods, and the third order neighborhoods of a region to be those sharing a border with the second order neighborhoods, then for the northeast states of USA given in Figure 8.1, we will have Table 8.1 of the three‐order neighborhoods.
Table 8.1 Spatial neighborhoods of 11 states in the Northeast United States.
Region (m) | First Order Neighborhood |
Second Order Neighborhood |
Third Order Neighborhood |
CT(1) | MA, NY, RI | NH, NJ, PA, VT | DE, MD, ME |
DE(2) | MD, NJ, PA | NY | CT, MA, VT |
MA(3) | CT, NH, NY, RI, VT | ME, NJ, PA | DE, MD |
MD(4) | DE, PA | NJ, NY | CT, MA, VT |
ME(5) | NH | MA, VT | CT, NY, RI |
NH(6) | MA, ME, VT | CT, NY, RI | NJ, PA |
NJ(7) | DE, NY, PA | CT, MA, MD, VT | NH, RI |
NY(8) | CT, MA, NJ, PA, VT | DE, MD, NH, RI | ME |
PA(9) | DE, MD, NJ, NY | CT, MA, VT | NH, RI |
RI(10) | CT, MA | NH, NY, VT | ME, NJ, PA |
VT(11) | MA, NH, NY | CT, ME, NJ, PA, RI, | DE, MD |
If we assign equal weight to all the locations of the same spatial order, the weighting matrix of W(1) will be
Similarly, we have
and
⊲
Thus, incorporating the structure of spatial dependence into a vector VARIMA model, we have the following space–time autoregressive integrated moving average (STARIMA) model,
where
at is a Gaussian vector white noise process with zero‐mean vector 0, and covariance matrix structure
and Σ is a m × m symmetric positive definite matrix. D(B) is a proper differencing operator defined similarly as in Eq. (2.59) of Chapter 2. In this application, since component series are observations of the same nature in different locations, the differencing operator could simply be D(B) = (1 − B)dI. ϕk,ℓ and θk,ℓ are autoregressive and moving average parameters at time lag k and space lag ℓ. respectively. p is the autoregressive order, q is the moving average order, ak is the spatial order for the kth autoregressive term, and mk is the spatial order for the kth moving average term.
It should be noted that the variance of each time series Zi,t is assumed to be constant. If not, a suitable variance stabilization transformation is assumed to be used to obtain Zi,t. For simplicity, we will simply denote the model as a model.
The STARIMA model in Eq. (8.3) can be extended to a seasonal STARIMA model that contains both seasonal and non‐seasonal AR and MA polynomials as follows,
where
D(Bs) = (1 − Bs)DI, and s is a seasonal period. αk,ℓ and βk,ℓ are seasonal autoregressive and seasonal moving average parameters at time lag k and space lag ℓ, respectively. P is the seasonal autoregressive order, Q is the seasonal moving average order, λk is the spatial order for the kth seasonal autoregressive term, and τk is the spatial order for the kth seasonal moving average term. For simplicity, we will denote the seasonal STARIMA model with a seasonal period s in Eq. (8.5) as
For a zero‐mean stationary spatial time series, the model in Eq. (8.3) reduces to the following space–time autoregressive moving average model,
where the zeros of lie outside the unit circle.
The model becomes a space–time autoregressive () model when q = 0. It becomes a space–time moving average () model when p = 0. The STAR models were first introduced by Cliff and Ord (1975a, 1975b), and Martin and Oeppen (1975), and further extended to STARMA models by Pfeifer and Deutsch (1980a, b, c).
Since a stationary model can be approximated by an autoregressive model, because of its easier interpretation, the most widely used STARMA models in practice are models,
where Zt is a zero‐mean stationary spatial time series or a proper transformed and differenced series of a nonstationary spatial time series.
Similar to the vector time series models, the important characteristics of STARMA models are shown through their space–time autocorrelation function (ST‐ACF) and space–time partial autocorrelation function (ST‐PACF).
Without loss of generality, we will assume a zero‐mean spatial time series. To properly incorporate the space weighting matrices in the analysis, the ST‐ACF for ℓth neighbors at time lag k is given by
where the space–time autocovariance function between ith and jth neighbors at time lag k for the m‐dimensional space–time process is defined by:
The sample ST‐ACF, is thus given by
Similar to the ARMA models, we can use space–time analog of the Yule–Walker equations to derive the ST‐PACF. Specifically, in Eq. (8.7), we replace a1, …, and ap by λ = max(a1, …, ap). In other words, we consider the following STAR(pλ,…, λ) model
Pre‐multiplying both sides of Eq. (8.14) by [W(h)Zt − s]′ and taking expectation, we have the space–time analog of the Yule–Walker equations,
The ST‐PACF is obtained as the last coefficient, ϕp,ℓ, for ℓ = 0, 1, …, λ, from the solution of the successively solving the system of equations for each p = 1, 2, …, s = p, each set of ℓ = {0}, {0, 1}, …, {0, 1, …λ}, and h = {0}, {0, 1}, …, {0, 1, …λ}.
As an illustration, let us first consider the computation of some specific STAR models.
Sample ST‐ACF Table
Time lag/Space Lag | 0 | 1 | 2 | 3 | … |
1 | … | ||||
2 | … | ||||
3 | … | ||||
4 | … | ||||
. | . | . | . | . | . |
. | . | . | . | . | . |
Sample ST‐ACF Graph
Sample ST‐PACF Table
Time lag/Space Lag | 0 | 1 | 2 | 3 | … |
1 | … | ||||
2 | … | ||||
3 | … | ||||
4 | … | ||||
. | . | . | . | . | . |
. | . | . | . | . | . |
Sample ST‐PACF Graph
Under the assumption that the at are i.i.d. random vectors with E(at) = 0 and variance–covariance matrix σ2I, the variance of the sample ST‐ACF and ST‐PACF can be approximated by
and
where m is the number of locations and n is the length of the time series in each location.
It should be pointed out that the parameters ϕk,ℓ and θk,ℓ in the STARMA or STARIMA models discussed in Section 8.2 are all scalar, which is the same for all locations. To remove this undesirable restriction, the generalized STAR models, denoted by GSTAR models, were first introduced by Borovkova, Lopuhaa, and Ruchjana (2002), and later extended to the following generalized STARMA, denoted by GSTARMA models, by Di Giacinto (2006),
where the assumption of at is the same as Eq. (8.3). However, the time lag k and space lag ℓ parameters, Φk,ℓ and Θk,ℓ, are diagonal matrices instead of scalars. These parameters can vary depending on locations and hence offer more flexibility.
More generally, the model in Eq. (8.18) can be extended to generalized models,
where
and
To include a seasonal phenomenon, we can also extend the seasonal model in Eq. (8.5) to the following generalized seasonal model,
where
and
The coefficients, αk,ℓ and βk,ℓ, are (m × m) diagonal matrices of parameters at time lag k and space lag ℓ.
As shown in Example 8.2, a STAR(11) model in Eq. (8.8) is equivalent to a first order vector autoregressive VAR(1) model in Eq. (8.10), that is
Thus, one can treat STARMA models as special cases of the vector VARMA models, where the diagonal elements of the m × m autoregressive and moving average parameter matrices are assumed to be equal because the m series represent a single process operating at different locations and the off‐diagonal elements are assumed to be a linear combination of the W(ℓ) weighting matrices. Obviously, GSTARMA models are also special cases of vector VARMA models.
Similar to the vector VARMA models, STARMA and GSTARMA models are characterized by the ST‐ACF and ST‐PACF. For a STAR or a GSTAR model, its ST‐PACF cuts off after certain temporal and space lags. Similarly, for a STMA or a GSTMA model, its ST‐ACF cuts off after certain temporal and space lags. If neither of these ST‐ACF and ST‐PACF cuts off over time and space, it will be a STARMA or a GSTARMA model.
Once a STARMA model is identified, the parameters are estimated by minimizing
using various nonlinear optimization techniques, which is equivalent to maximize the likelihood function of
discussed in Chapter 2. The insignificant parameters should be removed and a simpler model be re‐estimated until all parameters are statistically significant.
Before a model can be used for forecasting and making inferences, a diagnostic check must be carefully done to make sure that the residuals are approximately white noise. Specifically, we can calculate the residual sample ST‐ACF and ST‐PACF and compare with their estimated standard deviation. If they are not white noise, based on the results of the residual analysis, we need to re‐identify a new model, re‐estimate the parameters, and perform a diagnostic check, which constitutes an iterative process.
Bordering districts are defined as first neighbors, and the weighting matrix for W(1) is given in Table 8.2. Models involving additional spatial lags were considered but the parameter estimates were not significant different from 0, and hence the weighting matrices W(ℓ) for ℓ > 1 are not given.
The patterns in Figure 8.6 show that the sample ST‐ACFs exponentially decay at both regular time lags, seasonal time lags of multiple of 12, and that ST‐PACFs are significant only at time lags 2, 12, and 24, and spatial lag at 1. For a monthly series of seasonal period of 12, we tentatively choose a STAR(21,1) × (21,1)12 as its possible underlying model. The estimation result is given as follows:
Since the estimate of ϕ2,1 is only 1.46 standard errors away from 0, we re‐fit a STAR(21,0) × (21,1)12 model with the following result,
Now, the estimate of α2,0 is insignificant, so we reduce it to the following STAR(21,0) × (20,1)12 model,
Now all estimates are significant. In addition, the white noise pattern of the ST‐ACFs and ST‐PACFs of residuals from the fitted model shown in Figure 8.7 indicates that the model is adequate.
The sample ST‐ACF decays either exponentially or in a sinusoidal fashion. In addition, the sample ST‐PACFs cut‐off after time lag 1 and spatial lag 1, indicating a STAR(11) model.
Estimation through minimizing the trace of the error covariance matrix (the determinant was not easily worked with for such a large matrix) leads to the following STAR(11) model:
Both parameters are significant, and almost all the sample ST‐ACF and ST‐PACF of the model residual are small or at least within two standard errors of 0 as shown in Tables 8.5 and 8.6, indicating an adequate model.
Table 8.5 Sample ST‐ACF of residuals.
k | ||||||||
1 | 0.011 | (0.074) | −0.007 | (0.039) | 0.035 | (0.035) | −0.014 | (0.036) |
2 | 0.008 | (0.075) | 0.050 | (0.040) | 0.034 | (0.035) | 0.029 | (0.036) |
3 | −0.030 | (0.076) | −0.047 | (0.040) | −0.027 | (0.036) | 0.022 | (0.037) |
4 | −0.017 | (0.077) | −0.001 | (0.041) | −0.027 | (0.036) | −0.009 | (0.037) |
5 | −0.134 | (0.078) | −0.014 | (0.041) | 0.014 | (0.037) | 0.097 | (0.038) |
6 | −0.107 | (0.079) | 0.011 | (0.042) | −0.022 | (0.037) | 0.043 | (0.039) |
7 | 0.007 | (0.081) | −0.010 | (0.043) | 0.030 | (0.038) | 0.051 | (0.039) |
8 | 0.117 | (0.082) | 0.039 | (0.043) | 0.051 | (0.039) | −0.024 | (0.040) |
9 | −0.097 | (0.083) | −0.002 | (0.044) | −0.002 | (0.039) | 0.071 | (0.041) |
10 | 0.164 | (0.085) | 0.014 | (0.045) | 0.009 | (0.040) | −0.051 | (0.041) |
Table 8.6 Sample ST‐PACF of residuals.
k | ||||||||
1 | 0.011 | (0.074) | −0.019 | (0.039) | 0.099 | (0.035) | −0.055 | (0.036) |
2 | 0.006 | (0.075) | 0.093 | (0.040) | 0.033 | (0.035) | 0.027 | (0.036) |
3 | −0.031 | (0.076) | −0.080 | (0.040) | −0.009 | (0.036) | 0.050 | (0.037) |
4 | −0.020 | (0.077) | 0.007 | (0.041) | −0.066 | (0.036) | −0.015 | (0.037) |
5 | −0.128 | (0.078) | 0.042 | (0.041) | 0.103 | (0.037) | 0.135 | (0.038) |
6 | −0.109 | (0.079) | 0.065 | (0.042) | −0.022 | (0.037) | 0.035 | (0.039) |
7 | 0.001 | (0.081) | −0.022 | (0.043) | 0.100 | (0.038) | 0.073 | (0.039) |
8 | 0.107 | (0.082) | 0.020 | (0.043) | 0.061 | (0.039) | 0.002 | (0.040) |
9 | −0.112 | (0.083) | 0.039 | (0.044) | −0.005 | (0.039) | 0.110 | (0.041) |
10 | 0.144 | (0.085) | −0.086 | (0.045) | −0.009 | (0.040) | 0.000 | (0.041) |
For more related discussions of Examples 8.5 and 8.6, we refer readers to a Ph.D. dissertation by Gehman (2016), and papers by Gehman and Wei (2014, 2015a, b).
The information in Table 8.8 is only for a one‐step ahead forecast. A table for a five‐step forecast may not be very practical. In many applications, cancer researchers are interested in the forecasts for the nine Morbidity and Mortality Weekly Report (MMWR) regions or four STD regions shown in Figure 8.10. They can easily be obtained from the forecasts from Eq. (8.28) through aggregation, and the five‐step forecasts are shown in Table 8.9 for the nine MMWR regions and Table 8.10 for the four STD regions, respectively.
Table 8.9 Five‐step forecasts for nine MMWR regions.
Forecast Step | Region | Forecast | Actual | Error | Forecast Step | Region | Forecast | Actual | Error |
1 | 1 | 0.769 | −1.102 | −1.871 | 3 | 6 | 0.107 | −0.072 | −0.179 |
1 | 2 | 0.190 | −0.497 | −0.687 | 3 | 7 | −0.007 | −0.761 | −0.754 |
1 | 3 | 0.109 | 0.359 | 0.250 | 3 | 8 | −0.386 | 0.209 | 0.595 |
1 | 4 | −0.286 | −0.633 | −0.347 | 3 | 9 | −0.138 | 1.938 | 2.076 |
1 | 5 | 1.093 | 1.213 | 0.120 | 4 | 1 | −0.085 | 2.272 | 2.357 |
1 | 6 | 0.323 | 0.295 | −0.028 | 4 | 2 | 0.150 | 0.468 | 0.318 |
1 | 7 | 0.273 | 0.988 | 0.715 | 4 | 3 | 0.641 | 1.322 | 0.681 |
1 | 8 | −0.132 | −1.324 | −1.192 | 4 | 4 | 0.828 | 1.887 | 1.059 |
1 | 9 | −0.274 | −0.996 | −0.722 | 4 | 5 | 0.661 | 1.286 | 0.625 |
2 | 1 | −0.202 | 3.936 | 4.138 | 4 | 6 | 0.420 | 0.275 | −0.145 |
2 | 2 | 0.243 | 0.374 | 0.131 | 4 | 7 | 0.565 | 0.388 | −0.177 |
2 | 3 | 1.186 | 1.967 | 0.781 | 4 | 8 | 0.962 | 2.822 | 1.860 |
2 | 4 | 1.399 | 0.402 | −0.997 | 4 | 9 | 0.441 | 1.444 | 1.003 |
2 | 5 | 1.046 | −0.126 | −1.172 | 5 | 1 | 0.256 | 1.861 | 1.605 |
2 | 6 | 0.737 | 0.038 | −0.699 | 5 | 2 | 0.055 | 0.962 | 0.907 |
2 | 7 | 0.938 | 0.224 | −0.714 | 5 | 3 | −0.092 | 0.806 | 0.898 |
2 | 8 | 1.820 | 1.521 | −0.299 | 5 | 4 | −0.170 | 9.339 | 9.509 |
2 | 9 | 0.446 | 1.835 | 1.389 | 5 | 5 | 0.182 | 3.097 | 2.915 |
3 | 1 | 0.431 | −1.537 | −1.968 | 5 | 6 | 0.024 | −0.243 | −0.267 |
3 | 2 | 0.100 | 0.270 | 0.170 | 5 | 7 | −0.066 | 1.825 | 1.891 |
3 | 3 | −0.074 | 0.488 | 0.562 | 5 | 8 | −0.320 | 1.776 | 2.096 |
3 | 4 | −0.168 | −0.692 | −0.524 | 5 | 9 | −0.072 | 1.426 | 1.498 |
3 | 5 | 0.521 | 1.540 | 1.019 |
Table 8.10 Five‐step forecasts for four STD regions.
Forecast Step | Region | Forecast | Actual | Error | Forecast Step | Region | Forecast | Actual | Error |
1 | 1 | −0.406 | −2.320 | −1.914 | 3 | 3 | 0.531 | −1.267 | −1.798 |
1 | 2 | −0.178 | −0.274 | −0.096 | 3 | 4 | 0.621 | 0.708 | 0.087 |
1 | 3 | 0.959 | −1.599 | −2.558 | 4 | 1 | 1.404 | 4.266 | 2.862 |
1 | 4 | 1.690 | 2.496 | 0.806 | 4 | 2 | 1.469 | 3.209 | 1.740 |
2 | 1 | 2.266 | 3.357 | 1.091 | 4 | 3 | 0.065 | 2.740 | 2.675 |
2 | 2 | 2.585 | 2.370 | −0.215 | 4 | 4 | 1.646 | 1.950 | 0.304 |
2 | 3 | 0.040 | 4.309 | 4.269 | 5 | 1 | −0.392 | 3.202 | 3.594 |
2 | 4 | 2.721 | 0.135 | −2.586 | 5 | 2 | −0.262 | 10.146 | 10.408 |
3 | 1 | −0.524 | 2.147 | 2.671 | 5 | 3 | 0.312 | 2.824 | 2.512 |
3 | 2 | −0.243 | −0.204 | 0.039 | 5 | 4 | 0.139 | 4.679 | 4.540 |
The forecasts for the nine MMWR regions and four STD regions shown in Tables 8.9 and 8.10 are examples of spatial aggregation. With the forecast from Eq. (8.28), the forecast for the nine regions is obtained by
where is a (44 × 1) vector, A is the (9 × 44) spatial aggregation matrix shown next by its transpose.
State | Row 1 | Row 2 | Row 3 | Row 4 | Row 5 | Row 6 | Row 7 | Row 8 | Row 9 |
CT | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
ME | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
MA | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
NH | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
RI | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
NJ | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
NY | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
DE | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 |
DC | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 |
MD | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 |
PA | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
VA | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 |
WV | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 |
AL | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 |
FL | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 |
GA | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 |
KY | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 |
MS | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 |
NC | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 |
SC | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 |
TN | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 |
IL | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 |
IN | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 |
MI | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 |
MN | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 |
OH | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 |
WI | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 |
AR | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 |
LA | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 |
NM | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 |
OK | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 |
TX | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 |
IA | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 |
KS | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 |
MO | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 |
NE | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 |
CO | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 |
UT | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 |
AZ | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 |
CA | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 |
NV | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 |
ID | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 |
OR | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 |
WA | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 |
Similarly, the forecast for the four regions is obtained by with the (4 × 44) spatial aggregation matrix A shown next by its transpose.
State | Row 1 | Row 2 | Row 3 | Row 4 |
CT | 0 | 0 | 1 | 0 |
ME | 0 | 0 | 1 | 0 |
MA | 0 | 0 | 1 | 0 |
NH | 0 | 0 | 1 | 0 |
RI | 0 | 0 | 1 | 0 |
NJ | 0 | 0 | 1 | 0 |
NY | 0 | 0 | 1 | 0 |
DE | 0 | 0 | 0 | 1 |
DC | 0 | 0 | 0 | 1 |
MD | 0 | 0 | 0 | 1 |
PA | 0 | 0 | 1 | 0 |
VA | 0 | 0 | 0 | 1 |
WV | 0 | 0 | 0 | 1 |
AL | 0 | 0 | 0 | 1 |
FL | 0 | 0 | 0 | 1 |
GA | 0 | 0 | 0 | 1 |
KY | 0 | 0 | 0 | 1 |
MS | 0 | 0 | 0 | 1 |
NC | 0 | 0 | 0 | 1 |
SC | 0 | 0 | 0 | 1 |
TN | 0 | 0 | 0 | 1 |
IL | 0 | 1 | 0 | 0 |
IN | 0 | 1 | 0 | 0 |
MI | 0 | 1 | 0 | 0 |
MN | 0 | 1 | 0 | 0 |
OH | 0 | 1 | 0 | 0 |
WI | 0 | 1 | 0 | 0 |
AR | 0 | 0 | 0 | 1 |
LA | 0 | 0 | 0 | 1 |
NM | 1 | 0 | 0 | 0 |
OK | 0 | 0 | 0 | 1 |
TX | 0 | 0 | 0 | 1 |
IA | 0 | 1 | 0 | 0 |
KS | 0 | 1 | 0 | 0 |
MO | 0 | 1 | 0 | 0 |
NE | 0 | 1 | 0 | 0 |
CO | 1 | 0 | 0 | 0 |
UT | 1 | 0 | 0 | 0 |
AZ | 1 | 0 | 0 | 0 |
CA | 1 | 0 | 0 | 0 |
NV | 1 | 0 | 0 | 0 |
ID | 1 | 0 | 0 | 0 |
OR | 1 | 0 | 0 | 0 |
WA | 1 | 0 | 0 | 0 |
Clearly, the nine‐region and four‐region forecasts can also be achieved by first getting the nine‐region and four‐region time series through spatial data aggregation,
and then computing the forecast using the aggregate models constructed from the aggregate series obtained from Eq. (8.31). We will leave this as a data analysis project for readers. The related theory and concept of aggregation will be discussed further in Chapter 10.
For further information on space–time models and applications, we refer readers to Stoffer (1986), Smirnov and Anselin (2001), Lichstein et al. (2002), Zhou and Buongiorno (2006), Craigmile and Guttorp (2011), Eshel (2012), Chang et al. (2014), Clements et al. (2014), Blangiardo and Cameletti (2015), Casals et al. (2016), Cseke et al. (2016), Jalbert et al. (2017), Lin et al. (2017), Rao and Terdik (2017), Shand and Li (2017), Quick et al. (2018), and Singh and Lalitha (2018), among others.
3.141.192.183