International Journal of Statistics and Applications
p-ISSN: 2168-5193 e-ISSN: 2168-5215
2022; 12(2): 30-41
doi:10.5923/j.statistics.20221202.02
Received: Feb. 12, 2022; Accepted: Mar. 2, 2022; Published: Mar. 15, 2022

Erhard Reschenhofer1, Barbara Katharina Reschenhofer2
1Department of Statistics and Operations Research, University of Vienna, Oskar-Morgenstern-Platz 1, Vienna, Austria
2Department of English and American Studies, University of Vienna, Spitalgasse 2-4, Vienna, Austria
Correspondence to: Erhard Reschenhofer, Department of Statistics and Operations Research, University of Vienna, Oskar-Morgenstern-Platz 1, Vienna, Austria.
| Email: |  | 
Copyright © 2022 The Author(s). Published by Scientific & Academic Publishing.
This work is licensed under the Creative Commons Attribution International License (CC BY). 
                    	http://creativecommons.org/licenses/by/4.0/
                    	
Choosing the value of 0.5 for the fractional differencing operator can be helpful for the determination of the stationarity of a time series. A pole at frequency zero of the spectral density of the fractionally differenced time series may indicate nonstationarity of the original time series (underdifferencing) whereas a vanishing spectral density at frequency zero may indicate stationarity of the original time series (overdifferencing). In addition to this frequency-domain analysis, it is advantageous to check in the time domain whether the autocorrelation function of the fractionally differenced time series is positive and slowly decaying. Unfortunately, carrying out fractional differencing is not a simple task unless the time series is extremely long, which is rarely the case in practice. We therefore propose a simple approximation which is based on a parsimonious ARMA(1,1) model. The new method is applied to climatological and socioeconomic datasets. The hypothesis of stationarity is rejected for the global surface temperature, economic growth, and migration.
Keywords: Nonstationarity, Fractional differencing, Root differencing, Global surface temperature, GDP per capita, Immigration, Emigration
Cite this paper: Erhard Reschenhofer, Barbara Katharina Reschenhofer, Investigating the Persistence of Shocks: Global Warming, Economic Growth and Migration, International Journal of Statistics and Applications, Vol. 12 No. 2, 2022, pp. 30-41. doi: 10.5923/j.statistics.20221202.02.
 which satisfies the difference equation
 which satisfies the difference equation |  | (1) | 
 is white noise with mean 0 and variance
 is white noise with mean 0 and variance  as
 as|  | (2) | 
 on the current value
 on the current value  is only temporary and vanishes as
 is only temporary and vanishes as  if
 if  In contrast, if
 In contrast, if  we have
 we have |  | (3) | 
 on
 on  is therefore persistent. In the former case,
 is therefore persistent. In the former case,  is a stationary autoregressive (AR) process of order 1 with variance
 is a stationary autoregressive (AR) process of order 1 with variance |  | (4) | 
 of size n, it is often extremely difficult to distinguish between the two cases. Indeed, there is hardly any difference between the two samples of size
 of size n, it is often extremely difficult to distinguish between the two cases. Indeed, there is hardly any difference between the two samples of size  shown in Figure 1.a
 shown in Figure 1.a  and Figure 1.c
 and Figure 1.c  respectively. At first glance, both look nonstationary. The situation improves when the sample size is increased to
 respectively. At first glance, both look nonstationary. The situation improves when the sample size is increased to  . While the sample from the random walk still looks nonstationary (see Figure 1.b), the sample from the AR process now looks quite stationary (see Figure 1.d). In practice, the dataset is given and can usually not be increased arbitrarily. Only if we are lucky and the parameter
. While the sample from the random walk still looks nonstationary (see Figure 1.b), the sample from the AR process now looks quite stationary (see Figure 1.d). In practice, the dataset is given and can usually not be increased arbitrarily. Only if we are lucky and the parameter  of the data generating process is sufficiently smaller than 1 for the given sample size, the unit root hypothesis
 of the data generating process is sufficiently smaller than 1 for the given sample size, the unit root hypothesis |  | (5) | 
|  | (6) | 
 we can always choose a suitable value for the parameter
 we can always choose a suitable value for the parameter  so that the sample will look stationary even if
 so that the sample will look stationary even if  (see Figure 1.h). This is due to the fact that the terms in the numerator and denominator of the lag polynomial representation
 (see Figure 1.h). This is due to the fact that the terms in the numerator and denominator of the lag polynomial representation |  | (7) | 
 is chosen only slighty greater than -1. Thus, we can never be sure whether a rejection of the unit root hypothesis is due to a small value of
 is chosen only slighty greater than -1. Thus, we can never be sure whether a rejection of the unit root hypothesis is due to a small value of  or a value of
 or a value of  close to -1 (for a more thorough line of reasoning see Pötscher, 2002).
 close to -1 (for a more thorough line of reasoning see Pötscher, 2002). |  | (8) | 
 is negative in both cases (see Figures 2.b and 2.f), only the second cumulative plot shows a clear downward trend. However, this visual significance is somewhat put into perspective when higher-order lags are also considered (see Figures 2.c and 2.g). All computations are carried out with the free statistical software R (R Core Team, 2018).Unfortunately, overdifferencing can also occur in the case of a nonstationary time series. When we consider the general class of autoregressive fractionally integrated moving average (ARFIMA) processes
 is negative in both cases (see Figures 2.b and 2.f), only the second cumulative plot shows a clear downward trend. However, this visual significance is somewhat put into perspective when higher-order lags are also considered (see Figures 2.c and 2.g). All computations are carried out with the free statistical software R (R Core Team, 2018).Unfortunately, overdifferencing can also occur in the case of a nonstationary time series. When we consider the general class of autoregressive fractionally integrated moving average (ARFIMA) processes |  | (9) | 
 and I(0) processes, i.e., processes that are integrated of order zero
 and I(0) processes, i.e., processes that are integrated of order zero  but there are also processes that are fractionally integrated. For stationarity, it is required that
 but there are also processes that are fractionally integrated. For stationarity, it is required that  Our goal is to distinguish stationary processes with
 Our goal is to distinguish stationary processes with  which include white noise processes as well as AR, MA and ARMA processes, from nonstationary processes with
 which include white noise processes as well as AR, MA and ARMA processes, from nonstationary processes with  which include random walks as well as ARIMA processes. We have seen above that a negative autocorrelation can be an indication of overdifferencing. Taking first differences reduces the memory parameter
 which include random walks as well as ARIMA processes. We have seen above that a negative autocorrelation can be an indication of overdifferencing. Taking first differences reduces the memory parameter  by 1. In the case of an I(1) process, the memory parameter decreases from one to zero and is therefore still nonnegative whereas in the case of an I(0) process, it decreases from 0 to -1. In contrast, both in the case of a fractionally integrated process with
 by 1. In the case of an I(1) process, the memory parameter decreases from one to zero and is therefore still nonnegative whereas in the case of an I(0) process, it decreases from 0 to -1. In contrast, both in the case of a fractionally integrated process with  which is stationary, and in the case of a fractionally integrated process with
 which is stationary, and in the case of a fractionally integrated process with  which is nonstationary, the memory parameter will become negative after differencing. Checking for negativity after differencing is therefore pointless. An obvious alternative is fractional differencing with the help of the fractional differencing operator, which is defined as a power series expansion in integer powers of
 which is nonstationary, the memory parameter will become negative after differencing. Checking for negativity after differencing is therefore pointless. An obvious alternative is fractional differencing with the help of the fractional differencing operator, which is defined as a power series expansion in integer powers of  i.e.,
 i.e.,|  | (10) | 
 in (10) will reduce the memory parameter by 0.5, hence we will observe overdifferencing exactly in the stationary case where the original order of integration is less than 0.5. Fractional differencing with
 in (10) will reduce the memory parameter by 0.5, hence we will observe overdifferencing exactly in the stationary case where the original order of integration is less than 0.5. Fractional differencing with  can also help to detect underdifferencing. For example, when a strong positive autocorrelation is not only present in the original, not differenced series (see, e.g., Figures 2.d and 2.h) but also after fractional differencing, albeit to a lesser extent.
 can also help to detect underdifferencing. For example, when a strong positive autocorrelation is not only present in the original, not differenced series (see, e.g., Figures 2.d and 2.h) but also after fractional differencing, albeit to a lesser extent.  by truncation, we will construct in Subsection 2.3 another approximation which is based on a parsimonious ARMA(1,1) model. But before that we will in the next subsection briefly leave the time domain and switch to the frequency domain where root differencing is a trivial exercise.
 by truncation, we will construct in Subsection 2.3 another approximation which is based on a parsimonious ARMA(1,1) model. But before that we will in the next subsection briefly leave the time domain and switch to the frequency domain where root differencing is a trivial exercise. |  | (11) | 
 either goes to infinity
 either goes to infinity  or to zero
 or to zero  Only in the case of a pure ARMA process
 Only in the case of a pure ARMA process  it converges (horizontally) to a positive number. After root differencing, the spectral density goes to zero if and only if
 it converges (horizontally) to a positive number. After root differencing, the spectral density goes to zero if and only if  i.e., exactly in the case of stationarity. In the frequency domain, root differencing can be accomplished simply by multiplying the spectral density by the factor
 i.e., exactly in the case of stationarity. In the frequency domain, root differencing can be accomplished simply by multiplying the spectral density by the factor |  | (12) | 
 Nonparametric estimates of the spectral density can be obtained by smoothing the periodogram
 Nonparametric estimates of the spectral density can be obtained by smoothing the periodogram|  | (13) | 
 A more elaborate method is the log periodogram regression which is based on the low-frequency approximation
A more elaborate method is the log periodogram regression which is based on the low-frequency approximation |  | (14) | 
 we must replace the spectral density in this approximation by the periodogram (see Geweke and Porter-Hudak, 1983) or a smoothed version of it (see Hassler, 1993, Peiris and Court, 1993, and Reisen, 1994) and choose a suitable neighborhood
 we must replace the spectral density in this approximation by the periodogram (see Geweke and Porter-Hudak, 1983) or a smoothed version of it (see Hassler, 1993, Peiris and Court, 1993, and Reisen, 1994) and choose a suitable neighborhood  of frequency zero. The parameter
 of frequency zero. The parameter  determines how many of the lowest Fourier frequencies are included in the regression (for a procedure involving non-Fourier frequencies, see Reschenhofer and Mangat, 2021). As always, there is a trade-off between bias and variance. A small value of
 determines how many of the lowest Fourier frequencies are included in the regression (for a procedure involving non-Fourier frequencies, see Reschenhofer and Mangat, 2021). As always, there is a trade-off between bias and variance. A small value of  increases the variance whereas a large value of
 increases the variance whereas a large value of  may introduce a bias caused by short-term autocorrelation. In order to illustrate the frequency-domain approach outlined above, we consider a financial application where
 may introduce a bias caused by short-term autocorrelation. In order to illustrate the frequency-domain approach outlined above, we consider a financial application where  is known a priori. Although stock price series cannot adequately be modeled by a random walk (and not even by a conditionally heteroskedastic random walk with drift), there is a broad agreement that they are integrated of order 1, hence
 is known a priori. Although stock price series cannot adequately be modeled by a random walk (and not even by a conditionally heteroskedastic random walk with drift), there is a broad agreement that they are integrated of order 1, hence  A disadvantage of financial datasets is that they are usually very short. Stock prices are rarely available for hundred years or more. A large sample size is not helping in this regard. E.g., for the investigation of climate change, a long annual temperature series from 1850 to 2020
 A disadvantage of financial datasets is that they are usually very short. Stock prices are rarely available for hundred years or more. A large sample size is not helping in this regard. E.g., for the investigation of climate change, a long annual temperature series from 1850 to 2020  is certainly more appropriate than a short daily series from 2001 to 2020
 is certainly more appropriate than a short daily series from 2001 to 2020  Indeed, Figure 3 shows that simply increasing the resolution of a time series does not change the shape of the periodogram in the low-frequency range. For each frequency, annual (3.a), monthly (3.b), and daily (3.c), the scatter plot based on the low-frequency relationship approximation (14) corroborates our suspicion that
 Indeed, Figure 3 shows that simply increasing the resolution of a time series does not change the shape of the periodogram in the low-frequency range. For each frequency, annual (3.a), monthly (3.b), and daily (3.c), the scatter plot based on the low-frequency relationship approximation (14) corroborates our suspicion that  Moreover, after multiplication by the factor (12), the periodograms still increase as
 Moreover, after multiplication by the factor (12), the periodograms still increase as  (see 3.d-f), which is inconsistent with stationarity of the original series.
 (see 3.d-f), which is inconsistent with stationarity of the original series. |  | (15) | 
 to a more complex process
 to a more complex process  In the frequency domain, this transformation is accomplished by multiplying the constant spectral density
 In the frequency domain, this transformation is accomplished by multiplying the constant spectral density|  | (16) | 
|  | (17) | 
|  | (18) | 
 to a possibly stationary process
 to a possibly stationary process  The spectral density of
 The spectral density of  is obtained from the spectral density of
 is obtained from the spectral density of  by multiplication with the factor
 by multiplication with the factor|  | (19) | 
|  | (20) | 
|  | (21) | 
 must be smaller than
 must be smaller than  or else the decline will vanish completely. The result of the dampened differencing is then given by
 or else the decline will vanish completely. The result of the dampened differencing is then given by|  | (22) | 
|  | (23) | 
 we need to find suitable values of
 we need to find suitable values of  and
 and  so that a plot of the log of (23) against
 so that a plot of the log of (23) against  has approximately a slope of 0.5 in the neighborhood of frequency zero. Table 1 gives pairs of values of
 has approximately a slope of 0.5 in the neighborhood of frequency zero. Table 1 gives pairs of values of  and
 and  for various sample sizes which mimic the effect of root differencing. These values were found by minimization of
 for various sample sizes which mimic the effect of root differencing. These values were found by minimization of |  | (24) | 
 and
 and  where
 where  The values in Table 1 can also be used for the approximation of the root integration operator
 The values in Table 1 can also be used for the approximation of the root integration operator  which is just the inverse operator of the root differencing operator
 which is just the inverse operator of the root differencing operator  Indeed, if
 Indeed, if  is obtained from
 is obtained from  by approximate root differencing, i.e.,
 by approximate root differencing, i.e., |  | (25) | 
 can be obtained from
 can be obtained from  by approximate root integrating, i.e.,
 by approximate root integrating, i.e., |  | (26) | 
| 
 | 
|  | (27) | 
 in Figures 4.b, e, h. Of special interest is the low frequency range (see Figures 4.a, d, g), where the graphs are approximately linear with slope 0.5. In this frequency range, the fit obtained by truncated series approximations is generally worse (see Figures 4.c, f, i). In the following, we will therefore change our setting slightly to make up for the shortcomings of the latter approximation. Firstly, we will introduce an initial settling period of length
 in Figures 4.b, e, h. Of special interest is the low frequency range (see Figures 4.a, d, g), where the graphs are approximately linear with slope 0.5. In this frequency range, the fit obtained by truncated series approximations is generally worse (see Figures 4.c, f, i). In the following, we will therefore change our setting slightly to make up for the shortcomings of the latter approximation. Firstly, we will introduce an initial settling period of length  and thereby reduce the analysis period from
 and thereby reduce the analysis period from  Secondly, we will use an expanding cut-off lag instead of a fixed one.
 Secondly, we will use an expanding cut-off lag instead of a fixed one.  and
 and  respectively, are overdifferenced as indicated by a log periodogram that decreases as the frequency decreases (see Figures 5.8, 9, 17, 18), and that a root differenced fractional series with
 respectively, are overdifferenced as indicated by a log periodogram that decreases as the frequency decreases (see Figures 5.8, 9, 17, 18), and that a root differenced fractional series with  is underdifferenced as indicated by a log periodogram that still increases as the frequency decreases (see Figures 5.25, 26). These findings are also in line with the results obtained by root differencing in the frequency domain (see Figures 5.7, 16, 25). Moreover, there is also agreement that the strong positive autocorrelation, which is present in the fractional series with
 is underdifferenced as indicated by a log periodogram that still increases as the frequency decreases (see Figures 5.25, 26). These findings are also in line with the results obtained by root differencing in the frequency domain (see Figures 5.7, 16, 25). Moreover, there is also agreement that the strong positive autocorrelation, which is present in the fractional series with  (see Figure 5.22), is (to a lesser extent) still present after root differencing (see Figures 5.23, 24). In contrast, the analogous plots for the stationary series are inconclusive (see Figures 5.4-6, 13-15). However, the truncated series approximation has at least managed to produce a negative first-order autocorrelation in each case (Figure 5.6, 15). In general, this method appears to remove a possible (stochastic) trend slightly more aggressively than the ARMA(1,1) approximation (see Figures 5.1-3, 20-12, 19-21).
 (see Figure 5.22), is (to a lesser extent) still present after root differencing (see Figures 5.23, 24). In contrast, the analogous plots for the stationary series are inconclusive (see Figures 5.4-6, 13-15). However, the truncated series approximation has at least managed to produce a negative first-order autocorrelation in each case (Figure 5.6, 15). In general, this method appears to remove a possible (stochastic) trend slightly more aggressively than the ARMA(1,1) approximation (see Figures 5.1-3, 20-12, 19-21).  The resulting reduction of the analysis period from 171 to 145 years can to some extent be justified by the fact that the global means of the first years are based on a significantly smaller number of measurements. In particular, this is   true for the years before 1880. Note that a similar surface temperature dataset, namely the GISTEMP v4 (see GISTEMP Team, 2021; Lenssen, 2019) provided by the NASA (https://data.giss.nasa.gov/gistemp/), is only available from 1880.The central question is whether the recent rise in temperature is just a transient phenomenon or an indication of nonstationarity. Applying the methods used for the production of Figure 5 to our global surface temperature series, we find that all indications point to nonstationarity. Firstly, the root differenced series (obtained from the mean-corrected original series) still exhibit some kind of an upward trend (see Figures 6.d, g). Secondly, there is also strong positive autocorrelation left after root differencing (see Figures 6.e, h). Thirdly, the log periodogram of the root differenced series always increases as the frequency decreases, regardless whether root differencing is carried out in the frequency domain (see Figure 6.c) or in the time domain (see Figures 6.f, i).
 The resulting reduction of the analysis period from 171 to 145 years can to some extent be justified by the fact that the global means of the first years are based on a significantly smaller number of measurements. In particular, this is   true for the years before 1880. Note that a similar surface temperature dataset, namely the GISTEMP v4 (see GISTEMP Team, 2021; Lenssen, 2019) provided by the NASA (https://data.giss.nasa.gov/gistemp/), is only available from 1880.The central question is whether the recent rise in temperature is just a transient phenomenon or an indication of nonstationarity. Applying the methods used for the production of Figure 5 to our global surface temperature series, we find that all indications point to nonstationarity. Firstly, the root differenced series (obtained from the mean-corrected original series) still exhibit some kind of an upward trend (see Figures 6.d, g). Secondly, there is also strong positive autocorrelation left after root differencing (see Figures 6.e, h). Thirdly, the log periodogram of the root differenced series always increases as the frequency decreases, regardless whether root differencing is carried out in the frequency domain (see Figure 6.c) or in the time domain (see Figures 6.f, i).  hence root differencing is definitely not enough.
 hence root differencing is definitely not enough.