User:William M. Connolley/Trend estimation
This page now released into the real world as: Trend_estimation
If a given set of measurements are taken of some process that is incompletely understood, one may construct a model, independent of anything known about the physics of the process, to explain the behaviour of the measurement. In particular, one may wish to know if the measurements exhibit an increasing or decreasing trend, that can be statistically distinguished from random behaviour.
This page does not attempt a full mathematical treatment, merely an exposition. It may help to consider the series at issue to be daily average temperatures at a given location, from winter to summer; or the global temperature series over the last 100 years. Particularly in the latter case, issues of homogeneity (is the series equally reliable throughout its length) and so forth are important; for the moment we shall simplify the discussion and neglect them.
Fitting a trend: least-squares
[edit]Given a set of data, and the desire to produce some kind of "model" of that data (model, in this case, meaning a function fitted through the data) there are a variety of functions that can be chosen for the fit. But if there is no prior understanding of the data, the simplest function to fit is a straight line and thus this is the "default".
Continuing, once it has been decided to fit a straight line, there are various ways to do so, but the overwhelming default is the least-squares fit, equivalent to minimisation of the L2 norm.
There is a rather poor wiki page on least squares. I'll continue here, for the while.
Thus, given a set of data points x_i, and data values y_i, one choses a and b so that sum(((a*x_i + b) - y_i))^2) is minimised. This can always be done. Someday I'll write this out neatly, for now you can find the formula at http://mathworld.wolfram.com/LeastSquaresFitting.html.
For the rest of this article, "trend" will mean the least squares trend. Its what it means in 99% of cases everywhere else.
Now, we have a trend. But is it significant? And what do we mean by significant?
Trends in Random data
[edit]Before we can consider trends in real data, we need to understand trends in random data.
If we take a series which is known to be random - fair dice falls; or computer-generated random numbers - and fit a trend line through the data, the chances of a truely zero trend are negligible. But we would probably expect the trend to be "small". If we take a series with a given degree of noise, and a given length (say, 100 points), and generate a large number of such series (say, 100,000 series), we can then calculate the trends from these 100,000 series, and empirically establish a distribution of trends that are to be expected from such random data. Such a distribution will be normal (Central limit theorem except in pathological cases, since (in a slightly non-obvious way of thinking about it) the trend is a linear combination of the y_i) and (if the series genuinely is random) centered on zero. We may now establish a level of statistical certainty, S, desired - 95% confidence is typical; 99% would be stricter, 90% rather looser - and say: what value, V, do we have to choose so that S% of trends are within V? (complication: we may be interested in positive and negative trends - 2-tailed - or may have prior knowledge that only positive, or only negative, trends are of interest).
In the above discussion the distribution of trends was calculated empirically, from a large number of trials. In simple cases (normally distributed random noise being a classic) the distribution of trends can be calculated exactly.
Suppose we then take another series, which we do not know in advance whether it "really" has a trend in it, calculate the trend, T, and discover that it is less than V. Then we may say that, at degree of certainty S, any trend in the data cannot be distinguished from random noise.
However, note that whatever value of S we choose, then a given fraction, 1-S, of truely random series will be declared (falsely, by construction) to have a significant trend. Conversely, a certain fraction of series that "really" have a trend, will be declared to have no trend.
Data as Trend plus Noise
[edit]To analyse a (time) series of data, we assume that it may be represented as trend plus noise:
where a and b are (usually unknown) constants and the e's are independent randomly distributed "errors". Unless something special is known about the e's, they will be assumed to have a normal distribution. It is simplest if the e's all have the same distribution, but if not (if some have higher variance, meaning that that data point is effectively less certain) then this can be taken into account during the least squares fitting, by weighting each point by the inverse of the variance of that point.
In most cases, where only a single time series exists to be analysed, the variance of the e's is estimated by fitting a trend, thus allowing a*t+b to be removed and leaving the e's as residuals, and calculating the variance of the e's from the residuals - this is often the only way of estimating the variance of the e's.
One particular special case of great interest, the (global) temperature time series, is known not to be homogeneous in time: apart from anything else, the number of weather observations has (generally) increased with time, and thus the error associated with estimating the global temperature from a limited set of observations has decreased with time. In fitting a trend to this data, this can be taken into account, as described above.
Once we know the "noise" of the series, we can then assess the significance of the trend by making the null hypothesis that the trend, a, is not significantly different from 0. From the above discussion of trends in random data with known variance, we know the distribution of trends to be expected from random (trendless) data. If the calculated trend, a, is larger than the value, V, then the trend is deemed significantly different from zero at significance level S.
Noisy time series, and an example
[edit]It is harder to see a trend in a noisy time series. For example, if the true series is 0, 1, 2, 3 all plus some independent normally distributed "noise" e of standard deviation E, and we have a sample series of length 50, then if E=0.1 the trend will be obvious; if E=100 the trend will probably be visible; but if E=10000 the trend will be buried in the noise.
If we consider a concrete example, the global surface temperature record of the past 140 years as presented by the IPCC: [1], then the interannual variation is about 0.2 oC and the trend about 0.6 oC over 140 years, with 95% confidence limits of 0.2 oC (by coincidence, about the same value as the interannual variation). Hence the trend is statistically different from 0. This alone, however, tells us nothing about the physical causes of the temperature change.
Real data is auto-correlated
[edit]The above discussion assumed the data could be represented as trend + noise, with the noise at each data point being independent. This is important, as it makes an enormous difference to the ease with which the statistics can be analysed. Real (climate) data rarely fulfills this criterion.
Other stuff not covered here but should be
[edit]Discussion of autocorrelation, reduced d.o.f, etc etc.
Averaging reducing noise.