Professional Documents
Culture Documents
Interpolation
Spatial Interpolation
Interpolation is the method of predicting the value of attributes at unsampled sites from measurements made at known
point locations within the same space. The rationale behind interpolation is that the values at points that lie close
together in the space are more likely to be similar than that of those which lie apart.
Figure 1: Interpolation
Interpolation finds its use in many areas such as estimation of rainfall and temperature at places where no weather
stations are available; contouring, where it is necessary to know where to place contours in between measured
locations; resampling a raster data, where raster is changed from one grid to another
An interpolation method which predicts the value of an attribute at sampled location equal to the value measured at that
location is called an exact interpolator. The rest of the methods are termed as inexact interpolators. The statistics of
difference (absolute or squared difference between the measured and predicted values at data points)
an indicator of the quality of the inexact interpolators.
There are two methods of interpolation:
Global method
Local method
is
Global methods make predictions using the entire dataset. On the other hand local methods make predictions using the
measured points within neighborhoods (smaller spatial areas within the larger study area). Global methods usually
create smooth surfaces, but they can be very sensitive to outliers (values that are significantly different as compared to
the other values). Compared with global methods, local methods yield less smooth surfaces and are not sensitive to
outliers as their effects do not influence the entire interpolated surface, but only local regions/parts of the interpolated
surface.
Where
b and b are the polynomial coefficients known as intercept and slope respectively.
0
In two dimension the multiple regression on x and y derives the polynomial surface of the following form
1x
2y
Kriging does not require the data to be normally distributed. Transformations and trend removal can help justify assumptions of
normality and stationarity. Prediction, using ordinary, simple, and universal kriging for general Box-Cox transformations and
arcsine transformations, is called transgaussian kriging. Prediction for log transformation is called lognormal kriging.
Semivariogram
Semi-variance is half the difference squared between the values of the paired locations separated by a distance h. This plot of
semi-variance with half the squared difference on the y-axis and the distance that separates the locations on the x-axis is called
semivariogram cloud. Plotting all the pairs becomes unmanageable therefore the pairs are grouped into lag bins and the
process of grouping is called binning. Binning form the pair of points and then group the pairs so that they have a common
distance and direction. For each bin, only the average distance and semi-variance for all the pairs in that bin are plotted as a
single point on the empirical semivariogram cloud. The plotted semivariogram is then fitted with a curve known as a
semivariogram model. The distance at which the curve flattens out is called the range. For distances greater than the range, it
is assumed that there exist no autocorrelation between the points. The value (on the y axis) the semivariogram attains at range
is known as sill. The intercept made by the semivariogram model on y axis of the semivariogram is called the nugget.
Figure 3: Semivariogram
Point-ID
Attribute
Area
Poly-ID
Attribute
10
200
10
20
550
20
30
475
30
40
525
40
The major disadvantage of this method is that the attribute inside a polygon is assumed to have homogenous value which
changes only at the boundary. Since a polygon has a single point inside it, the variation within a polygon cannot be estimated.
As the distance increases the weight decreases. The rate at which the weights decrease depends on the value of p. If p is zero
there is no decrease with distance, and because each weight is the same the predicted value is the average of all the data
values in the search neighborhood. As p increases, the weights for distant points decrease rapidly.
An approximately correct value of p can be determined by minimizing the root mean square prediction error (RMSPE), a
statistic which quantifies the error of the prediction surface and is calculated by cross validation. The process of cross validation
involves removal of one or more data locations and predicts their associated data using the data at the rest of the locations. In
this way, one can compare the predicted value to the observed value and obtain useful information about the quality of the
used model.
Splines
Spline functions are piece-wise functions, which are fitted to a small number of data points exactly and at the same time it is
ensured that the joins between one part of the curve and another are continuous. Unlike trend surface, it is possible to modify a
part of the curve without having to recompute the whole, and hence is an advantage.
A piecewise polynomial function is written as:
The points x ,.x are called break points. They divide an interval into k sub-intervals. The points of the curve at these values
of x are called knots. The functions P (x) are polynomials of degree m or less. The term r denotes the constraints on the spline.
i
k-1
References:
Lo, C.P. & Yeung, A. 2009, Concepts and techniques of Geographic Information Systems, PHI Learning Private Limited, New
Delhi.
Goodchild, M.F., Longley, P.A., Maguire, D. J. & Rhind, D.W 2001, Geographic information systems and science, John Wiley &
Sons Ltd. , England.