Ohio State is in the process of revising websites and program materials to accurately reflect compliance with the law. While this work occurs, language referencing protected class status or other activities prohibited by Ohio Senate Bill 1 may still appear in some places. However, all programs and activities are being administered in compliance with federal and state law.

Seminar: Alexander Kukush

Statistics Seminar
July 29, 2004
All Day
209 W. Eighteenth Ave. (EA), Room 170

Title

Asymptotic Properties of Covariance-Matching Constrained Kriging

Speaker

Dr. Alexander Kukush, Kiev National University, Belarus

Abstract

Consider a random field S(s), where s belongs to a domain D in d-dimensional Euclidean space, S(s) = µ(s) + d(s). Here µ(·) is the large-scale, deterministic, mean structure of the signal S(·), and d(·) is the small-scale stochastic structure that models te spatial dependence. The S(·) represents the spatial process of fundamental interest, namely the signal. The observations of S(·) at discrete points are contaminated with additive measurement error, according to the model Z(s) = S(s) + e(s). Here e(s) is a zero-mean, white-noise process representing measurement error. Following Aldworth and Cressie (2003) we investigate a predictor of nonlinear functional g (S(·)). The predictor is based on the kriging methodology with extra constraints and called covariance- matching constrained kriging. It is an optimal linear predictor that matches not only first moments but second moments as well.

The underlying optimization problem is linear programming problem, with both linear and quadratic constraints. By geometric reasoning we show that the problem always has a solution provided the constraints are consistent, and propose a new algorithm to compute a solution. We prove that the predictor is asymptotically unbiased. We give an upper bound for the L1-error of prediction. In general the error is not vanishing as the sample size tends to infinity, but it is asymptotically small if the second and third derivatives of nonlinear tranform g(·) are small enough.