This paper presents an approach for constrained Gaussian Process (GP)
regression where we assume that a set of linear transformations of the process
are bounded. It is motivated by machine learning applications for
high-consequence engineering systems, where this kind of information is often
made available from phenomenological knowledge. We consider a GP f over
functions on X⊂Rn taking values in
R, where the process Lf is still Gaussian when
L is a linear operator. Our goal is to model f under the
constraint that realizations of Lf are confined to a convex set of
functions. In particular, we require that a≤Lf≤b, given
two functions a and b where a<b pointwise. This formulation provides a
consistent way of encoding multiple linear constraints, such as
shape-constraints based on e.g. boundedness, monotonicity or convexity. We
adopt the approach of using a sufficiently dense set of virtual observation
locations where the constraint is required to hold, and derive the exact
posterior for a conjugate likelihood. The results needed for stable numerical
implementation are derived, together with an efficient sampling scheme for
estimating the posterior process.Comment: Published in JMLR: http://jmlr.org/papers/volume20/19-065/19-065.pd