Context¶
We are given some approximations of inputs"
Uncertainty Inputs:Model:xx∼N(μx,Σx)=μx+εx,εx∼N(0,Σx) We have the data likelihood given
y∼p(y∣x,θ)
We assume that this predictive density is given by a Gaussian distribution
y∼p(y∣x,θ)≈N(y∣μy,Σy) where hμ and hσ is the predictive mean and variance respectively.
For example, this predictive mean could be a basis function, a non-linear function or a neural network.
The predictive variance function could be constant or a simple linear function.
Predictive Mean:Predictive Variance:μyΣy=hμ(x;θ),=hσ2(x;θ),hμ:RDx×Θ→RDyhσ2:RDx×Θ→RDy×Dy We have ways to estimate these quantities as follows using the law of iterated expectations.
μy(x;θ)σy(x;θ)=Ex[hμ(x,θ)]=Ex[hσ2(x,θ)]+Ex[hμ2(x,θ)]−Ex2[hμ(x,θ)] In integral form, we can write this as:
μy(x;θ)σy(x;θ)=∫μx(x,θ)p(x)dx=∫hσ2(x,θ)p(x)dx+∫hμ2(x,θ)p(x)dx−[∫μx(x,θ)p(x)dx]2
Taylor Approximation¶
Ex[hμ(x,θ)]Ex[hμ(x,θ)]Ex[hμ(x,θ)]≈hμ(μx,θ)+21Tr[∂2hμ(μx,θ)Σx]≈hμ(μx,θ)+21Tr[∂2hμ(μx,θ)Σx]≈hμ(μx,θ)+21Tr[∂2hμ(μx,θ)Σx]
Moment-Matching¶