Error function of noisy input and target variables

by Robert Smith   Last Updated September 07, 2018 07:19 AM

Why is the sum of squares error function of noisy input and noisy target variables very similar to the error function for only noisy input?

This is the relevant part in Bishop's book:

Another viewpoint on kernel regression comes from a consideration of regression problems in which the input variables as well as the target variables are corrupted with additive noise. Suppose each target value $t_{n}$ is generated as usual by taking a function $y(z_{n})$ evaluated at a point $z_{n}$, and adding Gaussian noise. The value of $z_{n}$ is not directly observed, however, but only a noise corrupted version $x_{n}=z_{n}+\xi_{n}$ where the random variable $\xi$ is governed by some distribution $g(\xi)$.

$$E=\frac{1}{2}\sum_{n=1}^{N}\int \{y(x_{n}-\xi_{n})-t_{n}\}^{2}g(\xi_{n})d\xi_{n}$$

However, the error function when we consider a noisy input variable is extremely similar:

$$E=\frac{1}{2}\sum_{n=1}^{N}\int \{y(x_{n}+\xi_{n})-t_{n}\}^{2}\nu(\xi_{n})d\xi_{n}$$


Related Questions

Viewing kernel regression in a Bayesian framework

Updated March 20, 2019 14:19 PM

How do I calculate the smoother matrix?

Updated March 24, 2018 17:19 PM

accuracy conditional on feature values

Updated April 25, 2018 23:19 PM

How to use simpleSmootherC? (R Locpol)

Updated March 20, 2017 07:19 AM