The Geometry of Signal Extraction
Teasing out the Signal
There’s a classic signal extraction problem stated as follows: you observe a random variable as the sum of two normal distributions, and such that . Given an observation of , what is the conditional expectation of ?
The problem asks us to find . There are a number of reasons why we might want to do so. For starters, if we’re interested in the value of some Gaussian distribution , but we can only observe , the conditional expectation given above is exactly what we’re looking for. In the past I’ve seen it derived hammer and tongs via the definition of conditional expectation:
If and are statistically independent we can express the joint distribution as a product of marginal distributions, fix , and end up with the expression that we’re looking for:
Ouch. It works, but it’s not pretty. Last night I came up with a geometric interpretation for the normal case that I wanted to share. Googling around there are similar derivations but I figured that one more writeup with some deeper explanation wouldn’t hurt.
Regression as an Operator
To start we note a general propriety of conditional expectation: , where is some measurable function. We also need a simple decomposition lemma: any random variable can be written as: , where is a RV s.t. and . The intuition here is that almost by definition any variable can be expressed as a conditional expectation and an error term. The proof is simple:
We need this to prove the following result:
Proof:
From the decomposition property that we proved above
