8

Let $X$ be in integrable, with density $f$ with respect to the Lebesgue measure. Compute the conditional expectation : $ \operatorname{E} \left[ X\, \Big|\, |X| \,\right] $

My ansatz was : $ \operatorname{E} \left[ X\, \Big|\, |X| \,\right] = \operatorname{E} \left[ \operatorname{sgn}(X)\cdot |X| \, \Big|\, |X| \,\right] = |X| \operatorname{E} \left[ \operatorname{sgn}(X) \, \Big|\, |X| \,\right] $

At this point I am stuck since one cannot assume in general that $\operatorname{sgn}(X)$ is independent of $|X|$ nor measurable.

Did
  • 284,245
fred00
  • 83
  • The set ${0}$ has Lebesgue measure $0$, so you treat this as if $X:\Omega\to\mathbb R \setminus {0},$ and then $x\mapsto\operatorname{sgn}(x)$ is an everywhere continuous function, so for $\omega\in\Omega,$ the mapping $\omega \mapsto X(\omega) \mapsto \operatorname{sgn}(X(\omega))$ is measurable. $\qquad$ – Michael Hardy Jan 21 '17 at 17:03
  • Thanks for your answer Michael, but i do not fully understand how do you get from the continouity of $x \rightarrow sgn(x)$ to the measurability w.r.t $\mathcal{\sigma}(|X|)$ – fred00 Jan 21 '17 at 18:12
  • So what you meant was you don't know if $\operatorname{sgn}(X)$ is measurable with respect to $\sigma(|X|)\text{?}\qquad$ – Michael Hardy Jan 21 '17 at 21:09
  • If $\operatorname{sgn}(X)$ were measurable with respect to $\sigma(|X|)$ then the conditional expected value of $\operatorname{sgn}(X)$ given $|X|$ would just be $\operatorname{sgn}(X).$ But maybe I haven't yet understood your question. $\qquad$ – Michael Hardy Jan 21 '17 at 21:19
  • Dear Michael, I thought you claim that $sgn(X)$ is measurable w.r.t $\sigma(|X|)$ and i didn't understand why. My original question is to compute the conditional expectation: $\mathbb{E}[X | |X|]$ – fred00 Jan 21 '17 at 23:01
  • For example if X is uniform on [-1,1] then I calculated this cond. expectation as follows : $\mathbb{E}[X| |X|]=|X|\cdot\mathbb{E}[sgn(X)| |X| ]=|X|\cdot\mathbb{E}[sgn(X)|]=|X|\cdot \frac{1}{2}$. Where I used that sgn(X) is indipendent of |X| because of the symmetry of the distribution around 0. Now I wanted to know if this cond. expectation can be also simplified in the case of a general $X\sim f$. In this case I can't use the independence of sgn(X) and |X|. However my ansatz might be wrong. Hopefully you understand now my question. – fred00 Jan 21 '17 at 23:13
  • Similar: https://math.stackexchange.com/q/829779/321264 – StubbornAtom Sep 20 '22 at 14:37

1 Answers1

8

Disclaimer: This answer is currently downvoted. Needless to say, it is perfectly correct, and it answers the question, apparently to the OP's satisfaction. The downvote might be due to extra-mathematical reasons not worthy of being further commented upon. Happy reading!

Let $Y=|X|$. When $X$ has a PDF $f$, you might convince yourself that $$E(X\mid Y)=g(Y)$$ where $g(y)$ is defined on $y\geqslant0$ by

$$g(y)=y\cdot\frac{f(y)-f(-y)}{f(y)+f(-y)}$$

In the general case, some more measure theoretical machinery is needed. Consider the measures defined by $$\mu_+(B)=P(X\in B)\quad \mu_-(B)=P(-X\in B)\quad \mu(B)=P(Y\in B)$$ for every Borel set $B\subset\mathbb R_+$. Then $\mu_\pm\leqslant\mu$ hence $\mu_\pm\ll\mu$ hence there exists some measurable functions $f_\pm$ such that $$\mu_\pm(B)=\int_Bf_\pm d\mu$$ for every Borel set $B$. Then $E(X\mid Y)=g(Y)$ where $g(y)$ is defined on $y\geqslant0$ by

$$g(y)=y\cdot\frac{f_+(y)-f_-(-y)}{f_+(y)+f_-(-y)}$$

Edit 1: This conditional expectation reflects the fact that, for every $y>0$, the conditional distribution of $X$ conditionally on $Y=y$ is given by $$P(X=y\mid Y=y)=\frac{f_+(y)}{f_+(y)+f_-(-y)}$$ and $$P(X=-y\mid Y=y)=\frac{f_-(y)}{f_+(y)+f_-(-y)}$$ which can be summarized as $$P(X=Y\mid Y)=\frac{f_+(Y)}{f_+(Y)+f_-(-Y)}=1-P(X=-Y\mid Y)$$ Edit 2: As often on this site, one meets a specific problem when answering questions involving conditional expectations, which is that every nontrivial one requires a solid definition of the concept. To understand why the OPs interested in such questions regularly lack any such definition would require a separate analysis, hence, for the time being, we will concentrate on adding some sketchy explanations on this definition, referring the interested reader to some accessible and rigorous source (such as the small blue textbook Probability with martingales, by David Williams).

Thus, in full generality, one considers random variables $X$ and $Y$ defined on the same probability space with $X$ integrable, then $E(X\mid Y)$ is defined as the (class of) random variable(s) $u(Y)$, for some measurable function $u$ such that $u(Y)$ is integrable and, for every bounded measurable function $w$, $$E(Xw(Y))=E(u(Y)w(Y))$$ Equivalently, one can ask that $u(Y)$ is integrable and that, for every event $A$ in $\sigma(Y)$, $$E(X\mathbf 1_A)=E(u(Y)\mathbf 1_A) $$ In the present case, to prove the formulas proposed in this post when $X$ has PDF $f$, one should find $g$ such that, for every bounded measurable function $w$, $$E(Xw(Y))=E(g(Y)w(Y))$$ that is, since $Y=|X|$ and $X$ has PDF $f$, $$\int xw(|x|)f(x)dx=\int g(|x|)w(|x|)f(x)dx$$ which is equivalent to $$\int_{y>0}yw(y)(f(y)-f(-y))dy=\int_{y>0}g(y)w(y)(f(y)+f(-y))dy$$ This identity holds for every function $w$ if and only if $$y\cdot(f(y)-f(-y))=g(y)\cdot(f(y)+f(-y))$$ almost everywhere for $y>0$, and $g$ follows.

Did
  • 284,245
  • First of all thank you Did for your detailed explanation. I have just one more question. For the part $E(X|Y)=g(Y)$: I need the conditional distribution $ X | Y=y = \begin{cases} y, & \text{with probability p}\ -y ,& \text{with probability 1-p} \end{cases}$. As I can conclude after seeing your result you claim that $p=\frac{f(y)}{f(y)+f(-y)}$. How did you got that ? – fred00 Jan 22 '17 at 22:55
  • I know that the density of Y is $f_{y}(y)=f(y)+f(-y) \quad ,y>0$. I tried $\mathbb{P}[X=y | Y=y ]=\frac{\mathbb{P}[X=y,Y=y]}{P[Y=y]}=\frac{\mathbb{P}[X=y]}{P[Y=y]}=\frac{f(y)}{f(y)+f(-y)}$. But I do not fully understand why the last equality should hold? For example if X has a cont. distribution. Thank you in advance for your support. – fred00 Jan 22 '17 at 23:05
  • Of course, when $X$ has a PDF, $P(Y=y)=0$ for every $y$ hence none of the formulas in your comment makes any sense. I added a brief reminder of the definition of conditional expectations in the general case. – Did Jan 23 '17 at 11:37
  • Thank you for the solution Did ! – fred00 Jan 23 '17 at 14:58
  • @Did Let $X$ be standard normal, then $f_x(y)-f_x(-y)=0$ and hence $E(X | |X|)=0$, this looks like an interesting result and a bit funny ... can it actually be? – Math-fun Jan 23 '17 at 16:12
  • @Math-fun Yes, it is quite natural if you think about it... and it holds for every symmetric random variable. – Did Jan 23 '17 at 16:23
  • @Did I think your result is correct for more general case. Let $\mu_X,\mu_{-X}$ be the distributions of $X, -X$ on $\mathbb{R}$, and $f_+=\frac{d\mu_X}{d(\mu_X+\mu_{-X})}$ be the Radon-Nikodym derivative, then $E(X| |X|)=|X|(2f_+(|X|)-1)$. – JGWang Jan 24 '17 at 08:59
  • @JGWang The "more general case" you describe is exactly the case solved in my answer, starting at "In the general case, some more measure theoretical machinery is needed". Not sure I am following your point... – Did Jan 24 '17 at 09:05
  • @Did:“more general case": The random variable X is discrete distributed, the distribution function of X is absolutely continuous and other cases. – JGWang Jan 24 '17 at 09:14
  • @JGWang All dealt with in my answer (did you read it?). – Did Jan 24 '17 at 09:26
  • @Did: In your answer, I read the following sentence: "to prove the formulas proposed in this post when $X$ has PDF $f$''. Now, if the $X$ doesn't have a PDF, is the conclusion still effective. In my comment, I just say that, "$E(X| |X|)=|X|(2f_+(|X|)-1)$, $f_+=\frac{d\mu_X}{d(\mu_X+\mu_{-X})}$. It is effective in any case, even $X$ doesn't have a PDF. – JGWang Jan 25 '17 at 03:01
  • @JGWang Yeah, my answer explicitely states that the conclusion holds in the general case, in the general formulation given using $f_\pm$. But perhaps what you mean is that the proof (not only the formula) in this general case is similar to the proof I explain in the restricted case where $X$ has a PDF $f$? Then you are right. – Did Jan 25 '17 at 08:21
  • @Did Thanks for your answer. – JGWang Jan 26 '17 at 01:46