I learned from this answer that I can use the Iverson bracket notation to get the expectation of the geometric distribution. The theory behind it seems unfathomable to me now and I just want to intuit the following practice:
Let $X$ be a random variable which is $1$ with probability $\frac{1}{100}$ and $0$ with probability $\frac{99}{100}$. Suppose that I draw independent random variables $X_1$, $X_2$ ... and I stop when I see the first "$1$". For example, if I draw
$$X_1=0, X_2=0, X_3 = 0, X_4=1$$
then I would stop at $X_4$. Let N be the last index that I draw. How big would the expected $N$ be?
Here are my thoughts:
If it comes up $1$(with probability $1/100$), then the expectation is $\frac{1}{100} \cdot 1$; otherwise, I start again except I've already used one trial, but I don't understand why the expectation, in this case, is $(1-\frac{1}{100}) \cdot (1+E[N])$?
Is it for the rest of all trials or just the second trial? If it is the expectation of the rest of all trials why its probability is $1-\frac{1}{100}$? It seems comprehensible that the expectation is $E[N]+1$ because the first trial is fixed and the expectation for the rest is the same as that for the whole trials.
If it is also for the first trial(learned from the aforementioned answer as he/she stated that, quote: "$ X= 1\cdot[\text{H on 1st toss}] + (1+Y)\cdot[\text{T on 1st toss}]$") what is the $(1+E[N])$? I can understand why the probability is $1-\frac{1}{100}$.
Reference: CS109: Probability for Computer Scientists