I am having a hard time understanding the link between the ergodicity of a random process and the ergodicity of what is called a measure-preserving transformation T as stated in wikipedia here.
Let $(X,\; \Sigma ,\; \mu\,)$ be a probability space, and $T:X \to X$ be a measure-preserving transformation. We say that T is ergodic with respect to $\mu$ (or alternatively that $\mu$ is ergodic with respect to T) if the following equivalent conditions hold:
for every $E \in \Sigma$ with $T^{-1}(E)=E\,$ either $\mu(E)=0\,$ or $\mu(E)=1\,$;
for every $E \in \Sigma$ with ${\displaystyle \mu > (T^{-1}(E)\bigtriangleup E)=0}$ we have $\mu(E)=0$ or $\mu(E)=1\,$ (where $\bigtriangleup$ denotes the symmetric difference);
for every $E \in \Sigma$ with positive measure we have ${\displaystyle \mu \left(\bigcup _{n=1}^{\infty }T^{-n}(E)\right)=1}$;
for every two sets E and H of positive measure, there exists an n > 0 such that ${\displaystyle \mu ((T^{-n}(E))\cap H)>0}$;
Every measurable function $f:X\to\mathbb{R}$ with $f\circ T=f$ is almost surely constant.
I actually found a very similar question on SE here:
Definition of ergodicity and ergodic process
But I still can't really picture what does T represent, and what a measure-preserving transformation is in the world of random processes. Has it something to do with a time translation?
Right now, the idea I have in mind of an ergodic random process is a process for which the distribution over the whole signal is somehow similar to the distribution of that very signal at any specific point in time.
An answer with an illustrated example would be very welcome.